Learning to Attack: A Bandit Approach to Adversarial Context Poisoning
arXiv:2603.00567v1 Announce Type: new Abstract: Neural contextual bandits are vulnerable to adversarial attacks, where subtle perturbations to rewards, actions, or contexts induce suboptimal decisions. We introduce AdvBandit, a black-box adaptive attack that formulates context poisoning as a continuous-armed bandit problem, enabling the attacker to jointly learn and exploit the victim’s evolving policy. The attacker requires no access to the victim’s internal parameters, reward function, or gradient information; instead, it constructs a surrogate model using a maximum-entropy inverse reinforcement […]