Abstract
Stackelberg games have been widely used to model interactions between attackers and defenders in a broad array of security domains. One related approach involves plan interdiction, whereby a defender chooses a subset of actions to block (remove), and the attacker constructs an optimal plan in response. In previous work, this approach has been introduced in the context of Markov decision processes (MDPs). The key challenge, however, is that the state space of MDPs grows exponentially in the number of state variables. We propose a novel scalable MDP interdiction framework which makes use of factored representation of state, using a parity function basis for representing a value function over a Boolean space. We demonstrate that our approach is significantly more scalable than prior art, while resulting in near-optimal interdiction decisions.
| Original language | English |
|---|---|
| State | Published - 2017 |
| Event | 33rd Conference on Uncertainty in Artificial Intelligence, UAI 2017 - Sydney, Australia Duration: Aug 11 2017 → Aug 15 2017 |
Conference
| Conference | 33rd Conference on Uncertainty in Artificial Intelligence, UAI 2017 |
|---|---|
| Country/Territory | Australia |
| City | Sydney |
| Period | 08/11/17 → 08/15/17 |