Alpcan, T., and T. Basar (2004) “A game theoretic analysis of intrusion detection in access control systems” Proceedings of 43 rd IEEE Conference on Decision.

Slides:



Advertisements
Similar presentations
Nash’s Theorem Theorem (Nash, 1951): Every finite game (finite number of players, finite number of pure strategies) has at least one mixed-strategy Nash.
Advertisements

Price Of Anarchy: Routing
Mixed Strategies CMPT 882 Computational Game Theory Simon Fraser University Spring 2010 Instructor: Oliver Schulte.
Introduction to Game theory Presented by: George Fortetsanakis.
Joint Strategy Fictitious Play Sherwin Doroudi. “Adapted” from J. R. Marden, G. Arslan, J. S. Shamma, “Joint strategy fictitious play with inertia for.
Calibrated Learning and Correlated Equilibrium By: Dean Foster and Rakesh Vohra Presented by: Jason Sorensen.
An Introduction to... Evolutionary Game Theory
MIT and James Orlin © Game Theory 2-person 0-sum (or constant sum) game theory 2-person game theory (e.g., prisoner’s dilemma)
EC3224 Autumn Lecture #04 Mixed-Strategy Equilibrium
Markov Game Analysis for Attack and Defense of Power Networks Chris Y. T. Ma, David K. Y. Yau, Xin Lou, and Nageswara S. V. Rao.
EC941 - Game Theory Lecture 7 Prof. Francesco Squintani
How Bad is Selfish Routing? By Tim Roughgarden Eva Tardos Presented by Alex Kogan.
Presenter: Raghu Ranganathan ECE / CMR Tennessee Technological University March 22th, 2011 Smart grid seminar series Yao Liu, Peng Ning, and Michael K.
1. 2 APPENDIX B A PROCEDURE FOR GENERATING AN EQUILIBRIUM POINT FOR 2-PERSON GAMES (That sometimes works!)
Part 3: The Minimax Theorem
Equilibrium Concepts in Two Player Games Kevin Byrnes Department of Applied Mathematics & Statistics.
Chapter 6 © 2006 Thomson Learning/South-Western Game Theory.
1 A Game Theoretic Approach for Active Defense Peng Liu Lab. for Info. and Sys. Security University of Maryland, Baltimore County Baltimore, MD OASIS,
Christos alatzidis constantina galbogini.  The Complexity of Computing a Nash Equilibrium  Constantinos Daskalakis  Paul W. Goldberg  Christos H.
Rational Learning Leads to Nash Equilibrium Ehud Kalai and Ehud Lehrer Econometrica, Vol. 61 No. 5 (Sep 1993), Presented by Vincent Mak
The Cat and The Mouse -- The Case of Mobile Sensors and Targets David K. Y. Yau Lab for Advanced Network Systems Dept of Computer Science Purdue University.
Lecture 1 - Introduction 1.  Introduction to Game Theory  Basic Game Theory Examples  Strategic Games  More Game Theory Examples  Equilibrium  Mixed.
Dynamic Adversarial Conflict with Restricted Information Jason L. Speyer Research Asst.: Ashitosh Swarup Mechanical and Aerospace Engineering Department.
Defending Complex System Against External Impacts Gregory Levitin (IEC, UESTC)
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
Learning in Games. Fictitious Play Notation! For n Players we have: n Finite Player’s Strategies Spaces S 1, S 2, …, S n n Opponent’s Strategies Spaces.
Location Estimation in Sensor Networks Moshe Mishali.
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
An Introduction to Game Theory Part III: Strictly Competitive Games Bernhard Nebel.
APEC 8205: Applied Game Theory Fall 2007
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
A Scalable Network Resource Allocation Mechanism With Bounded Efficiency Loss IEEE Journal on Selected Areas in Communications, 2006 Johari, R., Tsitsiklis,
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley Asynchronous Distributed Algorithm Proof.
May 11, 2005 Tracking on a Graph Songhwai Oh Shankar Sastry Target trajectoriesEstimated tracks Tracking in Sensor Networks Target tracking is a representative.
Games in the normal form- An application: “An Economic Theory of Democracy” Carl Henrik Knutsen 5/
UNIT II: The Basic Theory Zero-sum Games Nonzero-sum Games Nash Equilibrium: Properties and Problems Bargaining Games Bargaining and Negotiation Review.
March 24, 2003Upadhyaya – IWIA A Tamper-resistant Framework for Unambiguous Detection of Attacks in User Space Using Process Monitors R. Chinchani.
Data Selection In Ad-Hoc Wireless Sensor Networks Olawoye Oyeyele 11/24/2003.
By: Gang Zhou Computer Science Department University of Virginia 1 A Game-Theoretic Framework for Congestion Control in General Topology Networks SYS793.
MAKING COMPLEX DEClSlONS
Mechanisms for Making Crowds Truthful Andrew Mao, Sergiy Nesterko.
Algorithms for a large sparse nonlinear eigenvalue problem Yusaku Yamamoto Dept. of Computational Science & Engineering Nagoya University.
Derivative Action Learning in Games Review of: J. Shamma and G. Arslan, “Dynamic Fictitious Play, Dynamic Gradient Play, and Distributed Convergence to.
ECO290E: Game Theory Lecture 12 Static Games of Incomplete Information.
General Polynomial Time Algorithm for Near-Optimal Reinforcement Learning Duke University Machine Learning Group Discussion Leader: Kai Ni June 17, 2005.
1. 2 Non-Cooperative games Player I Player II I want the maximum payoff to Player I I want the maximum payoff to Player II.
Presenter: Jen Hua Chi Adviser: Yeong Sung Lin Network Games with Many Attackers and Defenders.
Dynamic Games of complete information: Backward Induction and Subgame perfection - Repeated Games -
Report on Intrusion Detection and Data Fusion By Ganesh Godavari.
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Lecture 2: two-person non.
Game Theory: introduction and applications to computer networks Game Theory: introduction and applications to computer networks Introduction Giovanni Neglia.
Dominance Since Player I is maximizing her security level, she prefers “large” payoffs. If one row is smaller (element- wise) than another,
Issues on the border of economics and computation נושאים בגבול כלכלה וחישוב Speaker: Dr. Michael Schapira Topic: Dynamics in Games (Part III) (Some slides.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley.
1 - CS7701 – Fall 2004 Review of: Detecting Network Intrusions via Sampling: A Game Theoretic Approach Paper by: – Murali Kodialam (Bell Labs) – T.V. Lakshman.
Information Theory for Mobile Ad-Hoc Networks (ITMANET): The FLoWS Project Competitive Scheduling in Wireless Networks with Correlated Channel State Ozan.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
CHAPTER 10 Widrow-Hoff Learning Ming-Feng Yeh.
1. 2 You should know by now… u The security level of a strategy for a player is the minimum payoff regardless of what strategy his opponent uses. u A.
1 a1a1 A1A1 a2a2 a3a3 A2A Mixed Strategies When there is no saddle point: We’ll think of playing the game repeatedly. We continue to assume that.
Resource Allocation in Hospital Networks Based on Green Cognitive Radios 王冉茵
Mixed Strategies Keep ‘em guessing.
Non-additive Security Games
Game Theory in Wireless and Communication Networks: Theory, Models, and Applications Lecture 2 Bayesian Games Zhu Han, Dusit Niyato, Walid Saad, Tamer.
Multiagent Systems Game Theory © Manfred Huber 2018.
Stochastic Game Models in Cyber Security
Information Theoretical Analysis of Digital Watermarking
Game Theory: The Nash Equilibrium
Collaboration in Repeated Games
Normal Form (Matrix) Games
Presentation transcript:

Alpcan, T., and T. Basar (2004) “A game theoretic analysis of intrusion detection in access control systems” Proceedings of 43 rd IEEE Conference on Decision and Control. A review by Matthew H. Henry October 12, 2005

Description of Game Network intrusion scenario Three players: –Intruder –Virtual Sensor Network (VSN) –Intrusion Detection System (IDS) Finite and dynamic games

Virtual Sensor Network Network of software sensors: S = {s 1, s 2, …, s max } Each sensor is an autonomous agent that either –Seeks to match known network intrusion activity signatures; or –Looks for anomalies in network usage that might indicate nefarious activity Each sensor reports findings to IDS core directly or via sensor hierarchy Sensors are “mobile” and can be instantiated and deployed at will by the IDS to monitor different subsystems In general, each sensor is capable of identifying one or more intrusion mechanisms The “strategy” of the VSN consists of a fixed probability distribution for each mode of attack and corresponds to the VSN output during that attack

The IDS and the Attacker The target system is decomposed into t max subsystems: {t 1, t 2, …, t max } There exist I max possible modes of attack: {I 1, I 2, …, I max } Each attack is an ordered pair: a k = (t i, I j ) The game, aside from the imperfect information afforded by the VSN, is a non-cooperative non-zero sum game played by the Attacker and the IDS –Attacker benefits from a successful intrusion and suffers a cost at being detected –The IDS benefits from a successful detection and suffers a penalty (in the form of network performance reduction) from a false positive – the IDS must manage a security tradeoff

Sensor Network Overlay on Protected System Attack LAN SCADA Control Center PCS 1 PCS 2 PCS M SCADA MTU 1 SCADA MTU N SCADA MTU 2 PCS m-1 PCS m PCS M-1 Process 1 Process 2 Process P Process p Process P-1 Intrusion Detection System (IDS) S S S S S SS S S S S S S S S S S S S S S S

Simple Example: single-move finite game Target system comprises a single subsystem and there exists a single possible mode of attack  a 0 = (1, 1) IDS strategy set includes two possible moves: {Take action against attacker, Do nothing} Attacker strategy set includes two possible moves: {Attack, Not attack} VSN “strategy” set includes two probability distributions: {[p 10, p 11 ], [p 00, p 01 ]}, where –p 10 = P(No attack detected | Attack occurred) –p 11 = P(Attack detected | Attack occurred) –p 00 = P(No attack detected | No attack occurred) –p 01 = P(Attack detected | No attack occurred) –Action taken when Attack Detected: Set Alarm

Simple Example: single-move finite game Unique Nash equilibrium in mixed strategies with probability distributions and payoffs shown above. (Solution found using GAMBIT)

Continuous Game Problems with finite game: –Exhibits poor scalability for large systems and high-dimensional action spaces –Payoff values must be separately defined for each possible outcome Propose continuous-kernel game with continuous strategy spaces and cost functions to improve scalability and generalization

Attacker Strategy Space Let A max denote the cardinality of the attack set of ordered pairs a k = (t i, I j ) The strategy space of the attacker is now a subset of  A max Attacker strategy u A  U A   A max, with elements u A i  0, i = 1, 2, …, A max

IDS Strategy Space Let R max denote the cardinality of the response set available to the IDS The strategy space of the IDS is now a subset of  R max IDS strategy u I  U I   R max, with elements u I i  0, i = 1, 2, …, R max

Virtual Sensor Network Sensor output as functions of attacker actions represented as a linear transformation P in the space U S   A max × A max The matrix P = [p ij ], i,j = 1…A max, maps attacker actions to sensor output Sensor output = ( u A ) T P  e.g. ideal P would be the Identity matrix: sensors perfectly detect and report attacker strategy Detection metric for attack a i : dq(i)= p ij /rowsum(p ij ) Define P = [p ij ] = [-p ij ] for i=j, [p ij ] otherwise  this provides positive cost for erroneous detection and negative cost (positive benefit) for correct detection

IDS Cost Function J I ( u A, u I, P) = γ( u A ) T PQu I (cost of false detection/benefit of correct detection) + ( u I ) T diag(  ) u I (cost of resource allocation) + ( c I ) T (Q u A – Q u I ) (cost of successful attack) γ – scalar gain for cost/benefit of false/correct detection Q – A max × R max matrix of binary values (0/1) that maps IDS response actions to attacks Q - A max × A max diagonal matrix with elements  1, signifying the degree of vulnerability of specific subsystems to attacks diag(  ) = diag([  1  2 …  Rmax ])– cost of response actions c I = [c I 1 c I 2 … c I Amax ] – cost of each attack to IDS

IDS Cost Function (Example) 2-Dimensional attack space: u A = [u A 1 u A 2 ] corresponding to one attack mode on two subsystems 1-dimensional IDS response space u I γ =1 Q = [1 1] T – IDS response is same for both attacks Q = 2-Dim Identity Matrix – both subsystems equally vulnerable to this attack diag(  ) =  = 1 c I = [1 2] – attack on subsystem 2 twice as costly as an attack on subsystem 1 P = [.8.2;.3.7]

Attacker Cost Function J A ( u A, u I, P) = -γ( u A ) T PQu I (cost of capture/benefit of successful intrusion) + ( u A ) T diag(  ) u A (cost of resource allocation) + ( c A ) T (Q u I – Q u A ) (benefit of successful attack) diag(  ) = diag([  1  2 …  A max ])– cost of attack resources c A = [c I 1 c I 2 … c I A max ] – benefit of each attack to attacker

“Optimal” Trajectories Minimizing the cost functions yield the following reaction functions u I ( u A, P) = [  I – γ[diag(2  )] -1 Q T P T u A ] + u A ( u I, P) = [  A + γ[diag(2  )] -1 PQ u I ] + Where –  I = [( c I Q) 1 /(2  1 )…( c I Q) Rmax /(2  Rmax )] –  A = [( c A Q) 1 /(2  1 )…( c A Q) Amax /(2  Amax )] –[] + indicates that negative elements are mapped to zero note: this is not best response in the sense of fictitious play since u A is unknown to the IDS, and u I is unknown to the Attacker

Nash Equilibrium Strategy pair ( u I*, u A* ) is in Nash equilibrium if they jointly minimize cost: – u I* = argmin uI { J I ( u A*, u I, P)} – u A* = argmin uA { J I ( u A, u I*, P)} The authors prove that a unique interior Nash equilibrium exists for constrained values of γ which force u I* to be positive in the equilibrium solution Their proof uses the convexity of the cost functions and derives a Hessian for the coupled cost vector [J I J A ] to show uniqueness of the interior solution

Repeated Games Incorporates dynamics associated with improving sensing capability (learning) and sensor reconfiguration (reallocation of IDS/VSN resources) Reflected in dynamic P matrix e.g. P(n+1)=[P(n)+2  (  +  )(diag(diag( u A)Q u I ) -  col(diag( u A )Q u I ))+  W(n)] N Where – , ,  are small positive constants –  ~ U([-1,1]) –W(n) = [w ij ]   Amax×Rmax, w ij are I.I.D. and ~ U([-1,1]), models transients and imperfections in the sensor grid –[] N maps the elements to the interval (0,1)

Repeated Games Given sensor network performance P(n), players optimizes next moves – similar to best response to expected cost Assumes some (limited?) mutual knowledge of P(n) and some estimation of opponent play history (?) Note: it is not clear from this paper how the estimates of opponent strategy or P(n) are made (in fact, the authors do not explicitly suggest that they are estimates – this is my inference) Any ideas?

Convergence to Nash Equilibrium Authors demonstrate that since N.E. exists for fixed P, it is sufficient for convergence to equilibrium with dynamic P to show convergence of P P converges for small positive  and  Both players have an incentive to vary strategies over time since, otherwise, the opponent of a player with an unchanging strategy will adapt to exploit weaknesses left open by the static strategy