A Learning-Based Approach to Reactive Security * Ben Rubinstein Microsoft Research Silicon Valley With: Adam Barth 1, Mukund Sundararajan 2, John Mitchell.

Slides:



Advertisements
Similar presentations
Routing Complexity of Faulty Networks Omer Angel Itai Benjamini Eran Ofek Udi Wieder The Weizmann Institute of Science.
Advertisements

Routing and Congestion Problems in General Networks Presented by Jun Zou CAS 744.
Risk Models and Controlled Mitigation of IT Security R. Ann Miura-Ko Stanford University February 27, 2009.
Online Learning for Online Pricing Problems Maria Florina Balcan.
1 A Graph-Theoretic Network Security Game M. Mavronicolas , V. Papadopoulou , A. Philippou  and P. Spirakis § University of Cyprus, Cyprus  University.
Markov Decision Process
Lecturer: Moni Naor Algorithmic Game Theory Uri Feige Robi Krauthgamer Moni Naor Lecture 8: Regret Minimization.
An Introduction to... Evolutionary Game Theory
Markov Game Analysis for Attack and Defense of Power Networks Chris Y. T. Ma, David K. Y. Yau, Xin Lou, and Nageswara S. V. Rao.
Markov Game Analysis for Attack and Defense of Power Networks Chris Y. T. Ma, David K. Y. Yau, Xin Lou, and Nageswara S. V. Rao.
Regret Minimizing Audits: A Learning-theoretic Basis for Privacy Protection Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha Carnegie Mellon.
Online learning, minimizing regret, and combining expert advice
Extraction and Transfer of Knowledge in Reinforcement Learning A.LAZARIC Inria “30 minutes de Science” Seminars SequeL Inria Lille – Nord Europe December.
Seminar In Game Theory Algorithms, TAU, Agenda  Introduction  Computational Complexity  Incentive Compatible Mechanism  LP Relaxation & Walrasian.
Maximizing the Spread of Influence through a Social Network
Cryptography and Data Security: Long-Term Challenges Burt Kaliski, RSA Security Northeastern University CCIS Mini Symposium on Information Security November.
Class Project Due at end of finals week Essentially anything you want, so long as it’s AI related and I approve Any programming language you want In pairs.
Online Vertex-Coloring Games in Random Graphs Revisited Reto Spöhel (joint work with Torsten Mütze and Thomas Rast; appeared at SODA ’11)
Proactive Learning: Cost- Sensitive Active Learning with Multiple Imperfect Oracles Pinar Donmez and Jaime Carbonell Pinar Donmez and Jaime Carbonell Language.
Research Direction Introduction Advisor : Frank, Y.S. Lin Presented by Yu Pu Wu.
Advisor: Yeong-Sung Lin Presented by I-Ju Shih 2011/10/25 1 Research Direction Introduction.
Regret Minimizing Audits: A Learning-theoretic Basis for Privacy Protection Jeremiah Blocki, Nicolas Christin, Anupam Datta, Arunesh Sinha Carnegie Mellon.
Defending Complex System Against External Impacts Gregory Levitin (IEC, UESTC)
1 Computing Nash Equilibrium Presenter: Yishay Mansour.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Online Algorithms for Network Design Adam Meyerson UCLA.
AWESOME: A General Multiagent Learning Algorithm that Converges in Self- Play and Learns a Best Response Against Stationary Opponents Vincent Conitzer.
Achieving End-to-End Fairness in Wireless Networks Ananth Rao Ion Stoica OASIS Retreat, Jul 2005.
Query Incentive Networks Jon Kleinberg and Prabhakar Raghavan - Presented by: Nishith Pathak.
Dynamic Network Security Deployment under Partial Information George Theodorakopoulos (EPFL) John S. Baras (UMD) Jean-Yves Le Boudec (EPFL) September 24,
A Scalable Network Resource Allocation Mechanism With Bounded Efficiency Loss IEEE Journal on Selected Areas in Communications, 2006 Johari, R., Tsitsiklis,
Job Scheduling Lecture 19: March 19. Job Scheduling: Unrelated Multiple Machines There are n jobs, each job has: a processing time p(i,j) (the time to.
Finding an Unpredictable Target in a Workspace with Obstacles LaValle, Lin, Guibas, Latombe, and Motwani, 1997 CS326 Presentation by David Black-Schaffer.
Game-based Analysis of Denial-of- Service Prevention Protocols Ajay Mahimkar Class Project: CS 395T.
Commitment without Regrets: Online Learning in Stackelberg Security Games Nika Haghtalab Carnegie Mellon University Joint work with Maria-Florina Balcan,
1 Network Creation Game A. Fabrikant, A. Luthra, E. Maneva, C. H. Papadimitriou, and S. Shenker, PODC 2003 (Part of the Slides are taken from Alex Fabrikant’s.
Self-Adaptive QoS Guarantees and Optimization in Clouds Jim (Zhanwen) Li (Carleton University) Murray Woodside (Carleton University) John Chinneck (Carleton.
Utility Theory & MDPs Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart.
Online Oblivious Routing Nikhil Bansal, Avrim Blum, Shuchi Chawla & Adam Meyerson Carnegie Mellon University 6/7/2003.
Preserving Link Privacy in Social Network Based Systems Prateek Mittal University of California, Berkeley Charalampos Papamanthou.
The Power of the Defender M. Gelastou  M. Mavronicolas  V. Papadopoulou  A. Philippou  P. Spirakis §  University of Cyprus, Cyprus § University of.
5. Alternative Approaches. Strategic Bahavior in Business and Econ 1. Introduction 2. Individual Decision Making 3. Basic Topics in Game Theory 4. The.
Reputations Based On Transitive Trust Slides by Josh Albrecht.
Resisting Denial-of-Service Attacks Using Overlay Networks Ju Wang Advisor: Andrew A. Chien Department of Computer Science and Engineering, University.
Algorithms  Al-Khwarizmi, arab mathematician, 8 th century  Wrote a book: al-kitab… from which the word Algebra comes  Oldest algorithm: Euclidian algorithm.
Peter van Emde Boas: Games and Computer Science 1999 GAMES AND COMPUTER SCIENCE Theoretical Models 1999 Peter van Emde Boas References available at:
Maximization of Network Survivability against Intelligent and Malicious Attacks (Cont’d) Presented by Erion Lin.
On a Network Creation Game PoA Seminar Presenting: Oren Gilon Based on an article by Fabrikant et al 1.
Online Algorithms By: Sean Keith. An online algorithm is an algorithm that receives its input over time, where knowledge of the entire input is not available.
1 The Price of Defense M. Mavronicolas , V. Papadopoulou , L. Michael ¥, A. Philippou , P. Spirakis § University of Cyprus, Cyprus  University of Patras.
Games. Adversaries Consider the process of reasoning when an adversary is trying to defeat our efforts In game playing situations one searches down the.
Decapitation of networks with and without weights and direction : The economics of iterated attack and defense Advisor : Professor Frank Y. S. Lin Presented.
Jennifer Campbell November 30,  Problem Statement and Motivation  Analysis of previous work  Simple - competitive strategy  Near optimal deterministic.
Online Social Networks and Media
Adversarial Search Chapter Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time limits.
Information Theory for Mobile Ad-Hoc Networks (ITMANET): The FLoWS Project Competitive Scheduling in Wireless Networks with Correlated Channel State Ozan.
Ken Brumfield | Premier Field Engineer Ward Ralston| Group Product Manager Microsoft Corporation.
© 2009 Ilya O. Ryzhov 1 © 2008 Warren B. Powell 1. Optimal Learning On A Graph INFORMS Annual Meeting October 11, 2009 Ilya O. Ryzhov Warren Powell Princeton.
Grid Defense Against Malicious Cascading Failure Paulo Shakarian, Hansheng Lei Dept. Electrical Engineering and Computer Science, Network Science Center,
Slide 1/20 Defending Against Strategic Adversaries in Dynamic Pricing Markets for Smart Grids Paul Wood, Saurabh Bagchi Purdue University
Advisor: Yeong-Sung Lin Presented by I-Ju Shih 2011/11/29 1 Research Direction Introduction.
Automating Cyber- Defense Management By: Zach Archer COSC 316.
On a Network Creation Game
COMPUTATIONAL MODELS.
Firewall Configuration and Administration
When Security Games Go Green
Professor Arne Thesen, University of Wisconsin-Madison
CASE − Cognitive Agents for Social Environments
Adviser: Frank,Yeong-Sung Lin Present by Ying-Ju Chen
Multiagent Systems Repeated Games © Manfred Huber 2018.
Presentation transcript:

A Learning-Based Approach to Reactive Security * Ben Rubinstein Microsoft Research Silicon Valley With: Adam Barth 1, Mukund Sundararajan 2, John Mitchell 3, Dawn Song 1, Peter Bartlett 1 1 UC Berkeley 2 Google 3 Stanford * Appeared at Financial Crypto. & Data Security 2010

Proactive vs. Reactive Security What's important is to understand the delineation between what’s considered “acceptable” and “unacceptable” spending. The goal is to prevent spending on reactive security “firefighting”. – John N. Stewart, VP (CSO), Cisco Systems Conventional wisdom for CISOs – Adopt forward-looking, proactive, approach to managing security risks – Reactive security is akin to myopic bug chasing TRUST Conference F'10Reactive Security2

Strategic Reactive Security Good reactive security – Should be strategic and not “firefighting” – Under certain conditions keeps up with or beats proactive approaches – Machine Learning & Economics can help TRUST Conference F'10Reactive Security3

Focus on Truly Adversarial Attacker No probabilistic assumptions on attacker Allow attacker to be omniscient Consider reactive defender with limited knowledge of – System vulnerabilities – Attacker’s incentives – Attacker’s rationality TRUST Conference F'10Reactive Security4

Focus on Incentives We model attacker cost and payoff, combined as – additive profit; or multiplicative ROA TRUST Conference F'10Reactive Security5 An effective defense need not be perfect–but it should reduce attacker’s utility relative to attacking other systems.

Results in a Nutshell If… – Security budget is fungible – Attack costs linear in defense allocation – No catastrophic attacks to defender Attacker’s utility against reactive defense approaches utility under fixed proactive In many cases reactive is much better TRUST Conference F'10Reactive Security6

Formal Model: Attack Graph System as directed graph – Nodes: states – Edges: state transitions Attacks are paths Examples – Compromised machines connected by a network – Components in a complex software system – Internet fraud “battlefield” TRUST Conference F'10Reactive Security7 Peering Points Gateway Application Servers Database Servers Internet

Formal Model: Iterated Game Fixed properties of graph – Node v’s reward r(v)≥0 – Edge e’s attack surface w(e) Repeated game – Defender allocates total budget B, with d t (e) to edge e – Attacker launches attack a t – Attacker pays and receives Attacker sees defense prior to attack Defender sees edges/weights only once attacked TRUST Conference F'10Reactive Security8 Attack surface Defense allocation

Proactive Defender(s) Pro’s of analysis: includes defenders who – Have perfect knowledge of the entire graph – Have perfect knowledge of the attacks – Play rationally given in/complete information Con’s of analysis – We (mostly) assume proactive plays fixed strategy TRUST Conference F'10Reactive Security9

Strategic Reactive Defender Based on Multiplicative Weights algorithm of Online Learning Theory Unseen edges get no allocation Budget is increased on attacked edges Allocation due to “the past” is exponentially down-weighed since 0<β<1 TRUST Conference F'10Reactive Security10 All edges initially unseen Observe attacked edges Count #times edge attacked Multiplicative update Re-normalize in [0,1]; allocate this times budget B

Main Theorems Attacker’s utility – Profit = Payoff – Cost – ROA = (Total Payoff) / (Total Cost) Compared to any proactive strategy d *, the reactive strategy achieves – – for any α TRUST Conference F'10Reactive Security11

Robustness & Extensions Robustness – Proactive not robust to uncertainty in attacker’s utility; reactive is!! – Reactive can do much better under uncertain payoffs Extensions – Hypergraphs / Datalog – Multiple attackers – Adaptive proactive defenders TRUST Conference F'10Reactive Security12

Conclusions Incentives-based, fully-adversarial risk model Learning-based defender performs close to or better than fixed proactive defenders Recommendations for CISOs – Employ monitoring tools to help focus on real attacks – Make security organization more agile – Avoid overreacting to the most recent attack; consider past attacks (down-weighed exponentially) TRUST Conference F'10Reactive Security13

Thanks!!

Model Case Studies Perimeter defense – Non-zero reward at one vertex – Rational attacker will select minimum-cost path from start to reward – Rational defense is to maximize minimum-cost path: allocate budget to minimum-cut TRUST Conference F'10Reactive Security15

Model Case Studies Defense in Depth – Allocate budget evenly to edges – ROA = 1 TRUST Conference F'10Reactive Security16

Proof Sketch Profit when edges are known – Simple reduction to standard regret bound of Freund- Schapire for Multiplicative Update alg Profit under hidden edges – Simulation argument shows that a slight modification to MultUp produces same allocations as MultUp on observed graph – Care taken with – Algorithms’ profits bounded by ROA under hidden edges – Ratio of two numbers is small if numbers are large & similar. Need: TRUST Conference F'10Reactive Security17

Lower Bound TRUST Conference F'10Reactive Security18 s r:1 w:1 Budget=1

Learning Rewards Consider star configuration with unknown rewards Proactive defense – Allocates budget equally – Competitive ratio for ROA is #{leaf vertices} Reactive defense – Learns the rewards TRUST Conference F'10Reactive Security19

Robustness to Objective Given defense budget of 9 Proactive defender assuming profit-seeking – Allocates 9 to right-hand edge: 1 profit for all attacks – ROA for left-hand edge is infinite!! Reactive defender’s play is invariant TRUST Conference F'10Reactive Security20