Detection and Mitigation of Cyber-Attacks Using Game Theory and Learning
|
|
- Rudolph Maxwell
- 6 years ago
- Views:
Transcription
1 Detection and Mitigation of Cyber-Attacks Using Game Theory and Learning João P. Hespanha Kyriakos G. Vamvoudakis
2 Cyber Situation Awareness Framework Mission Cyber-Assets Simulation/Live Security Exercises Observations: Netflow, Probing, Time analysis Analysis to get up-to-date view of cyber-assets Analysis to determine dependencies between assets and missions Data Data Analyze and Characterize Attackers Data Mission Model Data Cyber-Assets Model Predict Future Actions COAs Sensor Alerts Correlation Engine Data Impact Analysis Data Create semantically-rich view of cyber-mission status
3 Outline... Large matrix games (summary of results) Multi-agent learning under cyber-attack using Q-learning (summary of results) Integration of online optimization for real-time attack prediction and visualization (summary of results) Observability of dynamical systems under attacks to sensors (summary and new results)
4 Outline... Large matrix games (summary of results) Multi-agent learning under cyber-attack using Q-learning (summary of results) Integration of online optimization for real-time attack prediction and visualization (summary of results) Observability of dynamical systems under attacks to sensors (summary and new results) Bopardikar- UTRC Prandini-Milano Poly. Milano
5 Network Security Games intrusion detection system chat software web proxy attack graph from [J. Wing 2007] intruder s target! sequence of intruder actions that compromise database server, not detected by IDS Even trivially small network security games can lead to games with very large decision trees Problem statistics of ictf 2010 over 7800 distinct mission states (defender observations) over 2500 distinct observations available to the attacker defender can choose among about distinct policies attacker can choose among distinct policies, depending on attacker's level of expertise
6 attack graph from [J. Wing 2007] Network Security Games intrusion detection system Developed sample-based approach to solving zero-sum games Approach provides probabilistic guarantees on the performance of the policies (in terms of security levels) Results applicable to very general classes of games that can include stochastic actions, partial information, etc. chat software web proxy intruder s target! sequence of intruder actions that compromise database server, not detected by IDS Even trivially small network security games can lead to games with very large decision trees Problem statistics of ictf 2010 over 7800 distinct mission states (defender observations) over 2500 distinct observations available to the attacker defender can choose among about distinct policies attacker can choose among distinct policies, depending on attacker's level of expertise
7 Application to ictf 2010 services S6, S2 We were able to Provide Cyber-security Litya office receives estimates avg. 314 of units mission for completion success of all 4 missions Provide dynamic rules to control firewall Take into account the effect of attacks & counter measures services S4, S5 Response can be a function of attacker sophistication Play what-if scenarios (vulnerabilities, information, etc.) services S3, S7 service S1 Increasing level of attacker sophistication services S2, S3 Level of attacker sophistication # units received by Litya for 1 round of missions [Option I, no bribes] services S0, S2 service S8 service S6 # units received by Litya for 1 round of missions [Option I, with bribes] service S9 # units received by Litya for 1 round of missions [Option II, with bribes] no service vulnerable (baseline) services S3, S9 S2 services S0, S1 service S2 (vulnerable to 38 teams) S2, S6, S9 (vulnerable to at least 6 team) S0, S2, S4, S6, S7, S8, S9 services S3, S8 service S1 service S0 (vulnerable to at least 1 team) all services vulnerable
8 Outline... Large matrix games (summary of results) Multi-agent learning under cyber-attack using Q-learning (summary of results) Integration of online optimization for real-time attack prediction and visualization (summary of results) Observability of dynamical systems under attacks to sensors (summary and new results)
9 Resilient Cyber-Mission Architectures In complex cyber missions, human operators define policies and rules computing elements automate processes of distributed resource allocation, scheduling, inventory management, etc. What is the impact of attacks on this type of automated/optimization process? Can we devise algorithms with built-in attack prediction/awareness capabilities?
10 Focus: Distributed Consensus/Agreement Classical problem in distributed computing: A group of computing elements must agree on a common scalar value x (e.g., priority, resources allocated, inventory decision, database value) Decision done iteratively & distributed using peer-to-peer communication 2 nd order adjustment rule x i pk ` 1q x i pkq` x i pk ` 1q value at processor i, iteration k x i pkq x i pkq`u i ` v i adjustment on x i by processor i, at iteration k Goal: minimize errors between values of agents and their neighbors e i ÿ xi x j x i x j peers of agent i jpn i (self-included) Attacker: maximize errors using stealth attacks (small v i ) correct update on adjustment update on adjustment by attacker
11 Focus: Distributed Consensus/Agreement Classical problem in distributed computing: A group of computing elements must agree on a common scalar value x (e.g., priority, resources allocated, inventory decision, database value) Decision done iteratively & distributed using peer-to-peer communication 2 nd order adjustment rule x i pk ` 1q x i pkq` x i pk ` 1q value at processor i, iteration k x i pkq x i pkq`u i ` v i correct update on adjustment update on adjustment by attacker adjustment on x i by processor i, at iteration k Nash equilibrium formulation: e i ÿ xi x j x i x j jpn i J i ÿ ˆ }e i } 2 ` ÿ }u 2 j } 2 2ij}v j } k jpn error min. by us max. by attacker our updates (small means smooth) min. by us max. by attacker attacker updates (small means stealth) max. by us min. by attacker
12 Optimal Solution Bellman Equation Optimal Control and Attacker Policies number of peers u i d i 0 I BV i Be i Under appropriate regularity assumptions (smoothness) J i pu i,u i,v q J i pu i,u i,v q J i pu, v i,v iq J i pu, v i,v i u i * is optimal (minimal) for us v i * is optimal (maximal) for attacker Moreover, Consensus will be reached asymptotically All variables will remain bounded through the transient (in fact, Lyapunov stability)
13 Optimal Solution Bellman Equation Optimal Control and Attacker Policies number of peers But Bellman equation difficult BV i to solve (curse of dimensionality) u i d i 0 I Be i Approach: Under Machine appropriate learning regularity based algorithm assumptions to solve (smoothness) this distributed consensus problem J i u i * is optimal (minimal) for us pu i,u i,v q J i pu i,u i,v Key contributions i JApplies i pu, v i,v iq to second-order J i pu, vupdates i,v iq (and@v even i more v i * is complex optimal (maximal) dynamics) for attacker Algorithms do not require global knowledge of the communication graph Algorithms do not require knowledge of the update rules used by other agents Moreover, Formal guarantees of correctness (convergence) Consensus will be reached asymptotically All variables will remain bounded through the transient (in fact, Lyapunov stability)
14 Actor/Critic Learning Approach Critic = Model free (distributed) algorithm to evaluate the current algorithm & estimate attacker actions Actor = Model free (distributed) algorithm to enact optimal decisions (based on critic s findings) Actor & Critic based on Approximate Dynamic Programming: Critic learns Q-function (action dependent) & Actor learns optimal control laws
15 Outline... Large matrix games (summary of results) Multi-agent learning under cyber-attack using Q-learning (summary of results) Integration of online optimization for real-time attack prediction and visualization (summary of results) Observability of dynamical systems under attacks to sensors (summary and new results)
16 Cyber Missions Complexity Challenges to real-time cyber-mission protection: cyber assets shared among missions cyber asset requirements change over time missions can use different configurations of resources complex network of cyber-asset dependencies Mission 1 services S0, S1 services S5, S6 service S9 services S0, S1 services S5, S6, S7 service S6 Attack on service S0 can result in multiple mission failure But, damage only realized if missions follows particular paths Mission 2 services S0, S2 service S2 service S ICTF Cyber Awareness Questions: When & where is an attacker most likely to strike? When & where is an attacker most damaging to mission completion? How will the answer depend on attacker resources? attacker skills? attacker knowledge? (real-time what-if analysis)
17 Mapping Service Attacks to Mission Damage Developed an optimization formalism to predict (most likely/damaging) attacks Damage equation: (for service s at time t) Uncertainty equation: (for service s at time t) PD s t «a s t ` b s t AR s t Potential damage probability of realizing damage attack resources p s t «r0,1s`cs t d s t AR s t attack resources equation parameters vary with time as mission progresses (learned from data in ictf exercises) Optimal attacks: maximize total damage to mission constrained by: ÿ ARt s TR s total attack resources at time t Formalism and predictions validated in ICTF 2011 exercise (89 teams, participants)
18 Multi-Resolution Visualization Enabling a multi-resolution attack analysis 1. High-level attack predictions based on online optimization 2. Potential damage & uncertainty associated with attacks to different services 3. Parameters that determine damage and uncertainty AlloSphere Integration High-level predictions permit fast action Low-level parameters permits investigating rationale for predictions
19 Outline... Large matrix games (summary of results) Multi-agent learning under cyber-attack using Q-learning (summary of results) Integration of online optimization for real-time attack prediction and visualization (summary of results) Observability of dynamical systems under attacks to sensors (summary and new results) Sinopoli-CMU Y. Mo-Caltech
20 Detection in Adversarial Environments How to interpret & access the reliability of sensors that have been manipulated? Sensors relevant to cyber missions? Measurement sensors (e.g., SCADA systems) Computational sensors (e.g., weather forecasting simulation engines) Data retrieval sensors (e.g., database queries) Cyber-security sensors (e.g., IDSs) Sensor-error Domains Deterministic sensors: with n sensors, one can get correct answer as long as m < n/2 sensors have been manipulated Stochastic sensors without manipulation: solution given by hypothesis testing/estimation Ppsensor errorq p err ñ Ppn sensor errorq «` n n{2 Stochastic sensors with potential manipulation: open problem? pn{2 err
21 Problem formulation X binary random variable to be estimated PpX 0q PpX 1q 1 2 for simplicity (papers treat general case) Y 1, Y 2,, Y n noisy measurements of X produced by n sensors PpY i Xq p err PpY i Xq 1 p per-sensor error probability (not necessarily very small) Z 1, Z 2,, Z n measurements actually reported by the n sensors # Y i sensor i not attacked Z i? sensor i attacked at most m sensors attacked p attack probability that we are under attack (very hard to know!) interpretation of sensor data should be mostly independent of p attack
22 Theorem: Result for small # of sensors (n<2/p err ) X binary random variable to be estimated Y 1, Y 2,, Y n noisy measurements of X produced by n sensors Z 1, Z 2,, Z n measurements # actually reported by the n sensors Y i sensor i not attacked Z i? sensor i attacked at most m sensors attacked p attack probability that we are under attack (very hard to know!) The optimal estimator is go with the majority of the (potentially manipulated) sensor readings # µ majority w.p. 1 y 2 µ no consensus w.p. y 2, y 2 # r0,1s go with the majority, EXCEPT if there is consensus p1 p err q ÿ ˆ n 1 2ÿ k m n 1 2ÿ k 0 p ˆn m p n k err p1 p err q k m k m ˆn m k 1 p attack p attack n m n m`perr q p n m k err p1 p err q k `p1 perr q n p n err p n m err 0 p n m err The optimal estimator is largely independent of p attack (hard to know)
23 Result for small # of sensors (n<2/p err ) X binary random variable to be estimated Y 1, Y 2,, Y n noisy measurements of X produced by n sensors Z 1, Z 2,, Z n measurements # actually reported by the n sensors Y i sensor i not attacked Z i? sensor i attacked What about the at estimation most m sensors of time-varying attacked variables: e.g., the state of a mission! p attack probability that we are under attack (very hard to know!) Previously Theorem: X constant binary random variable to be estimated The optimal estimator is Now go with the majority of the (potentially manipulated) sensor readings # X(t) µ majority time-varying w.p. state 1 y 2 variable to be estimated, # r0,1s n m y 2 p1 p based p err qn m`p on n m err err µ no consensus 1. sensor w.p. measurements y 2, that may 0 have been manipulated p n m 2. system dynamics E.g., the state of a cyber mission go with the majority, EXCEPT if there is consensus The optimal estimator is largely independent of p attack (hard to know) ÿ ˆ n 1 2ÿ k m n 1 2ÿ k 0 p ˆn m p n k err p1 p err q k m k m ˆn m k 1 p attack p attack q p n m k err p1 p err q k `p1 perr q n p n err err
24 Continuous Linear Systems dynamical evolution of systems s state ẋ = Ax + Bu control signals y i = C i x + D i u, i 2{1,...,N} ( y i sensor i not attacked z i =? sensor i attacked N measurements produced by sensor N measurements reported by sensor at most M sensors can be manipulated by the attackers Under what conditions can one reconstruct the state from (potentially corrupted) sensor measurements?
25 Continuous Linear Systems dynamical evolution of systems s state ẋ = Ax + Bu control signals y i = C i x + D i u, i 2{1,...,N} ( y i sensor i not attacked z i =? sensor i attacked N measurements produced by sensor N measurements reported by sensor at most M sensors can be manipulated by the attackers Theorem: Exact state reconstruction is possible if and only if system is observable through every subset of N - 2M measurements ê state could be reconstructed through only N - 2M measurements in the absence of attacks ê potential attack at M sensors, effectively disables 2M sensors
26 Estimation algorithms Gramian-based estimator: batch, finite-time estimation inversion of the observability matrix at each time step Observer-based estimator: asymptotic estimation recursive low-computation algorithm provably robust with respect to noise on all sensors (including non attacked ones) Algorithm outline: 1. Build an estimate removing by ignoring a set S of M sensors 2. Build additional estimates by removing, in addition, all combinations of M additional sensors 3. If all attacked sensors were in set S, then the estimates in steps 1. and 2. will be consistent (modulo noise) (all estimates can be constructed without combinatorial complexity, by using finite dimensionality)
27 Discrete Event Systems - Background alphabet: c Y u language: L Ä supervisor: f : L Ñ sets with (at least) all uncontrolled symbols language controller by supervisor: w P L f ô w P L f,w P L, P fpwq Theorem. There exists a supervisor f such that L f K i K is controllable K u X L Ä K observation map: P : Ñp o Yt uq observation alphabet P -supervisor: g : P plq Ñ language controller by P-supervisor: w P L g ô w P L g,w P L, P g`p pwq Theorem. There exists a P -supervisor g such that L g K i K is controllable and P -observable P pwq P pw 1 q ñ E P : w P K, w 1 P LzK or w P LzK, w 1 P K
28 Supervised DES under attacks language observation map P-supervisor language observation map P-supervisor attack A 1 attack A 2 attack A n Attack model: m out of n attacks active (which?) each attack symbol distortion, erasure, insertion nondeterminism A i : o Ñ 2 o
29 Supervised DES under attacks language observation map P-supervisor attack A 1 attack A 2 attack A n Attack model: m out of n attacks active (which?) each attack symbol distortion, erasure, insertion nondeterminism A i : o Ñ 2 o maximal language L max g,a controlled by P -supervisor g under the attack A: w P L max g,a ô w P L max g,a,w P L, Dy P AP pwq, P gpyq, set of observed symbols minimal language L max g,a w controlled by P -supervisor g under the attack A: P L max g,a ô w P L max g,a,w P P AP pwq, P gpyq, languages with largest/smallest number of word that attacker could enforce
30 Supervised DES under attacks language observation map P-supervisor attack A 1 attack A 2 attack A n Attack model: m out of n attacks active (which?) each attack symbol distortion, erasure, insertion nondeterminism A i : o Ñ 2 o Theorem. There exists a P -supervisor g such that L min g,a Lmax g,a K i K is controllable and P -observable for set of attacks A DA, A 1 P A AP pwqxa 1 P pw 1 q H ñ E P : w P K, w 1 P LzK or w P LzK, w 1 P K.
31 Output-symbol attacks language observation map P-supervisor attack A 1 attack A 2 attack A n Attack model: Each output symbol produced by one sensor One sensor could be manipulated (which?) Sensor manipulation: symbol erasure, insertion nondeterminism A i = arbitrary insertions/deletions of i-th output symbol Theorem. There exists a P -supervisor g such that L min g,a Lmax g,a K is controllable and P j P o K i remove from output string the symbols i & j potential attack at 1 sensor, effectively disables 2 sensor
32 Summary... Large matrix games (summary of results) Multi-agent learning under cyber-attack using Q-learning (summary of results) Integration of online optimization for real-time attack prediction and visualization (summary of results) Observability of dynamical systems under attacks to sensors (summary and new results)
Cyber Security Games with Asymmetric Information
Cyber Security Games with Asymmetric Information Jeff S. Shamma Georgia Institute of Technology Joint work with Georgios Kotsalis & Malachi Jones ARO MURI Annual Review November 15, 2012 Research Thrust:
More informationPrivacy and Fault-Tolerance in Distributed Optimization. Nitin Vaidya University of Illinois at Urbana-Champaign
Privacy and Fault-Tolerance in Distributed Optimization Nitin Vaidya University of Illinois at Urbana-Champaign Acknowledgements Shripad Gade Lili Su argmin x2x SX i=1 i f i (x) Applications g f i (x)
More informationCyber-Awareness and Games of Incomplete Information
Cyber-Awareness and Games of Incomplete Information Jeff S Shamma Georgia Institute of Technology ARO/MURI Annual Review August 23 24, 2010 Preview Game theoretic modeling formalisms Main issue: Information
More informationGrundlagen der Künstlichen Intelligenz
Grundlagen der Künstlichen Intelligenz Formal models of interaction Daniel Hennes 27.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Taxonomy of domains Models of
More informationcs/ee/ids 143 Communication Networks
cs/ee/ids 143 Communication Networks Chapter 5 Routing Text: Walrand & Parakh, 2010 Steven Low CMS, EE, Caltech Warning These notes are not self-contained, probably not understandable, unless you also
More informationOLSO. Online Learning and Stochastic Optimization. Yoram Singer August 10, Google Research
OLSO Online Learning and Stochastic Optimization Yoram Singer August 10, 2016 Google Research References Introduction to Online Convex Optimization, Elad Hazan, Princeton University Online Learning and
More informationADMM and Fast Gradient Methods for Distributed Optimization
ADMM and Fast Gradient Methods for Distributed Optimization João Xavier Instituto Sistemas e Robótica (ISR), Instituto Superior Técnico (IST) European Control Conference, ECC 13 July 16, 013 Joint work
More informationState Estimation of Linear and Nonlinear Dynamic Systems
State Estimation of Linear and Nonlinear Dynamic Systems Part I: Linear Systems with Gaussian Noise James B. Rawlings and Fernando V. Lima Department of Chemical and Biological Engineering University of
More informationModern Information Retrieval
Modern Information Retrieval Chapter 8 Text Classification Introduction A Characterization of Text Classification Unsupervised Algorithms Supervised Algorithms Feature Selection or Dimensionality Reduction
More informationMarkov Decision Processes Chapter 17. Mausam
Markov Decision Processes Chapter 17 Mausam Planning Agent Static vs. Dynamic Fully vs. Partially Observable Environment What action next? Deterministic vs. Stochastic Perfect vs. Noisy Instantaneous vs.
More informationA Reservoir Sampling Algorithm with Adaptive Estimation of Conditional Expectation
A Reservoir Sampling Algorithm with Adaptive Estimation of Conditional Expectation Vu Malbasa and Slobodan Vucetic Abstract Resource-constrained data mining introduces many constraints when learning from
More informationFinal Exam December 12, 2017
Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes
More informationAttack-Resilient Supervisory Control of Discrete-Event Systems
1 Attack-Resilient Supervisory Control of Discrete-Event Systems Yu Wang, Alper Kamil Bozkurt and Miroslav Pajic arxiv:194.3264v1 [cs.fl] 5 Apr 219 Abstract In this work, we study the problem of supervisory
More informationDynamic and Adversarial Reachavoid Symbolic Planning
Dynamic and Adversarial Reachavoid Symbolic Planning Laya Shamgah Advisor: Dr. Karimoddini July 21 st 2017 Thrust 1: Modeling, Analysis and Control of Large-scale Autonomous Vehicles (MACLAV) Sub-trust
More informationLoopy Belief Propagation for Bipartite Maximum Weight b-matching
Loopy Belief Propagation for Bipartite Maximum Weight b-matching Bert Huang and Tony Jebara Computer Science Department Columbia University New York, NY 10027 Outline 1. Bipartite Weighted b-matching 2.
More informationMARKOV DECISION PROCESSES (MDP) AND REINFORCEMENT LEARNING (RL) Versione originale delle slide fornita dal Prof. Francesco Lo Presti
1 MARKOV DECISION PROCESSES (MDP) AND REINFORCEMENT LEARNING (RL) Versione originale delle slide fornita dal Prof. Francesco Lo Presti Historical background 2 Original motivation: animal learning Early
More informationA Stochastic Framework for Quantitative Analysis of Attack-Defense Trees
1 / 35 A Stochastic Framework for Quantitative Analysis of R. Jhawar K. Lounis S. Mauw CSC/SnT University of Luxembourg Luxembourg Security and Trust of Software Systems, 2016 ADT2P & TREsPASS Project
More informationA Robust Event-Triggered Consensus Strategy for Linear Multi-Agent Systems with Uncertain Network Topology
A Robust Event-Triggered Consensus Strategy for Linear Multi-Agent Systems with Uncertain Network Topology Amir Amini, Amir Asif, Arash Mohammadi Electrical and Computer Engineering,, Montreal, Canada.
More informationRecent Developments in Statistical Dialogue Systems
Recent Developments in Statistical Dialogue Systems Steve Young Machine Intelligence Laboratory Information Engineering Division Cambridge University Engineering Department Cambridge, UK Contents Review
More informationAustralian National University WORKSHOP ON SYSTEMS AND CONTROL
Australian National University WORKSHOP ON SYSTEMS AND CONTROL Canberra, AU December 7, 2017 Australian National University WORKSHOP ON SYSTEMS AND CONTROL A Distributed Algorithm for Finding a Common
More informationLittlestone s Dimension and Online Learnability
Littlestone s Dimension and Online Learnability Shai Shalev-Shwartz Toyota Technological Institute at Chicago The Hebrew University Talk at UCSD workshop, February, 2009 Joint work with Shai Ben-David
More information5. Density evolution. Density evolution 5-1
5. Density evolution Density evolution 5-1 Probabilistic analysis of message passing algorithms variable nodes factor nodes x1 a x i x2 a(x i ; x j ; x k ) x3 b x4 consider factor graph model G = (V ;
More informationHidden Markov models 1
Hidden Markov models 1 Outline Time and uncertainty Markov process Hidden Markov models Inference: filtering, prediction, smoothing Most likely explanation: Viterbi 2 Time and uncertainty The world changes;
More informationFor general queries, contact
PART I INTRODUCTION LECTURE Noncooperative Games This lecture uses several examples to introduce the key principles of noncooperative game theory Elements of a Game Cooperative vs Noncooperative Games:
More informationA Control-Theoretic Perspective on the Design of Distributed Agreement Protocols, Part
9. A Control-Theoretic Perspective on the Design of Distributed Agreement Protocols, Part Sandip Roy Ali Saberi Kristin Herlugson Abstract This is the second of a two-part paper describing a control-theoretic
More informationOnline Learning, Mistake Bounds, Perceptron Algorithm
Online Learning, Mistake Bounds, Perceptron Algorithm 1 Online Learning So far the focus of the course has been on batch learning, where algorithms are presented with a sample of training data, from which
More informationMulti-Robotic Systems
CHAPTER 9 Multi-Robotic Systems The topic of multi-robotic systems is quite popular now. It is believed that such systems can have the following benefits: Improved performance ( winning by numbers ) Distributed
More informationPart I: Definitions and Properties
Turing Machines Part I: Definitions and Properties Finite State Automata Deterministic Automata (DFSA) M = {Q, Σ, δ, q 0, F} -- Σ = Symbols -- Q = States -- q 0 = Initial State -- F = Accepting States
More informationCryptographic Protocols Notes 2
ETH Zurich, Department of Computer Science SS 2018 Prof. Ueli Maurer Dr. Martin Hirt Chen-Da Liu Zhang Cryptographic Protocols Notes 2 Scribe: Sandro Coretti (modified by Chen-Da Liu Zhang) About the notes:
More informationFinal Exam December 12, 2017
Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes
More informationDP Project Development Pvt. Ltd.
Dear Sir/Madam, Greetings!!! Thanks for contacting DP Project Development for your training requirement. DP Project Development is leading professional training provider in GIS technologies and GIS application
More informationOnline Learning and Sequential Decision Making
Online Learning and Sequential Decision Making Emilie Kaufmann CNRS & CRIStAL, Inria SequeL, emilie.kaufmann@univ-lille.fr Research School, ENS Lyon, Novembre 12-13th 2018 Emilie Kaufmann Online Learning
More informationNon-Collision Conditions in Multi-agent Robots Formation using Local Potential Functions
2008 IEEE International Conference on Robotics and Automation Pasadena, CA, USA, May 19-23, 2008 Non-Collision Conditions in Multi-agent Robots Formation using Local Potential Functions E G Hernández-Martínez
More informationDistributed and Real-time Predictive Control
Distributed and Real-time Predictive Control Melanie Zeilinger Christian Conte (ETH) Alexander Domahidi (ETH) Ye Pu (EPFL) Colin Jones (EPFL) Challenges in modern control systems Power system: - Frequency
More informationProbabilistic Model Checking and Strategy Synthesis for Robot Navigation
Probabilistic Model Checking and Strategy Synthesis for Robot Navigation Dave Parker University of Birmingham (joint work with Bruno Lacerda, Nick Hawes) AIMS CDT, Oxford, May 2015 Overview Probabilistic
More informationAttack Graph Modeling and Generation
Attack Graph Modeling and Generation Ratnesh Kumar, Professor, IEEE Fellow Electrical and Computer Engineering, Iowa State University PhD Students: Mariam Ibrahim German Jordanian University Attack Graph:
More informationCognitive Cyber-Physical System
Cognitive Cyber-Physical System Physical to Cyber-Physical The emergence of non-trivial embedded sensor units, networked embedded systems and sensor/actuator networks has made possible the design and implementation
More informationFinal. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.
CS 188 Spring 2014 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationan efficient procedure for the decision problem. We illustrate this phenomenon for the Satisfiability problem.
1 More on NP In this set of lecture notes, we examine the class NP in more detail. We give a characterization of NP which justifies the guess and verify paradigm, and study the complexity of solving search
More informationConsensus-Based Distributed Optimization with Malicious Nodes
Consensus-Based Distributed Optimization with Malicious Nodes Shreyas Sundaram Bahman Gharesifard Abstract We investigate the vulnerabilities of consensusbased distributed optimization protocols to nodes
More informationArchitectures and Algorithms for Distributed Generation Control of Inertia-Less AC Microgrids
Architectures and Algorithms for Distributed Generation Control of Inertia-Less AC Microgrids Alejandro D. Domínguez-García Coordinated Science Laboratory Department of Electrical and Computer Engineering
More informationModule 09 Decentralized Networked Control Systems: Battling Time-Delays and Perturbations
Module 09 Decentralized Networked Control Systems: Battling Time-Delays and Perturbations Ahmad F. Taha EE 5243: Introduction to Cyber-Physical Systems Email: ahmad.taha@utsa.edu Webpage: http://engineering.utsa.edu/
More informationCMU Noncooperative games 4: Stackelberg games. Teacher: Ariel Procaccia
CMU 15-896 Noncooperative games 4: Stackelberg games Teacher: Ariel Procaccia A curious game Playing up is a dominant strategy for row player So column player would play left Therefore, (1,1) is the only
More informationDecoupling Coupled Constraints Through Utility Design
1 Decoupling Coupled Constraints Through Utility Design Na Li and Jason R. Marden Abstract The central goal in multiagent systems is to design local control laws for the individual agents to ensure that
More informationMining Classification Knowledge
Mining Classification Knowledge Remarks on NonSymbolic Methods JERZY STEFANOWSKI Institute of Computing Sciences, Poznań University of Technology COST Doctoral School, Troina 2008 Outline 1. Bayesian classification
More informationSLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada
SLAM Techniques and Algorithms Jack Collier Defence Research and Development Canada Recherche et développement pour la défense Canada Canada Goals What will we learn Gain an appreciation for what SLAM
More informationQuantifying Cyber Security for Networked Control Systems
Quantifying Cyber Security for Networked Control Systems Henrik Sandberg ACCESS Linnaeus Centre, KTH Royal Institute of Technology Joint work with: André Teixeira, György Dán, Karl H. Johansson (KTH) Kin
More informationAutomata and Languages
Automata and Languages Prof. Mohamed Hamada Software Engineering Lab. The University of Aizu Japan Nondeterministic Finite Automata with empty moves (-NFA) Definition A nondeterministic finite automaton
More informationR E A D : E S S E N T I A L S C R U M : A P R A C T I C A L G U I D E T O T H E M O S T P O P U L A R A G I L E P R O C E S S. C H.
R E A D : E S S E N T I A L S C R U M : A P R A C T I C A L G U I D E T O T H E M O S T P O P U L A R A G I L E P R O C E S S. C H. 5 S O F T W A R E E N G I N E E R I N G B Y S O M M E R V I L L E S E
More informationA Quantum Computing Approach to the Verification and Validation of Complex Cyber-Physical Systems
A Quantum Computing Approach to the Verification and Validation of Complex Cyber-Physical Systems Achieving Quality and Cost Control in the Development of Enormous Systems Safe and Secure Systems and Software
More informationCommunication constraints and latency in Networked Control Systems
Communication constraints and latency in Networked Control Systems João P. Hespanha Center for Control Engineering and Computation University of California Santa Barbara In collaboration with Antonio Ortega
More informationReinforcement Learning. Introduction
Reinforcement Learning Introduction Reinforcement Learning Agent interacts and learns from a stochastic environment Science of sequential decision making Many faces of reinforcement learning Optimal control
More informationarxiv: v1 [math.oc] 24 Oct 2014
In-Network Leader Selection for Acyclic Graphs Stacy Patterson arxiv:1410.6533v1 [math.oc] 24 Oct 2014 Abstract We study the problem of leader selection in leaderfollower multi-agent systems that are subject
More informationOn Agreement Problems with Gossip Algorithms in absence of common reference frames
On Agreement Problems with Gossip Algorithms in absence of common reference frames M. Franceschelli A. Gasparri mauro.franceschelli@diee.unica.it gasparri@dia.uniroma3.it Dipartimento di Ingegneria Elettrica
More informationSimple Counter-terrorism Decision
A Comparative Analysis of PRA and Intelligent Adversary Methods for Counterterrorism Risk Management Greg Parnell US Military Academy Jason R. W. Merrick Virginia Commonwealth University Simple Counter-terrorism
More informationEquilibrium solutions in the observable M/M/1 queue with overtaking
TEL-AVIV UNIVERSITY RAYMOND AND BEVERLY SACKLER FACULTY OF EXACT SCIENCES SCHOOL OF MATHEMATICAL SCIENCES, DEPARTMENT OF STATISTICS AND OPERATION RESEARCH Equilibrium solutions in the observable M/M/ queue
More informationGeneralized Lowness and Highness and Probabilistic Complexity Classes
Generalized Lowness and Highness and Probabilistic Complexity Classes Andrew Klapper University of Manitoba Abstract We introduce generalized notions of low and high complexity classes and study their
More informationPracticable Robust Markov Decision Processes
Practicable Robust Markov Decision Processes Huan Xu Department of Mechanical Engineering National University of Singapore Joint work with Shiau-Hong Lim (IBM), Shie Mannor (Techion), Ofir Mebel (Apple)
More informationOnline Learning Class 12, 20 March 2006 Andrea Caponnetto, Sanmay Das
Online Learning 9.520 Class 12, 20 March 2006 Andrea Caponnetto, Sanmay Das About this class Goal To introduce the general setting of online learning. To describe an online version of the RLS algorithm
More informationLecture 16: Perceptron and Exponential Weights Algorithm
EECS 598-005: Theoretical Foundations of Machine Learning Fall 2015 Lecture 16: Perceptron and Exponential Weights Algorithm Lecturer: Jacob Abernethy Scribes: Yue Wang, Editors: Weiqing Yu and Andrew
More informationProvable Security against Side-Channel Attacks
Provable Security against Side-Channel Attacks Matthieu Rivain matthieu.rivain@cryptoexperts.com MCrypt Seminar Aug. 11th 2014 Outline 1 Introduction 2 Modeling side-channel leakage 3 Achieving provable
More informationOn the Complexity of Mapping Pipelined Filtering Services on Heterogeneous Platforms
On the Complexity of Mapping Pipelined Filtering Services on Heterogeneous Platforms Anne Benoit, Fanny Dufossé and Yves Robert LIP, École Normale Supérieure de Lyon, France {Anne.Benoit Fanny.Dufosse
More informationHandout 1: Introduction to Dynamic Programming. 1 Dynamic Programming: Introduction and Examples
SEEM 3470: Dynamic Optimization and Applications 2013 14 Second Term Handout 1: Introduction to Dynamic Programming Instructor: Shiqian Ma January 6, 2014 Suggested Reading: Sections 1.1 1.5 of Chapter
More informationToday s s Lecture. Applicability of Neural Networks. Back-propagation. Review of Neural Networks. Lecture 20: Learning -4. Markov-Decision Processes
Today s s Lecture Lecture 20: Learning -4 Review of Neural Networks Markov-Decision Processes Victor Lesser CMPSCI 683 Fall 2004 Reinforcement learning 2 Back-propagation Applicability of Neural Networks
More informationTutorial: PART 1. Online Convex Optimization, A Game- Theoretic Approach to Learning.
Tutorial: PART 1 Online Convex Optimization, A Game- Theoretic Approach to Learning http://www.cs.princeton.edu/~ehazan/tutorial/tutorial.htm Elad Hazan Princeton University Satyen Kale Yahoo Research
More informationREINFORCEMENT LEARNING
REINFORCEMENT LEARNING Larry Page: Where s Google going next? DeepMind's DQN playing Breakout Contents Introduction to Reinforcement Learning Deep Q-Learning INTRODUCTION TO REINFORCEMENT LEARNING Contents
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More information6 Reinforcement Learning
6 Reinforcement Learning As discussed above, a basic form of supervised learning is function approximation, relating input vectors to output vectors, or, more generally, finding density functions p(y,
More informationA Data-driven Approach for Remaining Useful Life Prediction of Critical Components
GT S3 : Sûreté, Surveillance, Supervision Meeting GdR Modélisation, Analyse et Conduite des Systèmes Dynamiques (MACS) January 28 th, 2014 A Data-driven Approach for Remaining Useful Life Prediction of
More informationSecret Sharing CPT, Version 3
Secret Sharing CPT, 2006 Version 3 1 Introduction In all secure systems that use cryptography in practice, keys have to be protected by encryption under other keys when they are stored in a physically
More informationCS 380: ARTIFICIAL INTELLIGENCE MACHINE LEARNING. Santiago Ontañón
CS 380: ARTIFICIAL INTELLIGENCE MACHINE LEARNING Santiago Ontañón so367@drexel.edu Summary so far: Rational Agents Problem Solving Systematic Search: Uninformed Informed Local Search Adversarial Search
More informationMarkov Decision Processes Chapter 17. Mausam
Markov Decision Processes Chapter 17 Mausam Planning Agent Static vs. Dynamic Fully vs. Partially Observable Environment What action next? Deterministic vs. Stochastic Perfect vs. Noisy Instantaneous vs.
More informationMathematical Optimization Models and Applications
Mathematical Optimization Models and Applications Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 1, 2.1-2,
More informationInformation Aggregation in Networks
Information Aggregation in Networks MSc Thesis Submitted by: Mattan Lebovich Supervisor: Dr. Kobi Gal, Dr. Anat Lerner Open University Ra anana, Israel 1 Abstract This thesis studies information aggregation
More informationBalancing and Control of a Freely-Swinging Pendulum Using a Model-Free Reinforcement Learning Algorithm
Balancing and Control of a Freely-Swinging Pendulum Using a Model-Free Reinforcement Learning Algorithm Michail G. Lagoudakis Department of Computer Science Duke University Durham, NC 2778 mgl@cs.duke.edu
More informationA Polynomial-time Nash Equilibrium Algorithm for Repeated Games
A Polynomial-time Nash Equilibrium Algorithm for Repeated Games Michael L. Littman mlittman@cs.rutgers.edu Rutgers University Peter Stone pstone@cs.utexas.edu The University of Texas at Austin Main Result
More informationA practical introduction to active automata learning
A practical introduction to active automata learning Bernhard Steffen, Falk Howar, Maik Merten TU Dortmund SFM2011 Maik Merten, learning technology 1 Overview Motivation Introduction to active automata
More information3 Probabilistic Turing Machines
CS 252 - Probabilistic Turing Machines (Algorithms) Additional Reading 3 and Homework problems 3 Probabilistic Turing Machines When we talked about computability, we found that nondeterminism was sometimes
More informationGaussian Processes as Continuous-time Trajectory Representations: Applications in SLAM and Motion Planning
Gaussian Processes as Continuous-time Trajectory Representations: Applications in SLAM and Motion Planning Jing Dong jdong@gatech.edu 2017-06-20 License CC BY-NC-SA 3.0 Discrete time SLAM Downsides: Measurements
More information, and rewards and transition matrices as shown below:
CSE 50a. Assignment 7 Out: Tue Nov Due: Thu Dec Reading: Sutton & Barto, Chapters -. 7. Policy improvement Consider the Markov decision process (MDP) with two states s {0, }, two actions a {0, }, discount
More informationReinforcement Learning. Yishay Mansour Tel-Aviv University
Reinforcement Learning Yishay Mansour Tel-Aviv University 1 Reinforcement Learning: Course Information Classes: Wednesday Lecture 10-13 Yishay Mansour Recitations:14-15/15-16 Eliya Nachmani Adam Polyak
More informationDistributed Optimization. Song Chong EE, KAIST
Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links
More informationRole of Synchronized Measurements In Operation of Smart Grids
Role of Synchronized Measurements In Operation of Smart Grids Ali Abur Electrical and Computer Engineering Department Northeastern University Boston, Massachusetts Boston University CISE Seminar November
More informationUncertainty and Bayesian Networks
Uncertainty and Bayesian Networks Tutorial 3 Tutorial 3 1 Outline Uncertainty Probability Syntax and Semantics for Uncertainty Inference Independence and Bayes Rule Syntax and Semantics for Bayesian Networks
More informationPOLYNOMIAL SPACE QSAT. Games. Polynomial space cont d
T-79.5103 / Autumn 2008 Polynomial Space 1 T-79.5103 / Autumn 2008 Polynomial Space 3 POLYNOMIAL SPACE Polynomial space cont d Polynomial space-bounded computation has a variety of alternative characterizations
More informationDecentralized Stabilization of Heterogeneous Linear Multi-Agent Systems
1 Decentralized Stabilization of Heterogeneous Linear Multi-Agent Systems Mauro Franceschelli, Andrea Gasparri, Alessandro Giua, and Giovanni Ulivi Abstract In this paper the formation stabilization problem
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationReconnect 04 Introduction to Integer Programming
Sandia is a multiprogram laboratory operated by Sandia Corporation, a Lockheed Martin Company, Reconnect 04 Introduction to Integer Programming Cynthia Phillips, Sandia National Laboratories Integer programming
More informationMULTIPLE CHOICE QUESTIONS DECISION SCIENCE
MULTIPLE CHOICE QUESTIONS DECISION SCIENCE 1. Decision Science approach is a. Multi-disciplinary b. Scientific c. Intuitive 2. For analyzing a problem, decision-makers should study a. Its qualitative aspects
More informationLearning in Zero-Sum Team Markov Games using Factored Value Functions
Learning in Zero-Sum Team Markov Games using Factored Value Functions Michail G. Lagoudakis Department of Computer Science Duke University Durham, NC 27708 mgl@cs.duke.edu Ronald Parr Department of Computer
More informationChapter 2. Reductions and NP. 2.1 Reductions Continued The Satisfiability Problem (SAT) SAT 3SAT. CS 573: Algorithms, Fall 2013 August 29, 2013
Chapter 2 Reductions and NP CS 573: Algorithms, Fall 2013 August 29, 2013 2.1 Reductions Continued 2.1.1 The Satisfiability Problem SAT 2.1.1.1 Propositional Formulas Definition 2.1.1. Consider a set of
More informationLecture 4: State Estimation in Hidden Markov Models (cont.)
EE378A Statistical Signal Processing Lecture 4-04/13/2017 Lecture 4: State Estimation in Hidden Markov Models (cont.) Lecturer: Tsachy Weissman Scribe: David Wugofski In this lecture we build on previous
More informationArtificial Neural Networks
Artificial Neural Networks 鮑興國 Ph.D. National Taiwan University of Science and Technology Outline Perceptrons Gradient descent Multi-layer networks Backpropagation Hidden layer representations Examples
More informationGame Theoretic Learning in Distributed Control
Game Theoretic Learning in Distributed Control Jason R. Marden Jeff S. Shamma November 1, 2016 May 11, 2017 (revised) Abstract In distributed architecture control problems, there is a collection of interconnected
More informationFINM6900 Finance Theory Noisy Rational Expectations Equilibrium for Multiple Risky Assets
FINM69 Finance Theory Noisy Rational Expectations Equilibrium for Multiple Risky Assets February 3, 212 Reference Anat R. Admati, A Noisy Rational Expectations Equilibrium for Multi-Asset Securities Markets,
More informationMining Classification Knowledge
Mining Classification Knowledge Remarks on NonSymbolic Methods JERZY STEFANOWSKI Institute of Computing Sciences, Poznań University of Technology SE lecture revision 2013 Outline 1. Bayesian classification
More informationLecture 11: Proofs, Games, and Alternation
IAS/PCMI Summer Session 2000 Clay Mathematics Undergraduate Program Basic Course on Computational Complexity Lecture 11: Proofs, Games, and Alternation David Mix Barrington and Alexis Maciel July 31, 2000
More informationCOMPUTER SCIENCE TRIPOS
CST.2016.6.1 COMPUTER SCIENCE TRIPOS Part IB Thursday 2 June 2016 1.30 to 4.30 COMPUTER SCIENCE Paper 6 Answer five questions. Submit the answers in five separate bundles, each with its own cover sheet.
More informationGeostatistics and Spatial Scales
Geostatistics and Spatial Scales Semivariance & semi-variograms Scale dependence & independence Ranges of spatial scales Variable dependent Fractal dimension GIS implications Spatial Modeling Spatial Analysis
More information