COMP 2208 Dr. Long Tran-Thanh University of Southampton Revision.

Slides:



Advertisements
Similar presentations
Decision Theory: Sequential Decisions Computer Science cpsc322, Lecture 34 (Textbook Chpt 9.3) Nov, 28, 2012.
Advertisements

1 Image Classification MSc Image Processing Assignment March 2003.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Data Mining Classification: Alternative Techniques
Intelligent Environments1 Computer Science and Engineering University of Texas at Arlington.
Ai in game programming it university of copenhagen Reinforcement Learning [Outro] Marco Loog.
Supervised Learning Recap
. Markov Chains as a Learning Tool. 2 Weather: raining today40% rain tomorrow 60% no rain tomorrow not raining today20% rain tomorrow 80% no rain tomorrow.
CPSC 322, Lecture 37Slide 1 Finish Markov Decision Processes Last Class Computer Science cpsc322, Lecture 37 (Textbook Chpt 9.5) April, 8, 2009.
Lecture 17: Supervised Learning Recap Machine Learning April 6, 2010.
CII504 Intelligent Engine © 2005 Irfan Subakti Department of Informatics Institute Technology of Sepuluh Nopember Surabaya - Indonesia.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
Reinforcement Learning
CS 590M Fall 2001: Security Issues in Data Mining Lecture 3: Classification.
Learning From Observations
COMP 328: Midterm Review Spring 2010 Nevin L. Zhang Department of Computer Science & Engineering The Hong Kong University of Science & Technology
Bayesian Reinforcement Learning with Gaussian Processes Huanren Zhang Electrical and Computer Engineering Purdue University.
Reinforcement Learning Mitchell, Ch. 13 (see also Barto & Sutton book on-line)
Representing Uncertainty CSE 473. © Daniel S. Weld 2 Many Techniques Developed Fuzzy Logic Certainty Factors Non-monotonic logic Probability Only one.
5/25/2005EE562 EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS Lecture 16, 6/1/2005 University of Washington, Department of Electrical Engineering Spring 2005.
Learning Programs Danielle and Joseph Bennett (and Lorelei) 4 December 2007.
Making Decisions CSE 592 Winter 2003 Henry Kautz.
Machine learning Image source:
Quiz 4: Mean: 7.0/8.0 (= 88%) Median: 7.5/8.0 (= 94%)
INTRODUCTION TO MACHINE LEARNING. $1,000,000 Machine Learning  Learn models from data  Three main types of learning :  Supervised learning  Unsupervised.
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Upper Confidence Trees for Game AI Chahine Koleejan.
Multi-Layer Perceptrons Michael J. Watts
Machine Learning Lecture 11 Summary G53MLE | Machine Learning | Dr Guoping Qiu1.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Artificial Intelligence Lecture No. 29 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Artificial Intelligence Techniques Multilayer Perceptrons.
Learning from Observations Chapter 18 Through
CPS 270: Artificial Intelligence Machine learning Instructor: Vincent Conitzer.
Non-Bayes classifiers. Linear discriminants, neural networks.
CHAPTER 8 DISCRIMINATIVE CLASSIFIERS HIDDEN MARKOV MODELS.
1 Chapter 10 Introduction to Machine Learning. 2 Chapter 10 Contents (1) l Training l Rote Learning l Concept Learning l Hypotheses l General to Specific.
COMP 2208 Dr. Long Tran-Thanh University of Southampton Bayes’ Theorem, Bayesian Reasoning, and Bayesian Networks.
COMP 2208 Dr. Long Tran-Thanh University of Southampton Neural Networks.
COMP 2208 Dr. Long Tran-Thanh University of Southampton K-Nearest Neighbour.
COMP 2208 Dr. Long Tran-Thanh University of Southampton Bandits.
COMP 2208 Dr. Long Tran-Thanh University of Southampton Decision Trees.
1 ECE 517: Reinforcement Learning in Artificial Intelligence Lecture 21: Dynamic Multi-Criteria RL problems Dr. Itamar Arel College of Engineering Department.
COMP 2208 Dr. Long Tran-Thanh University of Southampton Reinforcement Learning.
WHAT IS DATA MINING?  The process of automatically extracting useful information from large amounts of data.  Uses traditional data analysis techniques.
1 Learning Bias & Clustering Louis Oliphant CS based on slides by Burr H. Settles.
WHAT IS DATA MINING?  The process of automatically extracting useful information from large amounts of data.  Uses traditional data analysis techniques.
Reinforcement Learning Guest Lecturer: Chengxiang Zhai Machine Learning December 6, 2001.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Chapter 12. Probability Reasoning Fall 2013 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
DATA MINING and VISUALIZATION Instructor: Dr. Matthew Iklé, Adams State University Remote Instructor: Dr. Hong Liu, Embry-Riddle Aeronautical University.
Brief Intro to Machine Learning CS539
A Brief Introduction to Bayesian networks
Learning with Perceptrons and Neural Networks
Presented By S.Yamuna AP/CSE
Reinforcement Learning
Data Mining Lecture 11.
Basic Intro Tutorial on Machine Learning and Data Mining
A task of induction to find patterns
Memory-Based Learning Instance-Based Learning K-Nearest Neighbor
A task of induction to find patterns
Chapter 14 February 26, 2004.
Instructor: Vincent Conitzer
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 7
Presentation transcript:

COMP 2208 Dr. Long Tran-Thanh University of Southampton Revision

Some info about the exam 4 questions, 120 mins: you only need to answer to 3 (40mins/each on avg) 2 from Richard’s part, 2 from mine 1 = set of short questions 1 = more complex question Today’s goal: revise the materials of my part Lecture summary points: to help with short questions Worked examples (illustrative) Q & A

The concept of learning agents Environment Perception Behaviour Classification Update belief model Update decision making policy Decision making Perception Behaviour

Categories of learning Online – offline Supervised – unsupervised (- semisupervised) Lazy learning (see kNN) Reinforcement learning (see MDP) – kind of unsupervised

Classification 1: neural networks What is a perceptron? What is classification? What is linear regression? When is the data space linearly separable? Definition of activation functions Perceptron learning rule (high level explanation only) Expressiveness/limitations of perceptrons (what they can/can’t do) What is a multi-layered neural network? What does it mean that some neurons are hidden? Back-propagation (high level explanation only)

Classification 2: decision trees Occam’s razor How to build a decision tree: which attribute to choose first, when to stop? Entropy, conditional entropy, information gain Advantages of decision trees

Classification 3: k-nearest neighbour Generalisation power of classification algorithms Overfitting Training data vs. testing data Cross validation K-NN: how does it work? How to set the value of K? How to measure the distance?

Reasoning: the Bayesian approach Types of reasoning Bayes’ theorem Belief update with Bayes’ theorem Inference with joint distribution: advantages and issues Bayesian inference: Bayesian networks How to build Bayes nets? Properties of Bayes nets Worked example: a complex Bayesian inference problem

Decision making with bandits What is sequential decision making under uncertainty? What is the dilemma of exploration vs. exploitation? What does it mean that we need to find the trade-off between exploration and exploitation? The multi-armed bandit model Epsilon-first, epsilon-greedy Some applications and extensions

Reinforcement learning What is reinforcement learning What are the difficulties? States, actions, rewards Temporal difference (TD) learning Q-learning What is a Markov decision process? How to update the values in MDPs? Monte Carlo simulation Which actions we should take? – link back to bandits Some applications + extensions

Collaborative AI What is the aim of the classical AI (artificial general intelligence – AGI): build humanoid AI What is collaborative AI? 4 requirements of collaborative AI: Flexible autonomy Agile teaming Incentive engineering Accountable information

Bayesian inference example: GoT Q: will Jon survive? Let’s find out using Bayesian inference! Rumors say:

Bayesian inference example: GoT Bayesian network: W: the wound of Jon is lethal M: Melissandre is willing to help healing Jon J: Jon will not die We want to know P(J) = probability that Jon will not die

Bayesian inference example: GoT

Q1: what is the probability that Milessandre will help healing Jon? P(M) = ? Answer: P(M) = P(M|W = true)P(W=true) + P(M|W = false) P(W=false) P(M) = 0.2 * * 0.2 = 0.3

Bayesian inference example: GoT Q2: Let’s do a Monte Carlo simulation in this network We generate random states by taking a random value from zero to one inclusive Set a given state to True if the random value is less than or equal to the relevant probability of that state being True We need three such random values to generate one set of state values for the network Take three random values and use them in sequence of W, M, and J to generate a single random state of the network The 3 values are: [ 0.1, 0.7, 0.55 ]

Bayesian inference example: GoT Q2: Monte Carlo simulation with [ 0.1, 0.7, 0.55 ], order = W, M, J 1 st value = 0.1 W = True 2 nd value = 0.7 (W = True) 0.7 > 0.2 -> M = False 3 rd value = 0.55 (W = True, M = False) 0.55 > > J = False State values: W= T, M = F, J = F

Bayesian inference example: GoT Q3: Reconstruct the full joint distribution of W, M, and J Idea: use the truth table + new column representing the probability WMJProbability

Bayesian inference example: GoT Q3: Reconstruct the full joint distribution of W, M, and J How to fill the rows? Answer: use chain rule multiple times!!! WMJProbability P(W=1,M=1,J=1) = P(J=1|M=1,W=1)P(M=1,W=1) =P(J=1|M=1,W=1)P(M=1|W=1)P(W=1)

Bayesian inference example: GoT Q3: Reconstruct the full joint distribution of W, M, and J How to fill the rows? Answer: use chain rule multiple times!!! WMJProbability P(W=1,M=1,J=1) =P(J=1|M=1,W=1)P(M=1|W=1)P(W=1) = 0.5 * 0.2 * 0.8 =

Bayesian inference example: GoT Q3: Reconstruct the full joint distribution of W, M, and J WMJProbability = 0.5*0.2*0.8 = 0.01*0.8*0.8 = 0.99*0.8*0.8 = 0.95*0.7*0.2 = 0.05*0.7*0.2 = 0.7*0.3*0.2 = 0.3*0.3*0.2

Bayesian inference example: GoT Q4: What is the probability that Melissandre indeed helped Jon given that Jon did not survive? That is, p(M=True | J=False)? Idea: use Bayes’ theorem P(M=True| J = False) = P(J = False | M = True)*P(M=True)/P(J = False) P(J = False | M = True) = ? P(M=True) = ? P(J = False) = ?

Bayesian inference example: GoT Q4: What is the probability that Melissandre indeed helped Jon given that Jon did not survive? That is, p(M=True | J=False)? Idea 2: use the truth table P(J = False | M = True) = P(J = F, M = T)/P(M=T) WMJProbabili ty P(J = F, M = T) = sum of rows where J = F and M = T = = P(M = T) = sum of rows where M = T = = 0.3 = 0.087/0.29 = 0.29

Bayesian inference example: GoT Q4: What is the probability that Melissandre indeed helped Jon given that Jon did not survive? That is, p(M=True | J=False)? WMJProbabili ty P(M=True) = = 0.3

Bayesian inference example: GoT Q4: What is the probability that Melissandre indeed helped Jon given that Jon did not survive? That is, p(M=True | J=False)? WMJProbabili ty P(J=False) = =

Bayesian inference example: GoT Q4: What is the probability that Melissandre indeed helped Jon given that Jon did not survive? That is, p(M=True | J=False)? P(M=True| J = False) = P(J = False | M = True)*P(M=True)/P(J = False) P(J = False | M = True) = 0.29 P(M=True) = 0.3 P(J = False) = P(M=True| J = False) = 0.29*0.3/ =

Revision Q & A