1 ECE 517 Reinforcement Learning in Artificial Intelligence Lecture 21: Deep Machine Learning Dr. Itamar Arel College of Engineering Department of Electrical.

Slides:



Advertisements
Similar presentations
The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS Zemel
Advertisements

Sparse Coding in Sparse Winner networks Janusz A. Starzyk 1, Yinyin Liu 1, David Vogel 2 1 School of Electrical Engineering & Computer Science Ohio University,
A model of Consciousness With neural networks By: Hadiseh Nowparast.
Un Supervised Learning & Self Organizing Maps Learning From Examples
EE141 1 Broca’s area Pars opercularis Motor cortexSomatosensory cortex Sensory associative cortex Primary Auditory cortex Wernicke’s area Visual associative.
Tracking multiple independent targets: Evidence for a parallel tracking mechanism Zenon Pylyshyn and Ron Storm presented by Nick Howe.
Learning Sensorimotor Contingencies James J. Clark Centre for Intelligent Machines McGill University.
Instar Learning Law Adapted from lecture notes of the course CN510: Cognitive and Neural Modeling offered in the Department of Cognitive and Neural Systems.
IT 691 Final Presentation Pace University Created by: Robert M Gust Mark Lee Samir Hessami Mark Lee Samir Hessami.
Neural Networks (NN) Ahmad Rawashdieh Sa’ad Haddad.
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Last Words COSC Big Data (frameworks and environments to analyze big datasets) has become a hot topic; it is a mixture of data analysis, data mining,
Chapter 11: Artificial Intelligence
MIND: The Cognitive Side of Mind and Brain  “… the mind is not the brain, but what the brain does…” (Pinker, 1997)
Artificial Neural Network Theory and Application Ashish Venugopal Sriram Gollapalli Ulas Bardak.
Artificial Neural Nets and AI Connectionism Sub symbolic reasoning.
Introduction to Neural Networks. Neural Networks in the Brain Human brain “computes” in an entirely different way from conventional digital computers.
An Introduction to Artificial Intelligence and Knowledge Engineering N. Kasabov, Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering,
CS654: Digital Image Analysis Lecture 3: Data Structure for Image Analysis.
Self organizing maps 1 iCSC2014, Juan López González, University of Oviedo Self organizing maps A visualization technique with data dimension reduction.
Advances in Modeling Neocortex and its impact on machine intelligence Jeff Hawkins Numenta Inc. VS265 Neural Computation December 2, 2010 Documentation.
1 ECE-517: Reinforcement Learning in Artificial Intelligence Lecture 6: Optimality Criterion in MDPs Dr. Itamar Arel College of Engineering Department.
Machine Learning Dr. Shazzad Hosain Department of EECS North South Universtiy
Video Tracking Using Learned Hierarchical Features
1 ECE-517 Reinforcement Learning in Artificial Intelligence Lecture 7: Finite Horizon MDPs, Dynamic Programming Dr. Itamar Arel College of Engineering.
Hierarchical Temporal Memory as a Means for Image Recognition by Wesley Bruning CHEM/CSE 597D Final Project Presentation December 10, 2008.
A New Theory of Neocortex and Its Implications for Machine Intelligence TTI/Vanguard, All that Data February 9, 2005 Jeff Hawkins Director The Redwood.
1 ECE 517: Reinforcement Learning in Artificial Intelligence Lecture 17: TRTRL, Implementation Considerations, Apprenticeship Learning Dr. Itamar Arel.
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
Dr. Z. R. Ghassabi Spring 2015 Deep learning for Human action Recognition 1.
Last Words DM 1. Mining Data Steams / Incremental Data Mining / Mining sensor data (e.g. modify a decision tree assuming that new examples arrive continuously,
Introduction to Neural Networks and Example Applications in HCI Nick Gentile.
Chapter 6 – Architectural Design Lecture 1 1Chapter 6 Architectural design.
1 ECE 517: Reinforcement Learning in Artificial Intelligence Lecture 8: Dynamic Programming – Value Iteration Dr. Itamar Arel College of Engineering Department.
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 A self-organizing map for adaptive processing of structured.
Why Can't A Computer Be More Like A Brain?. Outline Introduction Turning Test HTM ◦ A. Theory ◦ B. Applications & Limits Conclusion.
CSC321 Lecture 5 Applying backpropagation to shape recognition Geoffrey Hinton.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 6: Applying backpropagation to shape recognition Geoffrey Hinton.
Cognitive models for emotion recognition: Big Data and Deep Learning
1 ECE 517: Reinforcement Learning in Artificial Intelligence Lecture 21: Dynamic Multi-Criteria RL problems Dr. Itamar Arel College of Engineering Department.
Artificial Intelligence: Research and Collaborative Possibilities a presentation by: Dr. Ernest L. McDuffie, Assistant Professor Department of Computer.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Maestro AI Vision and Design Overview Definitions Maestro: A naïve Sensorimotor Engine prototype. Sensorimotor Engine: Combining sensory and motor functions.
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
March 31, 2016Introduction to Artificial Intelligence Lecture 16: Neural Network Paradigms I 1 … let us move on to… Artificial Neural Networks.
HIERARCHICAL TEMPORAL MEMORY WHY CANT COMPUTERS BE MORE LIKE THE BRAIN?
National Taiwan Normal A System to Detect Complex Motion of Nearby Vehicles on Freeways C. Y. Fang Department of Information.
Brief Intro to Machine Learning CS539
Big data classification using neural network
Who am I? Work in Probabilistic Machine Learning Like to teach 
Neural Network Architecture Session 2
The Relationship between Deep Learning and Brain Function
Deep Learning Amin Sobhani.
What is an ANN ? The inventor of the first neuro computer, Dr. Robert defines a neural network as,A human brain like system consisting of a large number.
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
Lecture 22 Clustering (3).
Deep Learning Hierarchical Representations for Image Steganalysis
Deep learning Introduction Classes of Deep Learning Networks
Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang.
Sensorimotor Learning and the Development of Position Invariance
Creating Data Representations
Fundamentals of Neural Networks Dr. Satinder Bal Gupta
August 8, 2006 Danny Budik, Itamar Elhanany Machine Intelligence Lab
Learning Sensorimotor Contingencies
November 8, 2010 Dr. Itamar Arel College of Engineering
CSC 578 Neural Networks and Deep Learning
Presentation transcript:

1 ECE 517 Reinforcement Learning in Artificial Intelligence Lecture 21: Deep Machine Learning Dr. Itamar Arel College of Engineering Department of Electrical Engineering and Computer Science The University of Tennessee Fall 2010 November 8, 2010

ECE Reinforcement Learning in AI 2 RL and General AI RL seems like a good AI framework Some pieces are missing Long/short term memory: what is the optimal value (or cost- to-go) function to be used? Long/short term memory: what is the optimal value (or cost- to-go) function to be used? How do we treat multi-dimensional reward signals? How do we treat multi-dimensional reward signals? How do we deal with high-dimensional inputs (observations)? How do we deal with high-dimensional inputs (observations)? How to generalize to address an near-infinite state space? How to generalize to address an near-infinite state space? How long will it take to train such a system? How long will it take to train such a system? If we want to use hardware – how do we go about doing it? Storage capacity – human brain ~10 14 synapses (i.e. weights) Storage capacity – human brain ~10 14 synapses (i.e. weights) Processing power - ~10 11 neurons Processing power - ~10 11 neurons Communications – fully or partially connected architectures Communications – fully or partially connected architectures

ECE Reinforcement Learning in AI Why Deep Learning? Mimicking the way the brain represents information is a key challenge Deals efficiently with high-dimensionality Deals efficiently with high-dimensionality Handle multi-modal data fusion Handle multi-modal data fusion Capture temporal dependencies spanning large scales Capture temporal dependencies spanning large scales Incremental knowledge acquisition Incremental knowledge acquisition The challenge with high-dimensionality Real-world problem Real-world problem Curse of dimensionality (Bellman) Curse of dimensionality (Bellman) Spatial and temporal dependencies Spatial and temporal dependencies How to represent key features?? How to represent key features?? 3

ECE Reinforcement Learning in AI Main application: classification Hard (unsolved) problem due to … High-dimensionality data High-dimensionality data Distortions (noise, rotation, displacement, perspective, lighting conditions, etc.) Distortions (noise, rotation, displacement, perspective, lighting conditions, etc.) Partial observability Partial observability Mainstream approach … 4 ROI detectionFeature Extraction Classification

ECE Reinforcement Learning in AI The power of hierarchical representation Core idea: partition high-dimensional data to small patches, model them and discover dependencies between them Decomposes the problem Suggests a trade off more scope  less detail Key ideas: Basic cortical circuit Basic cortical circuit Massively parallel architecture Massively parallel architecture Discovers structure based on regularities in the observations Discovers structure based on regularities in the observations Multi-modal Multi-modal Goal: situation/state inference Goal: situation/state inference 5

ECE Reinforcement Learning in AI The power of hierarchical representation (con’t) Hypothesis: the brain represents information using a hierarchical architecture that comprises basic cortical circuits Effective way of dealing with large- scale POMDPs DL – state inference DL – state inference RL – for decision making under uncertainty RL – for decision making under uncertainty Suggest a semi-supervised learning framework Unsupervised – learns structure of natural data Unsupervised – learns structure of natural data Supervised – mapping states to classes Supervised – mapping states to classes 6

ECE Reinforcement Learning in AI 7 The Deep Learning Theory Basic idea is to decompose the large image into smaller images that can each be modeled The hierarchy is one of abstraction Higher levels of the state represent more abstract notions Higher levels of the state represent more abstract notions The higher the layer the more scope it encompasses and less detail it offers The higher the layer the more scope it encompasses and less detail it offers Multi-scale spatial-temporal context representation Multi-scale spatial-temporal context representation Lower levels interpret or control limited domains of experience, or sensory systems Lower levels interpret or control limited domains of experience, or sensory systems Connections from the higher level states predispose some selected transitions in the lower-level state machines

ECE Reinforcement Learning in AI 8 Inspiration: Role of Cerebral Cortex The cerebral cortex (aka neocortex), made up of four lobes, is involved in many complex cognitive functions including: memory, attention, perceptual awareness, "thinking", language and consciousness The cortex is the primary brain subsystem responsible for learning … Rich in neurons (>80% in human brain) Rich in neurons (>80% in human brain) It is the one embedding the hierarchical auto-associative memory architecture It is the one embedding the hierarchical auto-associative memory architecture Receives sensory information from many different sensory organs e.g.: eyes, ears, etc. and processes the information Receives sensory information from many different sensory organs e.g.: eyes, ears, etc. and processes the information Areas that receive that particular information are called sensory areas Areas that receive that particular information are called sensory areas

ECE Reinforcement Learning in AI 9 Deep Machine Learning – general framework The lower layers predict short-term sequences As you go higher in the hierarchy – “less accuracy, broader perspective” Analogy to a general commanding an army, or poem being recited “Surprise” sequences should propagate up to the appropriate layer

ECE Reinforcement Learning in AI 10 DL for Invariant Pattern Recognition Initial focus on the visual cortex Offers an invariant visual pattern recognition in the visual cortex Offers an invariant visual pattern recognition in the visual cortex Recognizing objects despite different scaling, rotations and translations is something humans perform without conscious effort Recognizing objects despite different scaling, rotations and translations is something humans perform without conscious effort Lighting conditions, various noises (additive, multiplicative) Lighting conditions, various noises (additive, multiplicative) Currently difficult for machines learning to achieve The approach taken is that geometric invariance is linked to motion When we look at an object, the patterns on our retina change a lot while the object (cause) remains the same When we look at an object, the patterns on our retina change a lot while the object (cause) remains the same Thus, learning persistent patterns on the retina would correspond to learning objects in the visual world Thus, learning persistent patterns on the retina would correspond to learning objects in the visual world Associating patterns with their causes corresponds to invariant pattern recognition Associating patterns with their causes corresponds to invariant pattern recognition

ECE Reinforcement Learning in AI 11 DL for Invariant Pattern Recognition (cont’) Each level in the system hierarchy has several modules that model cortical regions A module can have several children and one parent, thus modules are arranged in a tree structure A module can have several children and one parent, thus modules are arranged in a tree structure The bottom most level is called level 1 and the level number increases as you go up in the hierarchy The bottom most level is called level 1 and the level number increases as you go up in the hierarchy Inputs go directly to the modules at level 1 Inputs go directly to the modules at level 1 The level 1 modules have small receptive fields compared to the size of the total image, i.e., these modules receive their inputs from a small patch of the visual field The level 1 modules have small receptive fields compared to the size of the total image, i.e., these modules receive their inputs from a small patch of the visual field Several such level 1 modules tile the visual field, possibly with overlap Several such level 1 modules tile the visual field, possibly with overlap

ECE Reinforcement Learning in AI 12 General System Architecture Thus a level 2 module covers more of the visual field compared to a level 1 module. However, a level 2 module gets it information only through a level 1 module This pattern is repeated in the hierarchy This pattern is repeated in the hierarchy Receptive field sizes increase as one goes up the hierarchy Receptive field sizes increase as one goes up the hierarchy The module at the root of the tree covers the entire visual field, by pooling inputs from its child modules The module at the root of the tree covers the entire visual field, by pooling inputs from its child modules

ECE Reinforcement Learning in AI 13 Learning Framework Let X n (1) and X n (2) denote the sequence of inputs to modules 1 and 2 Learning occurs in three phases: First, each module learns the most likely sequences of its inputs Second, each module passes an index of its most-likely observed input sequence Third, each module learns the most frequent “coincidences” of indices originating from the lower layer modules Next …

ECE Reinforcement Learning in AI 14 Contextual Embedding (if exists) Feedback loop from layer 2 back to layer 1 (its children) This feedback provides contextual inference (from higher layers) This stage is initiated once the level 2 module has formed its alphabet, Y k This stage is initiated once the level 2 module has formed its alphabet, Y k Lower layer nodes eventually learn the CPD matrix P(X (1) |Y) Lower layer nodes eventually learn the CPD matrix P(X (1) |Y)

ECE Reinforcement Learning in AI 15 Bayesian Network Obtained Bottom layer random variables correspond to quantizations on input patterns The r.v. at the intermediate layers represent object parts that move together persistently R.V. at the top layer correspond to objects

ECE Reinforcement Learning in AI 16 Learning algorithm (cont.) After the system has learned (seen many example) and obtained the CPD at each layer, we seek where I is the image observed. A Bayesian Belief Propagation method is typically used to determine the above, based on hierarchy of beliefs Drawbacks of current schemes No “natural” spatiotemporal information representation No “natural” spatiotemporal information representation Layer-by-layer training is needed Layer-by-layer training is needed Modality independent (most current schemes limited to image data sets) Modality independent (most current schemes limited to image data sets)

ECE Reinforcement Learning in AI 17 Alternative Explanations for Biological Phenomena Physiological experiments found that neurons sometimes respond to illusory contours in a Kanizsa figure In other words, a neuron responds to a contour that does not exist in its receptive field Possible interpretation: activity of a neuron represents the probability that a contour should be present Possible interpretation: activity of a neuron represents the probability that a contour should be present Originates from its own state and the state information of higher-level neurons Originates from its own state and the state information of higher-level neurons DL based explanation for this phenomena Contrary to current hypothesis that assume “signal subtraction” occurs for some reason Contrary to current hypothesis that assume “signal subtraction” occurs for some reason