Committee Update Building a visual hierarchy Andrew Smith 30 July 2008.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

The Helmholtz Machine P Dayan, GE Hinton, RM Neal, RS Zemel
Naïve-Bayes Classifiers Business Intelligence for Managers.
Chapter 2.
Neural Network Models in Vision Peter Andras
Central Visual Processes. Anthony J Greene2 Central Visual Pathways I.Primary Visual Cortex Receptive Field Columns Hypercolumns II.Spatial Frequency.
Texture. Limitation of pixel based processing Edge detection with different threshold.
Object recognition and scene “understanding”
Object Recognition with Features Inspired by Visual Cortex T. Serre, L. Wolf, T. Poggio Presented by Andrew C. Gallagher Jan. 25, 2007.
Salvatore giorgi Ece 8110 machine learning 5/12/2014
HMAX Models Architecture Jim Mutch March 31, 2010.
Texture. Edge detectors find differences in overall intensity. Average intensity is only simplest difference. many slides from David Jacobs.
Thesis title: “Studies in Pattern Classification – Biological Modeling, Uncertainty Reasoning, and Statistical Learning” 3 parts: (1)Handwritten Digit.
CSC321: Neural Networks Lecture 3: Perceptrons
BEYOND SIMPLE FEATURES: A LARGE-SCALE FEATURE SEARCH APPROACH TO UNCONSTRAINED FACE RECOGNITION Nicolas Pinto Massachusetts Institute of Technology David.
Texture Turk, 91.
Un Supervised Learning & Self Organizing Maps Learning From Examples
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
A Theory of Cerebral Cortex (or, “How Your Brain Works”) Andrew Smith (CSE)
CS292 Computational Vision and Language Visual Features - Colour and Texture.
Ranking by Odds Ratio A Probability Model Approach let be a Boolean random variable: document d is relevant to query q otherwise Consider document d as.
October 5, 2010Neural Networks Lecture 9: Applying Backpropagation 1 K-Class Classification Problem Let us denote the k-th class by C k, with n k exemplars.
Computer vision: models, learning and inference Chapter 10 Graphical Models.
October 7, 2010Neural Networks Lecture 10: Setting Backpropagation Parameters 1 Creating Data Representations On the other hand, sets of orthogonal vectors.
CS Bayesian Learning1 Bayesian Learning. CS Bayesian Learning2 States, causes, hypotheses. Observations, effect, data. We need to reconcile.
Super-Resolution of Remotely-Sensed Images Using a Learning-Based Approach Isabelle Bégin and Frank P. Ferrie Abstract Super-resolution addresses the problem.
Time Series Data Analysis - II
Multiclass object recognition
Convolutional Neural Networks for Image Processing with Applications in Mobile Robotics By, Sruthi Moola.
Genetic Regulatory Network Inference Russell Schwartz Department of Biological Sciences Carnegie Mellon University.
Midterm Review Rao Vemuri 16 Oct Posing a Machine Learning Problem Experience Table – Each row is an instance – Each column is an attribute/feature.
Presented by Tienwei Tsai July, 2005
CS 8751 ML & KDDSupport Vector Machines1 Support Vector Machines (SVMs) Learning mechanism based on linear programming Chooses a separating plane based.
Neural Information in the Visual System By Paul Ruvolo Bryn Mawr College Fall 2012.
A Scalable Self-organizing Map Algorithm for Textual Classification: A Neural Network Approach to Thesaurus Generation Dmitri G. Roussinov Department of.
2 2  Background  Vision in Human Brain  Efficient Coding Theory  Motivation  Natural Pictures  Methodology  Statistical Characteristics  Models.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Perceptual Multistability as Markov Chain Monte Carlo Inference.
Stylization and Abstraction of Photographs Doug Decarlo and Anthony Santella.
Virtual Vector Machine for Bayesian Online Classification Yuan (Alan) Qi CS & Statistics Purdue June, 2009 Joint work with T.P. Minka and R. Xiang.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
Chapter 3: Neural Processing and Perception. Neural Processing and Perception Neural processing is the interaction of signals in many neurons.
1 CMSC 671 Fall 2001 Class #25-26 – Tuesday, November 27 / Thursday, November 29.
Group-7. Mental World What is Cognition? Cognition : Understanding and trying to make sense of the world Information processing Development of concepts.
Foundations (cont.) Complexity Testing explanations in psychology Cognitive Neuroscience.
Image Stabilization by Bayesian Dynamics Yoram Burak Sloan-Swartz annual meeting, July 2009.
MIT AI Lab / LIDS Laboatory for Information and Decision Systems & Artificial Intelligence Laboratory Massachusetts Institute of Technology A Unified Multiresolution.
© 2009 Robert Hecht-Nielsen. All rights reserved. 1 Andrew Smith University of California, San Diego Building a Visual Hierarchy.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Introduction to Sampling Methods Qi Zhao Oct.27,2004.
Machine Vision Edge Detection Techniques ENT 273 Lecture 6 Hema C.R.
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
Basics of Computational Neuroscience. What is computational neuroscience ? The Interdisciplinary Nature of Computational Neuroscience.
HIERARCHICAL TEMPORAL MEMORY WHY CANT COMPUTERS BE MORE LIKE THE BRAIN?
Ch 1. Introduction Pattern Recognition and Machine Learning, C. M. Bishop, Updated by J.-H. Eom (2 nd round revision) Summarized by K.-I.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
1 Nonlinear models for Natural Image Statistics Urs Köster & Aapo Hyvärinen University of Helsinki.
Convolutional Neural Network
CSC2535: Computation in Neural Networks Lecture 11 Extracting coherent properties by maximizing mutual information across space or time Geoffrey Hinton.
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Nearest-neighbor matching to feature database
Nearest-neighbor matching to feature database
Outline Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no.
Computer Vision Lecture 16: Texture II
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Texture.
The Naïve Bayes (NB) Classifier
Neural Network Models in Vision
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Outline Texture modeling - continued Markov Random Field models
Presentation transcript:

Committee Update Building a visual hierarchy Andrew Smith 30 July 2008

Outline Confabulation theory Summary Comparisons to other AI techniques Human Visual System Building A Visual Hierarchy Learning Inference Texture modeling (applications) Future work (dissertation defence, Spring 2009)

Confabulation Theory A theory of the mechanism of thought –Cortex/thalamus is divided into thousands of modules (1,000,000s of neurons). –Each module contains a lexicon of symbols. –Symbols are sparse populations (100s) of neurons within a module. –Symbols are stable states of a cortex-thalamus attractor circuit.

Confabulation theory (1/4) Key concept 1: Modules contain symbols, the atoms of our mental universe. Smell module: Apple, flower, rotten, … Word module: ‘rose’ ‘the’ ‘and’ ‘it’ ‘France’ ‘Joe’ … Abstract planning modules, etc. Modules are small patches of thalamocortical neurons. Each symbol is a sparse popuation of those neurons.

Confabulation theory (1/4)

Confabulation theory (2/4) Key concept 2: All cognitive knowledge is knowledge links between these symbols. Smell module: Apple, flower, rotten, … Word module: ‘the’ ‘and’ ‘it’ ‘France’ ‘Joe’ ‘apple’ … Only symbols that are meaningfully co-occurring may become linked.

Confabulation theory (3/4)

Key concept 3: A confabulation operation is the universal computational mechanism. Given evidence a, b, c pick answer x such that: x = argmax x’ p(a, b, c | x’) We say x has maximum cogency.

Confabulation theory (3/4) Fundamental Theorem of Cognition:[1] p(  ) 4 = p(  )/p(  ) ∙p(  )/p(  ) ∙p(  )/p(  ) ∙p(  )/p(  ) ∙p(  )p(  )p(g|  )p(  ) If the first four terms remain nearly constant w.r.t , maximizing the fifth term maximizes cogency (the conditional joint).

Confabulation theory (3/4)

Confabulation theory (4/4) Key concept 4: Each confabulation operation launches a control signal to other modules. Control mechanism of inference – studied by others in the lab. (not here)

Similarities to other AI / ML Bayesian networks – a special case A “confabulation network” is similar to a Bayesian Net with: Symbolic variables (discrete & finite & exclusive state) with equal priors. Naïve-Bayes assumption for CP tables. Can use similar learning algorithms (counting for CPs) Hinton’s (unrestricted) Bolzman Machines – generalized: Do not require complete connectivity (many) more than two states. Can use stochastic (Monte Carlo) ‘execution’

Outline Confabulation theory Summary Comparisons to other AI techniques Human Visual System A Visual Hierarchy Learning Inference Texture modeling Future Work (i.e. my thesis)

Human Visual System 1) Retina – “pixels” 2) Lateral Geniculate Nucleus (LGN) “center-surround” representation 3) Primary(…) Visual cortex (V1 …) Simple cells: Hubel Weisel (1959) Modeled by Dennis Gabor features [] Complex cells more complicated (end-stops, bars, ???) Take inspiration for our first and second-level features

Outline Confabulation theory Summary Comparisons to other AI techniques Human Visual System Building A Visual Hierarchy Learning Inference Texture modeling Future Work (i.e. my thesis)

Confabulation & vision Features (symbols) develop in a layer of the hierarchy as commonly seen inputs from their inputs. Knowledge links are simple conditional probabilities: p(  |  ) where  and  are symbols in connected modules) All knowledge can therefore be learned by simple co- occurrence counting. p(  |  ) = C( ,  ) / C(  )

Building a vision hierarchy Can no longer use SSE to evaluate model Instead, make use of generative model: –Always be able to generate a plausible image.

Data set 4, Mpix natural images (BW)

Vision Hierarch – level “0” We know the first transformation from neuroscience research: simple cells approximate Gabor filters. 5 scales, 16 orientations (odd + even)

Vision Hierarch – level “0” Does the full convolution preserve information in images? (inverted by LS) Very closely.

Vision Hierarchy – level 1 We now have a simple-cell like representation. How to create a symbolic representation? Apply principle: Collect common sets of inputs from simple cells: similar to a Vector Quantizer. Keep the 5-scales separate –(quantize 16-dimensions, not 80)

Vision Hierarchy – level 1 To create actual symbols, we use a vector quantizer –Trade-offs (threshold of quantizer) : Number of symbols Preservation of information Probability accuracy Solution Use angular distance metric (dot-product) –Keep only symbols that occurred in training set more than 200 times, to get accurate p(  ). –After training, ~95% of samples should be within threshold of at least one symbol. –Pick a threshold so images can be plausibly generated.

Vision Hierarchy – level 1 Oops! Ignoring wavelet magnitude makes all “texture features” equally prominent.

Vision Hierarchy – level 1 Solution, use binning (into 5 magnitudes), then apply vector quantizers).

Vision Hierarchy – level 1 ~10,000 symbols are learned for each of the 5 scales. Complex features develop.

Vision Hierarchy – level 1 Now image is re- represented as 5 “planes” of symbols:

Outline Confabulation theory Summary Comparisons to other AI techniques Human Visual System Building A Visual Hierarchy Learning Inference Texture modeling Future Work (i.e. my thesis)

Texture modeling - Learning We can now represent an image as five superimposed grids of symbols. Transform data set Learn which symbols are typically next to which. (knowledge links)

Knowledge links: Learn which symbols may be next to which symbols (conditional probabilities) Learn which symbols may be over/under which symbols. Go out to ‘radius’ 5.

Texture modeling – Inference 1 What if a portion of our image symbol representation is damaged? Blind spot CCD defect brain lesion We can use confabulation (generation) to infer a plausible replacement.

Texture modeling – Inference 1 Fill in missing region by confabulating from lateral & different scale neighbors (rad 5).

Texture modeling

Conclusions This visual hierarchy does an excellent job at capturing an image up to a certain order of complexity. Given this visual hierarchy and its learned knowledge links, missing regions could plausibly filled in. This could be a reasonable explanation for what animals do.

Texture modeling – Inference 2 Super-resolution: –If we have a low resolution image, can we confabulate (generate) a high- resolution version? –“Space out” the symbols, and confabulate values for the new neighbors

Texture modeling

Super-resolution: conclusions Having learned the statistics of natural images, the generative properties of this hierarchy can confabulate (generate) plausible high-resolution versions of its input.

Outline Confabulation theory Summary Comparisons to other AI techniques Human Visual System Building A Visual Hierarchy Learning Inference Texture modeling Future Work (Dissertation)

The next level… Level 2 symbol hierarchy Collect commonly recurring regions of level 1 symbols. Symbols at Level 2 will fit together like puzzle pieces. Thank you!