…. 2 Ongoing software project, not “theory” Encapsulated internals & interfaces Today: –Details of module internals –Details of architecture & signaling/feedback.

Slides:



Advertisements
Similar presentations
Clustering. How are we doing on the pass sequence? Pretty good! We can now automatically learn the features needed to track both people But, it sucks.
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
2806 Neural Computation Self-Organizing Maps Lecture Ari Visa.
Unsupervised Learning
Learning to make specific predictions using Slow Feature Analysis.
K Means Clustering , Nearest Cluster and Gaussian Mixture
Neural Networks Chapter 9 Joost N. Kok Universiteit Leiden.
Ch. 4: Radial Basis Functions Stephen Marsland, Machine Learning: An Algorithmic Perspective. CRC 2009 based on slides from many Internet sources Longin.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
Structure learning with deep neuronal networks 6 th Network Modeling Workshop, 6/6/2013 Patrick Michl.
Nearest Neighbor. Predicting Bankruptcy Nearest Neighbor Remember all your data When someone asks a question –Find the nearest old data point –Return.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
x – independent variable (input)
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Motion Analysis (contd.) Slides are from RPI Registration Class.
Segmentation Divide the image into segments. Each segment:
CS CS 175 – Week 2 Processing Point Clouds Registration.
Three Algorithms for Nonlinear Dimensionality Reduction Haixuan Yang Group Meeting Jan. 011, 2005.
Greg GrudicIntro AI1 Introduction to Artificial Intelligence CSCI 3202: The Perceptron Algorithm Greg Grudic.
Pattern Recognition. Introduction. Definitions.. Recognition process. Recognition process relates input signal to the stored concepts about the object.
Hazırlayan NEURAL NETWORKS Radial Basis Function Networks I PROF. DR. YUSUF OYSAL.
NonLinear Dimensionality Reduction or Unfolding Manifolds Tennenbaum|Silva|Langford [Isomap] Roweis|Saul [Locally Linear Embedding] Presented by Vikas.
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Monte Carlo Methods in Partial Differential Equations.
Radial Basis Function (RBF) Networks
Radial Basis Function Networks
Face Recognition Using Neural Networks Presented By: Hadis Mohseni Leila Taghavi Atefeh Mirsafian.
Nonlinear Dimensionality Reduction Approaches. Dimensionality Reduction The goal: The meaningful low-dimensional structures hidden in their high-dimensional.
October 8, 2013Computer Vision Lecture 11: The Hough Transform 1 Fitting Curve Models to Edges Most contours can be well described by combining several.
CSE 185 Introduction to Computer Vision
Manifold learning: Locally Linear Embedding Jieping Ye Department of Computer Science and Engineering Arizona State University
Soft Computing Colloquium 2 Selection of neural network, Hybrid neural networks.
Self Organizing Maps (SOM) Unsupervised Learning.
CS 551/651 Advanced Computer Graphics Warping and Morphing Spring 2002.
1 Lesson 3: Choosing from distributions Theory: LLN and Central Limit Theorem Theory: LLN and Central Limit Theorem Choosing from distributions Choosing.
Chapter 9 Neural Network.
COMMON EVALUATION FINAL PROJECT Vira Oleksyuk ECE 8110: Introduction to machine Learning and Pattern Recognition.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 16 Nov, 3, 2011 Slide credit: C. Conati, S.
EECS 274 Computer Vision Segmentation by Clustering II.
Computer Vision Lab. SNU Young Ki Baik Nonlinear Dimensionality Reduction Approach (ISOMAP, LLE)
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: ML and Simple Regression Bias of the ML Estimate Variance of the ML Estimate.
MACHINE LEARNING 8. Clustering. Motivation Based on E ALPAYDIN 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2  Classification problem:
Non-Bayes classifiers. Linear discriminants, neural networks.
Clustering.
Nonlinear Dimensionality Reduction Approach (ISOMAP)
Project 11: Determining the Intrinsic Dimensionality of a Distribution Okke Formsma, Nicolas Roussis and Per Løwenborg.
381 Self Organization Map Learning without Examples.
Radial Basis Function ANN, an alternative to back propagation, uses clustering of examples in the training set.
Clustering Instructor: Max Welling ICS 178 Machine Learning & Data Mining.
An Introduction of Independent Component Analysis (ICA) Xiaoling Wang Jan. 28, 2003.
Example Apply hierarchical clustering with d min to below data where c=3. Nearest neighbor clustering d min d max will form elongated clusters!
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Given a set of data points as input Randomly assign each point to one of the k clusters Repeat until convergence – Calculate model of each of the k clusters.
Clustering Approaches Ka-Lok Ng Department of Bioinformatics Asia University.
May 2003 SUT Color image segmentation – an innovative approach Amin Fazel May 2003 Sharif University of Technology Course Presentation base on a paper.
Machine Learning 12. Local Models.
Machine Learning Supervised Learning Classification and Regression
Spectral Methods for Dimensionality
Self-Organizing Network Model (SOM) Session 11
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Ch10 : Self-Organizing Feature Maps
Haim Kaplan and Uri Zwick
Synaptic Dynamics: Unsupervised Learning
Structure learning with deep autoencoders
ISOMAP TRACKING WITH PARTICLE FILTERING
KMeans Clustering on Hadoop Fall 2013 Elke A. Rundensteiner
Text Categorization Berlin Chen 2003 Reference:
EM Algorithm and its Applications
MIRA, SVM, k-NN Lirong Xia. MIRA, SVM, k-NN Lirong Xia.
Presentation transcript:

2 Ongoing software project, not “theory” Encapsulated internals & interfaces Today: –Details of module internals –Details of architecture & signaling/feedback –Single, clean, simple inputs –(26 slides) Not yet: –time –noise –robustness –multiple/partial hypotheses

3 One “compressor”: Generic memory unit Learns about low-dim structure in high-dim data Converts live data between low-dim   high-dim Hierarchy of compressors: Each learns from compressed & combined output of those below Bi-directional (feedback)

4 Compressor Internals Probability estimation Bi-directional mapping Matching to previous compression Compressing Quantizing & representing high-dim input old: P = p 1 + p 2 + …

5 Quantizing & representing high-dim input “Point” = position, weight, radius Two point-clouds: mapping vs. learning (sync occasionally) 1. Find 3 closest cloud-points X Result: Point-cloud approximates input cloud, with roughly equal weight per point 2. Choose the lightest 3. Move it to absorb new point, preserving center of mass 4. Increase weight 5. Update radius 6. (prune lightweight points) Online updates:

6 Compressing high to low (ISOMAP) 1. Find local distances in high-dim space 2. Create long-range distances from shortest piecewise path (“geodesic”) 3. Link “islands” until all D ij defined 4. Diagonalize F(Dij) to get low-dim cloud (arbitrary coordinates)

7 Keeping new maps consistent with old ones Old cloud The low-dim mapping is not always unique… …so rotate & stretch new cloud to minimize distance from old one (SVD) New cloud Rotated new cloud

8 Mapping new points using point-clouds 1. Find new point’s closest 4-5 neighbors 2. Express it as their center-of-mass (SVD) 3. Construct low-dim output from corresponding neighbors & weights 4. Also works mapping low  high = W2W2 W1W1 W3W3 W4W4 W2W2 W1W1 W3W3 W4W4 =

9 Prob. Estimation Each point is center of gaussian P = p 1 + p 2 + … RiRi P i = exp ( -0.5 r 2 / R 2 ) / (R D P tot ) “Probability” of test point is sum over local gaussians P = p 1 + p 2 + … Probability =“Closeness” to manifold = how much to trust this point … use it later in mixing estimates.

10 Compressors interacting Creating forward output Feedback mixed back in Settling

11 Creating output Map from high to low dim Expose result to all Compressors above Re-map output backwards to high dim Expose as feedback to Compressors below

12 P Mix feedback into output 2. Get probabilities of feedback and own output 1. Average feedback from above 3. Create weighted mixture of them P

13 2. Iterate a few times to settle Updating and settling 1. Expose mixture as updated output, and map downward as updated feedback --- done with description of system --

14 General simulation results 3-layer hierarchy with 2-1 convergence Input is 9x6 “pixel” space with random illumination Display low-dim output in 2-D  color

15 Simple 1-dim illumination How does each module map the input space? ?

16

17 Toroidal 1-dim illumination How does each module map the circular input space? ? =

18

19 2-dim spot illumination How does each module map the 2-D input space? ?

20

21 “Hallucinating” spots driven from above 1.Force activity at a single location in top module 2. Let feedback move down 3. Look at what lower modules think input ought to be ?? ?

22

23 2-dim clustered spots (left & right) How does each module map the 2-D input space? ?

24

25 Next steps Architecture –Time –Reference problem –Reference platform –Integration method –Separate streams for transforms vs. objects –Get people involved! Algorithms –Noise –Multiple hypotheses –Distributed representation –“neurons” –Better quantization, mapping, robustness