Andrew Ng CS228: Deep Learning & Unsupervised Feature Learning Andrew Ng TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:

Slides:



Advertisements
Similar presentations
Scalable Learning in Computer Vision
Advertisements

Elaine Hale, Wotao Yin, Yin Zhang Computational and Applied Mathematics Rice University , McMaster University, ICCOPT II TexPoint fonts used.
Advanced topics.
Rajat Raina Honglak Lee, Roger Grosse Alexis Battle, Chaitanya Ekanadham, Helen Kwong, Benjamin Packer, Narut Sereewattanawoot Andrew Y. Ng Stanford University.
Adam Coates Deep Learning and HPC Adam Coates Visiting Scholar at IU Informatics Post-doc at Stanford CS.
Tiled Convolutional Neural Networks TICA Speedup Results on the CIFAR-10 dataset Motivation Pretraining with Topographic ICA References [1] Y. LeCun, L.
Beam Sensor Models Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics TexPoint fonts used in EMF. Read.
Differentiable Sparse Coding David Bradley and J. Andrew Bagnell NIPS
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
Recent Developments in Deep Learning Quoc V. Le Stanford University and Google.
1 A scheme for racquet sports video analysis with the combination of audio-visual information Visual Communication and Image Processing 2005 Liyuan Xing,
Image classification by sparse coding.
Deep Learning.
STANFORD Hierarchical Apprenticeship Learning with Application to Quadruped Locomotion J. Zico Kolter, Pieter Abbeel, Andrew Y. Ng Goal Initial Position.
Self Taught Learning : Transfer learning from unlabeled data Presented by: Shankar B S DMML Lab Rajat Raina et al, CS, Stanford ICML 2007.
Chapter 11 Beyond Bag of Words. Question Answering n Providing answers instead of ranked lists of documents n Older QA systems generated answers n Current.
Learning to grasp objects with multiple contact points Quoc V. Le, David Kamm, Arda Kara, Andrew Y. Ng.
K-means Based Unsupervised Feature Learning for Image Recognition Ling Zheng.
AN ANALYSIS OF SINGLE- LAYER NETWORKS IN UNSUPERVISED FEATURE LEARNING [1] Yani Chen 10/14/
77 CHAPTER INPUT AND OUTPUT. © 2005 The McGraw-Hill Companies, Inc. All Rights Reserved. 7-2 Competencies Define input Describe keyboard entry, pointing.
Lecture 29: Recent work in recognition CS4670: Computer Vision Noah Snavely.
Multimodal Deep Learning
Image Classification using Sparse Coding: Advanced Topics
Introduce about sensor using in Robot NAO Department: FTI-FHO-FPT Presenter: Vu Hoang Dung.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Comp 5013 Deep Learning Architectures Daniel L. Silver March,
Computer Vision and Data Mining Research Projects Longin Jan Latecki Computer and Information Sciences Dept. Temple University
Deep Learning for Speech and Language Yoshua Bengio, U. Montreal NIPS’2009 Workshop on Deep Learning for Speech Recognition and Related Applications December.
Space-Indexed Dynamic Programming: Learning to Follow Trajectories J. Zico Kolter, Adam Coates, Andrew Y. Ng, Yi Gu, Charles DuHadway Computer Science.
End-to-End Text Recognition with Convolutional Neural Networks
A shallow introduction to Deep Learning
Large-scale Deep Unsupervised Learning using Graphics Processors
Input & Output Hardware Chapter 6 Taking Charge of Computing and Communications.
Zhenbao Liu 1, Shaoguang Cheng 1, Shuhui Bu 1, Ke Li 2 1 Northwest Polytechnical University, Xi’an, China. 2 Information Engineering University, Zhengzhou,
Andrew Ng Feature learning for image classification Kai Yu and Andrew Ng.
Project 2 code & artifact due Friday Midterm out tomorrow (check your ), due next Fri Announcements TexPoint fonts used in EMF. Read the TexPoint.
LeCun, Bengio, And Hinton doi: /nature14539
Andrew Ng Machine Learning and AI via Brain simulations Andrew Ng Stanford University Adam Coates Quoc Le Honglak Lee Andrew Saxe Andrew Maas Chris Manning.
Jack Pinches INFO410 & INFO350 S INFORMATION SCIENCE Computer Vision I.
Andrew Ng, Director, Stanford Artificial Intelligence Lab
A shallow look at Deep Learning
1 Andrew Ng, Associate Professor of Computer Science Robots and Brains.
Project 2 due today Project 3 out today Announcements TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAA.
1 Andrew Ng, Associate Professor of Computer Science Robots and Brains.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Lecture 5: Feature detection and matching CS4670 / 5670: Computer Vision Noah Snavely.
Yann LeCun Other Methods and Applications of Deep Learning Yann Le Cun The Courant Institute of Mathematical Sciences New York University
Information & Software Technologies
Computer Input Device: Digital Camera
Convolutional Neural Network
Deep Learning Amin Sobhani.
Article Review Todd Hricik.
Restricted Boltzmann Machines for Classification
Machine Learning Ali Ghodsi Department of Statistics
Unsupervised Learning of Tag-Sensitive Descriptors
State-of-the-art face recognition systems
Lecture 2: Edge detection
Computer Vision James Hays
Non-linear hypotheses
--- Stereoscopic Vision and Range Finders
--- Stereoscopic Vision and Range Finders
convolutional neural networkS
Multiple Feature Learning for Action Classification
Computer Basics: How Do Computers Work? Part II
Using Natural Language Processing to Aid Computer Vision
Android Sensor Programming
Neural Network Pipeline CONTACT & ACKNOWLEDGEMENTS
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Bidirectional LSTM-CRF Models for Sequence Tagging
Text Processing Like Humans Do:
Presentation transcript:

Andrew Ng CS228: Deep Learning & Unsupervised Feature Learning Andrew Ng TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAA Pieter Abbeel Adam Coates Zico Kolter Ian Goodfellow Quoc Le Honglak Lee Rajat Raina Andrew Saxe

Andrew Ng How is computer perception done? Image Low-level vision features Recognition Image Grasp point Low-level features Object detection Computer vision is hard!

Andrew Ng How is computer perception done? Image Grasp point Low-level features Image Vision features Recognition Object detection Audio Audio features Speaker ID Audio classification NLP Text Text features Text classification, MT, IR, etc.

Andrew Ng Sensor representations Input Learning/AI algorithm Low-level features

Andrew Ng Computer vision features SIFT Spin image HoG RIFT Textons GLOH

Andrew Ng Audio features ZCR Spectrogram MFCC Rolloff Flux

Andrew Ng NLP features Parser features NER/SRL Stemming POS tagging Anaphora WordNet features

Andrew Ng A plethora of sensors Camera array 3d range scan (laser scanner) 3d range scans (flash lidar) Audio A general-purpose algorithm for good sensor representations? Visible light image Thermal Infrared

Andrew Ng Sensor representation in the brain [BrainPort; Martinez et al; Roe et al.] Seeing with your tongue Human echolocation (sonar) Auditory cortex learns to see. Auditory Cortex

Andrew Ng Learning abstract representations pixels edges object parts (combination of edges) object models [Related work: Deep learning, Hinton, Bengio, LeCun, and others.]

Andrew Ng Feature learning for audio Learned features correspond to phonemes and other “basic units” of sound. Learned features Algorithm :

Andrew Ng TIMIT Phone classificationAccuracy Prior art (Clarkson et al.,1999) 79.6% Stanford Feature learning 80.3% TIMIT Speaker identificationAccuracy Prior art (Reynolds, 1995) 99.7% Stanford Feature learning 100.0% Audio Images Multimodal (audio/video) CIFAR Object classificationAccuracy Prior art (Yu and Zhang, 2010) 74.5% Stanford Feature learning 79.6% NORB Object classificationAccuracy Prior art (Ranzato et al., 2009) 94.4% Stanford Feature learning 97.0% AVLetters Lip readingAccuracy Prior art (Zhao et al., 2009) 58.9% Stanford Feature learning 65.8% Galaxy Other feature learning records: Different phone recognition task (Hinton), PASCAL VOC object classification (Yu) Hollywood2 ClassificationAccuracy Prior art (Laptev et al., 2004) 48% Stanford Feature learning 53% KTHAccuracy Prior art (Wang et al., 2010) 92.1% Stanford Feature learning 93.9% UCFAccuracy Prior art (Wang et al., 2010) 85.6% Stanford Feature learning 86.5% YouTubeAccuracy Prior art (Liu et al., 2009) 71.2% Stanford Feature learning 75.8% Video