Announcements  Final exam: Friday Dec 19, 8am-11am, 1 Pimentel  Closed-book, closed-notes except 1 page of notes  Grading  Your midterm score is replaced.

Slides:



Advertisements
Similar presentations
CS Spring 2009 CS 414 – Multimedia Systems Design Lecture 4 – Digital Image Representation Klara Nahrstedt Spring 2009.
Advertisements

September 2, 2014Computer Vision Lecture 1: Human Vision 1 Welcome to CS 675 – Computer Vision Fall 2014 Instructor: Marc Pomplun Instructor: Marc Pomplun.
Many slides based on P. FelzenszwalbP. Felzenszwalb General object detection with deformable part-based models.
Presented by: Mingyuan Zhou Duke University, ECE September 18, 2009
More sliding window detection: Discriminative part-based models Many slides based on P. FelzenszwalbP. Felzenszwalb.
Computer and Robot Vision I
December 5, 2013Computer Vision Lecture 20: Hidden Markov Models/Depth 1 Stereo Vision Due to the limited resolution of images, increasing the baseline.
CPSC 425: Computer Vision (Jan-April 2007) David Lowe Prerequisites: 4 th year ability in CPSC Math 200 (Calculus III) Math 221 (Matrix Algebra: linear.
Advanced Computer Vision Introduction Goal and objectives To introduce the fundamental problems of computer vision. To introduce the main concepts and.
PSY 5018H: Math Models Hum Behavior, Prof. Paul Schrater, Spring 2004 Vision as Optimal Inference The problem of visual processing can be thought of as.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 6: Low-level features 1 Computational Architectures in Biological.
Vision Computing An Introduction. Visual Perception Sight is our most impressive sense. It gives us, without conscious effort, detailed information about.
High-Quality Video View Interpolation
Object Recognition: Conceptual Issues Slides adapted from Fei-Fei Li, Rob Fergus, Antonio Torralba, and K. Grauman.
Spatial Pyramid Pooling in Deep Convolutional
1B50 – Percepts and Concepts Daniel J Hulme. Outline Cognitive Vision –Why do we want computers to see? –Why can’t computers see? –Introducing percepts.
November 29, 2004AI: Chapter 24: Perception1 Artificial Intelligence Chapter 24: Perception Michael Scherger Department of Computer Science Kent State.
Computer Vision Spring ,-685 Instructor: S. Narasimhan Wean Hall 5409 T-R 10:30am – 11:50am.
Generic object detection with deformable part-based models
A Brief Overview of Computer Vision Jinxiang Chai.
Nantes Machine Learning Meet-up 2 February 2015 Stefan Knerr CogniTalk
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 35 – Review for midterm.
Active Vision Key points: Acting to obtain information Eye movements Depth from motion parallax Extracting motion information from a spatio-temporal pattern.
Perception Introduction Pattern Recognition Image Formation
Computer Vision CS 776 Spring 2014 Recognition Machine Learning Prof. Alex Berg.
A shallow introduction to Deep Learning
Introduction EE 520: Image Analysis & Computer Vision.
December 4, 2014Computer Vision Lecture 22: Depth 1 Stereo Vision Comparing the similar triangles PMC l and p l LC l, we get: Similarly, for PNC r and.
Computer Vision: Summary and Discussion Computer Vision ECE 5554 Virginia Tech Devi Parikh 11/21/
Computer Vision Why study Computer Vision? Images and movies are everywhere Fast-growing collection of useful applications –building representations.
Computer Science Department Pacific University Artificial Intelligence -- Computer Vision.
Building high-level features using large-scale unsupervised learning Anh Nguyen, Bay-yuan Hsu CS290D – Data Mining (Spring 2014) University of California,
EIE426-AICV 1 Computer Vision Filename: eie426-computer-vision-0809.ppt.
Why is computer vision difficult?
Computer Vision 776 Jan-Michael Frahm 12/05/2011 Many slides from Derek Hoiem, James Hays.
School of Engineering and Computer Science Victoria University of Wellington Copyright: Peter Andreae, VUW Image Recognition COMP # 18.
Computing & Information Sciences Kansas State University Lecture 40 of 42 CIS 530 / 730 Artificial Intelligence Lecture 40 of 42 A Brief Survey of Computer.
Epitomic Location Recognition A generative approach for location recognition K. Ni, A. Kannan, A. Criminisi and J. Winn In proc. CVPR Anchorage,
Layered Object Detection for Multi-Class Image Segmentation UC Irvine Yi Yang Sam Hallman Deva Ramanan Charless Fowlkes.
September 3, 2013Computer Vision Lecture 1: Human Vision 1 Welcome to CS 675 – Computer Vision Fall 2013 Instructor: Marc Pomplun Instructor: Marc Pomplun.
Color and Brightness Constancy Jim Rehg CS 4495/7495 Computer Vision Lecture 25 & 26 Wed Oct 18, 2002.
Lecture 9 Feature Extraction and Motion Estimation Slides by: Michael Black Clark F. Olson Jean Ponce.
Perception and VR MONT 104S, Fall 2008 Lecture 8 Seeing Depth
Object Recognizing. Deep Learning Success in 2012 DeepNet and speech processing.
More sliding window detection: Discriminative part-based models
Computational Vision Jitendra Malik University of California, Berkeley.
Deep Learning Overview Sources: workshop-tutorial-final.pdf
Network Management Lecture 13. MACHINE LEARNING TECHNIQUES 2 Dr. Atiq Ahmed Université de Balouchistan.
COS 429: Computer Vision Thanks to Chris Bregler.
Xintao Wu University of Arkansas Introduction to Deep Learning 1.
Brief Intro to Machine Learning CS539
Processing visual information for Computer Vision
The Relationship between Deep Learning and Brain Function
Object detection with deformable part-based models
Lecture 24: Convolutional neural networks
R-CNN region By Ilia Iofedov 11/11/2018 BGU, DNN course 2016.
Cheng-Ming Huang, Wen-Hung Liao Department of Computer Science
Computer Vision James Hays
Image Classification.
“The Truth About Cats And Dogs”
Brief Review of Recognition + Context
?. ? White Fuzzy Color Oblong Texture Shape Most problems in vision are underconstrained White Color Most problems in vision are underconstrained.
KFC: Keypoints, Features and Correspondences
Deep Learning Some slides are from Prof. Andrew Ng of Stanford.
Fourier Transform of Boundaries
CSE (c) S. Tanimoto, 2007 Image Understanding
Optical flow and keypoint tracking
Wrap-up Computer Vision Spring 2019, Lecture 26
Week 3 Volodymyr Bobyr.
Presentation transcript:

Announcements  Final exam: Friday Dec 19, 8am-11am, 1 Pimentel  Closed-book, closed-notes except 1 page of notes  Grading  Your midterm score is replaced by final exam score if the final score is higher  Review sessions here TuTh ; practice final posted soon  HKN Survey 1.30pm today  Lecture on “Consciousness” by Christof Koch today 4pm 125 Li Ka Shing  Suggested reading: Tononi, G., & Koch, C. (2014). Consciousness: Here, There but Not Everywhere. arXiv:  Office hours today just pm 1

CS 188: Artificial Intelligence Computer Vision (all too briefly)* Instructor: Stuart Russell --- University of California, Berkeley

Outline  Perception generally  Approaches to vision  Edge detection  Image classification by object category 3

Perception generally  Stimulus (percept) S, world W  For some function g, S = g(W) (e.g., W=scene, g=rendering, S=image)  Can we do perception by inverting g?  W = g -1 (S) (e.g., vision = inverse graphics)  No! Percepts are usually massively ambiguous…

Percepts are evidence…  Can we just apply Bayes’ rule?  P(W|S) =  P(S|W) P(W)  P(W) is higher for students with heads attached  P(S|W) is just graphics, i.e., 1 when S=g(W), 0 otherwise  The problem is computation:  Sampling from P(W) until we find a world that agrees with S is hopeless!  Vision researchers have identified all sorts of cues in the image that suggest hypotheses about (parts of) the world  Need a combination of bottom-up and top-down reasoning 5

Top-down example: blind spot fill-in 6 Actual image Perceived image

More examples 7

Lightness constancy: Which has lighter pixels, A or B?

Vision cues and subsystems 9 scenes behaviors optical flow depth map edges objects disparity features regions image sequence tracking and data association object recognition segmentation filters edge detection matching shape from contour shape from stereo shape from motion shape from shading scene parsing

Images An image I(x,y)(t) is an array of integers [0..255] (or triples for RGB) 10 Cameras ~ 1Mpixel, human eyes ~240Mpixel x 25 fps

Edge detection Edges in image ⇐ discontinuities in scene: 1) depth 2) surface orientation 3) reflectance (surface markings) 4) illumination (shadows, etc.)

Edge detection contd. 1)Convolve image with spatially oriented filters (possibly multi-scale) E(x,y) =   f(u,v) I(x+u,y+v) du dv 2)Label above-threshold pixels with edge orientation 3)Find line segments by combining edge pixels with same orientation 12

Good, bad, and ugly 13

Illusory edges 14

Object Detection

Method 1: Deformable part-based models

Deformable part-based models  HOG features (Histogram of Oriented Gradients)  Whole-object + part-based  Model of P(part-pose | whole-pose)  Efficient dynamic programming algorithm for optimal matching  Automatically learns where the parts appear and what they look like 17

State-of-the-art results (2010) [Girschik, Felzenszwalb, McAllester] sofa bottle cat

State-of-the-art results (2010) [Girschik, Felzenszwalb, McAllester] person car horse

Method 2: Deep Learning

How Many Computers to Identify a Cat? [Le, Ng, Dean, et al, 2012] “Google Brain”

A typical deep learning architecture

Deep learning: Convolution network structure  Nodes in layer n take input from small patch in layer n-1  Weights are copied across all nodes in a given layer  translational invariance  fewer parameters => faster learning 23

Deep learning: Auto-encoding  Stacked auto-encoder for “unsupervised” training of hidden layers from huge collections of unlabeled images  Layer 0 = input image  Layer 1 = “compressed” version of Layer 0, trained with Layer 0 as output  layer 2 = “compressed” version of Layer 1, trained with Layer 1 as output  etc.  Then add a final layer for supervised training on categories  >0?     f1f1 f2f2 f3f3 f1f1 f2f2 f3f3

ImageNet Large-Scale Visual Recognition Challenge 25  1.2 million training images in 1000 categories  chosen from 14 million images in 21,841 categories  100,000 test images  “Correct” if true class is in top-5 prediction list

ILSVRC results % 6.7%

Results contd. 27

Next Time  Robotics  Interim results from EC3 contest  Where to go next to learn more  Final thoughts on the future of AI