Human Vision and Cameras CS 691 E Spring 2016. Outline Human vision system Human vision for computer vision Cameras and image formation Projection geometry.

Slides:



Advertisements
Similar presentations
The Camera : Computational Photography Alexei Efros, CMU, Fall 2006.
Advertisements

776 Computer Vision Jan-Michael Frahm, Enrique Dunn Spring 2013.
CSE 473/573 Computer Vision and Image Processing (CVIP) Ifeoma Nwogu Lecture 4 – Image formation(part I)
CS 691 Computational Photography Instructor: Gianfranco Doretto 3D to 2D Projections.
Computer Vision CS 776 Spring 2014 Cameras & Photogrammetry 1 Prof. Alex Berg (Slide credits to many folks on individual slides)
Paris town hall.
The Camera CS194: Image Manipulation & Computational Photography Alexei Efros, UC Berkeley, Fall 2014.
Projection Readings Szeliski 2.1. Projection Readings Szeliski 2.1.
Image formation and cameras CSE P 576 Larry Zitnick Many slides courtesy of Steve Seitz.
Cameras. Overview The pinhole projection model Qualitative properties Perspective projection matrix Cameras with lenses Depth of focus Field of view Lens.
Advanced Computer Vision Cameras, Lenses and Sensors.
Cameras (Reading: Chapter 1) Goal: understand how images are formed Camera obscura dates from 15 th century Basic abstraction is the pinhole camera Perspective.
Announcements. Projection Today’s Readings Nalwa 2.1.
Lecture 5: Projection CS6670: Computer Vision Noah Snavely.
Announcements Mailing list (you should have received messages) Project 1 additional test sequences online Talk today on “Lightfield photography” by Ren.
Lecture 5: Cameras and Projection CS6670: Computer Vision Noah Snavely.
Computer Vision - A Modern Approach
CSCE641: Computer Graphics Image Formation Jinxiang Chai.
Announcements Mailing list Project 1 test the turnin procedure *this week* (make sure it works) vote on best artifacts in next week’s class Project 2 groups.
Lecture 4a: Cameras CS6670: Computer Vision Noah Snavely Source: S. Lazebnik.
Lecture 12: Projection CS4670: Computer Vision Noah Snavely “The School of Athens,” Raphael.
Lecture 6: Image Warping and Projection CS6670: Computer Vision Noah Snavely.
Today: The Camera. Overview The pinhole projection model Qualitative properties Perspective projection matrix Cameras with lenses Depth of focus Field.
Single-view Metrology and Camera Calibration Computer Vision Derek Hoiem, University of Illinois 02/26/15 1.
Computer Vision Cameras, lenses and sensors Cosimo Distante Introduction to Image Processing Image.
The Camera : Computational Photography Alexei Efros, CMU, Fall 2008.
CSCE 641: Computer Graphics Image Formation & Plenoptic Function Jinxiang Chai.
Cameras, lenses, and calibration
Image formation and cameras
Perspective projection
Cameras CSE 455, Winter 2010 January 25, 2010.
EEM 561 Machine Vision Week 10 :Image Formation and Cameras
Building a Real Camera.
Image formation How are objects in the world captured in an image?
Northeastern University, Fall 2005 CSG242: Computational Photography Ramesh Raskar Mitsubishi Electric Research Labs Northeastern University September.
776 Computer Vision Jan-Michael Frahm Fall Camera.
776 Computer Vision Jan-Michael Frahm Spring 2012.
CPSC 641: Computer Graphics Image Formation Jinxiang Chai.
Recap from Wednesday Two strategies for realistic rendering capture real world data synthesize from bottom up Both have existed for 500 years. Both are.
776 Computer Vision Jan-Michael Frahm, Enrique Dunn Spring 2013.
COMPUTER VISION D10K-7C02 CV02: Camera Dr. Setiawan Hadi, M.Sc.CS. Program Studi S-1 Teknik Informatika FMIPA Universitas Padjadjaran.
Intelligent Vision Systems Image Geometry and Acquisition ENT 496 Ms. HEMA C.R. Lecture 2.
776 Computer Vision Jan-Michael Frahm Fall Last class.
EECS 274 Computer Vision Cameras.
Miguel Tavares Coimbra
CSE 185 Introduction to Computer Vision Cameras. Camera models –Pinhole Perspective Projection –Affine Projection –Spherical Perspective Projection Camera.
Projection Readings Nalwa 2.1 Szeliski, Ch , 2.1
Intelligent Vision Systems Image Geometry and Acquisition ENT 496 Ms. HEMA C.R. Lecture 2.
Instructor: Mircea Nicolescu Lecture 3 CS 485 / 685 Computer Vision.
Announcements Project 1 grading session this Thursday 2:30-5pm, Sieg 327 –signup ASAP:signup –10 minute slot to demo your project for a TA »have your program.
Lecture 14: Projection CS4670 / 5670: Computer Vision Noah Snavely “The School of Athens,” Raphael.
Lecture 18: Cameras CS4670 / 5670: Computer Vision KavitaBala Source: S. Lazebnik.
CSE 185 Introduction to Computer Vision
PNU Machine Vision Lecture 2a: Cameras Source: S. Lazebnik.
CS5670: Computer Vision Lecture 9: Cameras Noah Snavely
Jan-Michael Frahm Fall 2016
The Camera : Computational Photography
Announcements Project 1 Project 2
CSE 185 Introduction to Computer Vision
Jan-Michael Frahm Fall 2016
Lecture 13: Cameras and geometry
EECS 274 Computer Vision Cameras.
The Camera : Computational Photography
Announcements Midterm out today Project 1 demos.
Human Vision and Cameras
Projection Readings Nalwa 2.1.
Projection Readings Szeliski 2.1.
Announcements Midterm out today Project 1 demos.
The Camera.
Presentation transcript:

Human Vision and Cameras CS 691 E Spring 2016

Outline Human vision system Human vision for computer vision Cameras and image formation Projection geometry Reading: Chapter 1 (F&P book)

Human Eyes The human eye is the organ which gives us the sense of sight, allowing us to observe and learn more about the surrounding world than we do with any of the other four senses. The eye allows us to see and interpret the shapes, colors, and dimensions of objects in the world by processing the light they reflect or emit. The eye is able to detect bright light or dim light, but it cannot sense objects when light is absent.

Anatomy of the Human Eye

Some Concepts Retina – The retina is the innermost layer of the eye and is comparable to the film inside of a camera. It is composed of nerve tissue which senses the light entering the eye.

Concepts The macula lutea is the small, yellowish central portion of the retina. It is the area providing the clearest, most distinct vision. The center of the macula is called the fovea centralis, an area where all of the photoreceptors are cones; there are no rods in the fovea. Learn more concepts at

Rods and Cones The retina contains two types of photoreceptors, rods and cones. The rods are more numerous, some 120 million, and are more sensitive than the cones. However, they are not sensitive to color. The 6 to 7 million cones provide the eye's color sensitivity and they are much more concentrated in the central yellow spot known as the macula.

The Electromagnetic Spectrum

Human Vision System We do not “see” with our eyes, but with our brains

Human Vision for Computer Vision "Human vision is vastly better at recognition than any of our current computer systems, so any hints of how to proceed from biology are likely to be very useful." – David Lowe

Feedforward Processing Thorpe & Fabre-Thorpe 2001 LGN -- LGN -- The lateral geniculate nucleus V1 – The primary visual cortex V2 – Visual area V2 IT – Inferior temporal cortex

A Hierarchical Model Contains alternating layers called simple and complex cell units creating increasing complexity: [Hubel & Wiesel 1962, Riesenhuber & Poggio 1999, Serre et al. 2007] – Simple Cell (linear operation) – Selective – Complex Cell (nonlinear operation) – Invariant

S1 Layer Being selective Applying Gabor filters to the input image, setting parameters to match what is known about the primate visual system 8 bands and 4 orientations

Models with Four Layers (S1 C1 S2 C2) Powerful for object category recognition, Representative works include – Riesenhuber & Poggio ’99, Serre et al. ’05, ’07, Mutch & Lowe ’06 Air planes Motor bikes Faces Cars

S1C1 Gabor filtering Pixelwise MAX Spatial pooling MAX p1p1 p2p2 pNpN Pre-learned prototypes … v1v1 v2v2 vNvN S2 C2 Template matching Global MAX Off-line On-line Serre et al.’s model (PAMI’07)

Biologically Inspired Models T. Serre, L. Wolf, S. Bileschi, M. Riesenhuber, and T. Poggio. Robust object recognition with cortex-like mechanisms. IEEE Trans. Pattern Anal. Mach. Intell., 29(3):411–426, G. Guo, G. Mu, Y. Fu, and T. S. Huang. Human age estimation using bioinspired features. In IEEE CVPR, Any other biologically inspired models? Optional Homework: read Serre et al.’ paper or other newer models, discuss and compare those models, and think about the advantages and disadvantages

Cameras

Image Formation Digital Camera Film Alexei Efros’ slide

How do we see the world? Let’s design a camera – Idea 1: put a piece of film in front of an object – Do we get a reasonable image? Slide by Steve Seitz

Pinhole camera Add a barrier to block off most of the rays – This reduces blurring – The opening known as the aperture Slide by Steve Seitz

Pinhole camera model Pinhole model: – Captures pencil of rays – all rays through a single point – The point is called Center of Projection (focal point) – The image is formed on the Image Plane Slide by Steve Seitz

Figures © Stephen E. Palmer, 2002 Dimensionality Reduction Machine (3D to 2D) 3D world2D image What have we lost? Angles Distances (lengths) Slide by A. Efros

Projection properties Many-to-one: any points along same ray map to same point in image Points → points – But projection of points on focal plane is undefined Lines → lines (collinearity is preserved) – But line through focal point projects to a point Planes → planes (or half-planes) – But plane through focal point projects to line

Projection properties Parallel lines converge at a vanishing point – Each direction in space has its own vanishing point – But parallels parallel to the image plane remain parallel – All directions in the same plane have vanishing points on the same line How do we construct the vanishing point/line?

Vanishing points each set of parallel lines meets at a different point – The vanishing point for this direction Sets of parallel lines on the same plane lead to collinear vanishing points. – The line is called the horizon for that plane Good ways to spot faked images – scale and perspective don’t work – vanishing points behave badly

Distant objects are smaller Size is inversely proportional to distance.

Perspective distortion What does a sphere project to?

Shrinking the aperture Why not make the aperture as small as possible? – Less light gets through – Diffraction effects… Slide by Steve Seitz

Shrinking the aperture

The reason for lenses

Adding a lens A lens focuses light onto the film – Rays passing through the center are not deviated Slide by Steve Seitz

Adding a lens A lens focuses light onto the film – Rays passing through the center are not deviated – All parallel rays converge to one point on a plane located at the focal length f Slide by Steve Seitz focal point f

Adding a lens A lens focuses light onto the film – There is a specific distance at which objects are “in focus” other points project to a “circle of confusion” in the image “circle of confusion” Slide by Steve Seitz

Thin lens formula f D D’ Frédo Durand’s slide

Thin lens formula f D D’ Similar triangles everywhere! Frédo Durand’s slide

Thin lens formula f D D’ Similar triangles everywhere! y’ y y’/y = D’/D Frédo Durand’s slide

Thin lens formula f D D’ Similar triangles everywhere! y’ y y’/y = D’/D y’/y = (D’-f)/f Frédo Durand’s slide

Thin lens formula f D D’ 1 D 11 f += Any point satisfying the thin lens equation is in focus. Frédo Durand’s slide

Depth of Field Slide by A. Efros

How can we control the depth of field? Changing the aperture size affects depth of field – A smaller aperture increases the range in which the object is approximately in focus – But small aperture reduces amount of light – need to increase exposure Slide by A. Efros

Varying the aperture Large aperture = small DOFSmall aperture = large DOF Slide by A. Efros

Nice Depth of Field effect Source: F. Durand

Field of View (Zoom) Slide by A. Efros

Field of View (Zoom) Slide by A. Efros

f Field of View Smaller FOV = larger Focal Length Slide by A. Efros f FOV depends on focal length and size of the camera retina

Field of View / Focal Length Large FOV, small f Camera close to car Small FOV, large f Camera far from the car Sources: A. Efros, F. Durand

Same effect for faces standard wide-angletelephoto Source: F. Durand

Source: Hartley & Zisserman Approximating an affine camera

Lens systems A good camera lens may contain 15 elements and cost a thousand dollars The best modern lenses may contain aspherical elements

Lens Flaws: Chromatic Aberration Lens has different refractive indices for different wavelengths: causes color fringing Near Lens Center Near Lens Outer Edge

Lens flaws: Spherical aberration Spherical lenses don’t focus light perfectly Rays farther from the optical axis focus closer

Lens flaws: Vignetting

No distortionPin cushionBarrel Radial Distortion – Caused by imperfect lenses – Deviations are most noticeable for rays that pass through the edge of the lens

Digital camera A digital camera replaces film with a sensor array – Each cell in the array is light-sensitive diode that converts photons to electrons – Two common types Charge Coupled Device (CCD) Complementary metal oxide semiconductor (CMOS) – Slide by Steve Seitz

CCD vs. CMOS CCD: transports the charge across the chip and reads it at one corner of the array. An analog-to-digital converter (ADC) then turns each pixel's value into a digital value by measuring the amount of charge at each photosite and converting that measurement to binary form CMOS: uses several transistors at each pixel to amplify and move the charge using more traditional wires. The CMOS signal is digital, so it needs no ADC.

Color sensing in camera: Color filter array Source: Steve Seitz Estimate missing components from neighboring values (demosaicing) Why more green? Bayer grid Human Luminance Sensitivity Function

Problem with demosaicing: color moire Slide by F. Durand

The cause of color moire detector Fine black and white detail in image misinterpreted as color information Slide by F. Durand

Color sensing in camera: Prism Requires three chips and precise alignment More expensive CCD(B) CCD(G) CCD(R)

Color sensing in camera: Foveon X3 Source: M. Pollefeys CMOS sensor Takes advantage of the fact that red, blue and green light penetrate silicon to different depths better image quality

Issues with digital cameras Noise low light is where you most notice noisenoise light sensitivity (ISO) / noise tradeoff stuck pixels Resolution: Are more megapixels better? requires higher quality lens noise issues In-camera processing oversharpening can produce haloshalos RAW vs. compressed file size vs. quality tradeoff Blooming charge overflowing into neighboring pixelsoverflowing Color artifacts purple fringing from microlenses, artifacts from Bayer patterns purple fringingBayer patterns white balance More info online: – – Slide by Steve Seitz

Historical context Pinhole model: Mozi ( BCE), Aristotle ( BCE) Principles of optics (including lenses): Alhacen ( CE) Camera obscura: Leonardo da Vinci ( ), Johann Zahn ( ) First photo: Joseph Nicephore Niepce (1822) Daguerréotypes (1839) Photographic film (Eastman, 1889) Cinema (Lumière Brothers, 1895) Color Photography (Lumière Brothers, 1908) Television (Baird, Farnsworth, Zworykin, 1920s) First consumer camera with CCD: Sony Mavica (1981) First fully digital camera: Kodak DCS100 (1990) Niepce, “La Table Servie,” 1822 CCD chip Alhacen’s notes

Modeling projection The coordinate system – We will use the pinhole model as an approximation – Put the optical center (O) at the origin – Put the image plane ( Π’ ) in front of O x y z Source: J. Ponce, S. Seitz

x y z Modeling projection Projection equations – Compute intersection with Π’ of ray from P = (x,y,z) to O – Derived using similar triangles Source: J. Ponce, S. Seitz We get the projection by throwing out the last coordinate:

Homogeneous coordinates Is this a linear transformation? Trick: add one more coordinate: homogeneous image coordinates homogeneous scene coordinates Converting from homogeneous coordinates no—division by z is nonlinear Slide by Steve Seitz

divide by the third coordinate Perspective Projection Matrix Projection is a matrix multiplication using homogeneous coordinates:

divide by the third coordinate Perspective Projection Matrix Projection is a matrix multiplication using homogeneous coordinates: In practice: lots of coordinate transformations… World to camera coord. trans. matrix (4x4) Perspective projection matrix (3x4) Camera to pixel coord. trans. matrix (3x3) = 2D point (3x1) 3D point (4x1)

Weak perspective Assume object points are all at same depth -z 0

Orthographic Projection Special case of perspective projection – Distance from center of projection to image plane is infinite – Also called “parallel projection” – What’s the projection matrix? Image World Slide by Steve Seitz

Pros and Cons of These Models Weak perspective (including orthographic) has simpler mathematics – Accurate when object is small relative to its distance. – Most useful for recognition. Perspective is much more accurate for scenes. – Used in structure from motion. When accuracy really matters, we must model the real camera – Use perspective projection with other calibration parameters (e.g., radial lens distortion)