Analysis and classification of images based on focus

Slides:



Advertisements
Similar presentations
Automatic Photo Pop-up Derek Hoiem Alexei A.Efros Martial Hebert Carnegie Mellon University.
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
1 Building a Dictionary of Image Fragments Zicheng Liao Ali Farhadi Yang Wang Ian Endres David Forsyth Department of Computer Science, University of Illinois.
Texture Segmentation Based on Voting of Blocks, Bayesian Flooding and Region Merging C. Panagiotakis (1), I. Grinias (2) and G. Tziritas (3)
Exchanging Faces in Images SIGGRAPH ’04 Blanz V., Scherbaum K., Vetter T., Seidel HP. Speaker: Alvin Date: 21 July 2004.
CS 376b Introduction to Computer Vision 04 / 08 / 2008 Instructor: Michael Eckmann.
Binary Image Analysis: Part 2 Readings: Chapter 3: mathematical morphology region properties region adjacency 1.
Distinguishing Photographic Images and Photorealistic Computer Graphics Using Visual Vocabulary on Local Image Edges Rong Zhang,Rand-Ding Wang, and Tian-Tsong.
Example-Based Color Transformation of Image and Video Using Basic Color Categories Youngha Chang Suguru Saito Masayuki Nakajima.
Computer-Aided Diagnosis and Display Lab Department of Radiology, Chapel Hill UNC Julien Jomier, Erwann Rault, and Stephen R. Aylward Computer.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
KDD for Science Data Analysis Issues and Examples.
Track, Trace & Control Solutions © 2010 Microscan Systems, Inc. Machine Vision Tools for Solving Auto ID Applications Part 3 of a 3-part webinar series:
Multiclass object recognition
AdvisorStudent Dr. Jia Li Shaojun Liu Dept. of Computer Science and Engineering, Oakland University 3D Shape Classification Using Conformal Mapping In.
UNDERSTANDING DYNAMIC BEHAVIOR OF EMBRYONIC STEM CELL MITOSIS Shubham Debnath 1, Bir Bhanu 2 Embryonic stem cells are derived from the inner cell mass.
Texture. Texture is an innate property of all surfaces (clouds, trees, bricks, hair etc…). It refers to visual patterns of homogeneity and does not result.
Plug-in and tutorial development for GIMP- Cathy Irwin, 2004 The Development of Image Completion and Tutorial Plug-ins for the GIMP By: Cathy Irwin Supervisors:
COLOR HISTOGRAM AND DISCRETE COSINE TRANSFORM FOR COLOR IMAGE RETRIEVAL Presented by 2006/8.
Gianni Ramponi University of Trieste Images © 2002 Gonzalez & Woods Digital Image Processing Chapter 9 Morphological Image.
Stylization and Abstraction of Photographs Doug Decarlo and Anthony Santella.
In Defense of Nearest-Neighbor Based Image Classification Oren Boiman The Weizmann Institute of Science Rehovot, ISRAEL Eli Shechtman Adobe Systems Inc.
Competence Centre on Information Extraction and Image Understanding for Earth Observation 29th March 2007 Category - based Semantic Search Engine 1 Mihai.
1 A Compact Feature Representation and Image Indexing in Content- Based Image Retrieval A presentation by Gita Das PhD Candidate 29 Nov 2005 Supervisor:
CVPR2013 Poster Detecting and Naming Actors in Movies using Generative Appearance Models.
Daniel A. Keim, Hans-Peter Kriegel Institute for Computer Science, University of Munich 3/23/ VisDB: Database exploration using Multidimensional.
October Andrew C. Gallagher, Jiebo Luo, Wei Hao Improved Blue Sky Detection Using Polynomial Model Fit Andrew C. Gallagher, Jiebo Luo, Wei Hao Presented.
Effective Automatic Image Annotation Via A Coherent Language Model and Active Learning Rong Jin, Joyce Y. Chai Michigan State University Luo Si Carnegie.
Low level Computer Vision 1. Thresholding 2. Convolution 3. Morphological Operations 4. Connected Component Extraction 5. Feature Extraction 1.
Demosaicking for Multispectral Filter Array (MSFA)
Wonjun Kim and Changick Kim, Member, IEEE
Learning Photographic Global Tonal Adjustment with a Database of Input / Output Image Pairs.
1 Mathematic Morphology used to extract image components that are useful in the representation and description of region shape, such as boundaries extraction.
Machine Vision ENT 273 Hema C.R. Binary Image Processing Lecture 3.
Chapter 6 Skeleton & Morphological Operation. Image Processing for Pattern Recognition Feature Extraction Acquisition Preprocessing Classification Post.
Morphological Image Processing (Chapter 9) CSC 446 Lecturer: Nada ALZaben.
ICCV 2009 Tilke Judd, Krista Ehinger, Fr´edo Durand, Antonio Torralba.
Big data classification using neural network
Data Mining for Surveillance Applications Suspicious Event Detection
Medical Image Analysis
Year 10 GCSE Scheme of work
Yearbook/Photography Grade 9 Milroy
Instance Based Learning
Can Computer Algorithms Guess Your Age and Gender?
Visual Perception Principles
Mean Shift Segmentation
Brain Hemorrhage Detection and Classification Steps
Transformations Learning Target: I will be able to translate, reflect, rotate, and dilate figures.
Using Tensorflow to Detect Objects in an Image
Data Mining for Surveillance Applications Suspicious Event Detection
Ying Dai Faculty of software and information science,
Automated Recognition of Corn Embryos for Selective Breeding
Brief Review of Recognition + Context
Binary Image processing بهمن 92
Use of Python Scripts to Compare and Contrast Photo Images
Aline Martin ECE738 Project – Spring 2005
Using Tensorflow to Detect Objects in an Image
Department of Computer Engineering
Binary Image Analysis: Part 2 Readings: Chapter 3:
Karl R Gegenfurtner, Jochem Rieger  Current Biology 
Joshua Kahn, Scott Wiese ECE533 – Fall 2003 December 12, 2003
Data Mining for Surveillance Applications Suspicious Event Detection
Outline A. M. Martinez and A. C. Kak, “PCA versus LDA,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp , 2001.
Machine Learning for Visual Scene Classification with EEG Data
A Block Based MAP Segmentation for Image Compression
Analyzing social media data to monitor public health trends
A Novel Smoke Detection Method Using Support Vector Machine
MultiModality Registration using Hilbert-Schmidt Estimators
Morphological Filters Applications and Extension Morphological Filters
Presentation transcript:

Analysis and classification of images based on focus Quinton Smith Mentored by Dr. David Doria FlickrTM, 500pxTM, YogileTM and ZenfolioTM. These odd terms have one thing in common; they are all sites designed for photo sharing by photographers. Within each site there is a large variety of different photos - from portraits to landscapes. To allow users to find what they are looking for, these photos must be categorized. But how are these images placed in these categories? Typically the user manually assigns a label to each image. Automating this process would expedite this manual process, along with ensuring consistency in the categorization between photographers. To automate the categorization, features in the images must be quantified so that an algorithm can distinguish between classes of images using machine learning (Han, Qi, 2005). This has been done with details like texture (Haralick, Shanmugan, Dimstein, 1973). In this work, it is proposed that the use of analysis of the regions of the image that are in focus to classify images. In the majority of landscape photos, the photographer tries to keep the whole scene in focus so that nothing is blurry, while portraits tend to focus on the eyes of the intended target and leave the rest of the image out of focus. Introduction Methods and Materials (cont.) focused area of an image. To compare images, values derived from focus, such as the percent of focused pixels, and the number of blobs of focus pixels, can be compared between images. In order to study these regions of focus, techniques were drawn from morphology – the study of shape. The shape of the blobs of pixels were studied and changed. It can be used to increase the size of blobs so that small, one pixel gaps between blobs, which can be considered connected, are automatically filled. A process like this is called a morphological dilation, similar to when one’s pupil dilates, it becomes larger. Opposite to dilations, morphology can also be used to shrink a blob so that it no longer connects to other blobs. This is called a morphological erosion and is useful for cutting of unimportant pixels from the focused area. If there is a small bridge between two blobs that should not be connected, then a process called a morphological closing is performed. A closing allows small, one to two pixel thick bridges between blobs to be cut without actually decreasing the size of the blobs. Application of morphology created the focus maps in Figure 2. Figure 2: The progression of the code as it finds the most focused part of an image, from the original, left, to the focus blobs, middle, to the most focused blob in the image, right, for a landscape (Golightly, 2014), top, and a portrait, bottom. After determining the most focused area of the image, the next step was to collect data and build a classifier. The size of the largest in-focus region, as well as the total number of focus blobs that were present in the image were obtained. This data was collected for 20 training images, 10 from each category of portrait and landscape. Averages were found, and used to categorize a set of 100 test images, 50 from each category. By creating a linear discriminant line between the two average points, and plotting the test points on the same graph, one can visually see the category that each image would be placed in. The algorithm did this by finding the geometric distance (Graph 1) between a test image’s data points, and the averages of each category. The shortest distance was considered the image’s category. Results Graph 2: These two graphs show the number of correctly categorized images versus those incorrectly categorized for the two categories. The landscape photos had greater success at being categorized. Graph 1: Scatterplot containing points consisting of each image’s area and number of blobs, as well as the average for each image type. The closer a point is to the average, the smaller the geometric mean. Points to the left of the black line were considered portraits, those on the right were landscapes. An image’s focus is the area of the image that is “sharp” as compared to the rest of the image. What this means is that this area has the most detail in the image, and is not blurry in any way. The differences in the colors of the surrounding pixels can be considered to be detail, and thus could be used to determine the level of focus of a pixel relative to its surroundings. The variance, or spread, of the pixel values is a good estimate of the amount of detail at a point in an image. Using the Python programming language, and the built-in libraries SciPy and NumPy, the variance of each pixel with its surrounding pixels was found and used to create a new image, a map of the most focused areas of the image (Figure 1). This map can be used to find the most detailed and thus most Figure 1: Original images and the focus maps for a landscape, top, and a portrait, bottom. Methods and Materials Of the 100 images in our experiment, 71 of them were correctly categorized. This indicates that focus is a feasible feature to use to categorize images automatically. A major discrepancy is that the Landscape category had a 98% success rate, while the Portrait category had a 44% success rate (Graph 2). This could have been due to the wide variety of images that are present in the category of Portrait. Using additional features computed from the focus values would be beneficial. Additionally, using larger sets of training data for the learning phase of the algorithm would certainly improve the robustness of the system. Conclusions References Golightly, C. (Photographer). (2014, November 14). Quiraing Tree, Isle of Skye. [digital image]. Retrieved from https://www.flickr.com/photos/cgolightly/16490546877/ Han, Y., & Qi, X. (2005). Machine-learning-based image categorization. In Image Analysis and Recognition (pp. 585-592). Springer Berlin Heidelberg. Haralick, R. M., Shanmugam, K., & Dinstein, I. H. (1973). Textural features for image classification. Systems, Man and Cybernetics, IEEE Transactions on, (6), 610-621. Acknowledgements I owe my success to my mentor, Dr. David Doria, for his continued support and guidance through the entirety of this project. I also thank my faculty advisor, Mr. Davis, as well as Mr. Evans.