Information Extraction from Cricket Videos Syed Ahsan Ishtiaque Kumar Srijan.

Slides:



Advertisements
Similar presentations
Applications of one-class classification
Advertisements

QR Code Recognition Based On Image Processing
Road-Sign Detection and Recognition Based on Support Vector Machines Saturnino, Sergio et al. Yunjia Man ECG 782 Dr. Brendan.
Computational Biology, Part 23 Biological Imaging II Robert F. Murphy Copyright  1996, 1999, All rights reserved.
Recovering Human Body Configurations: Combining Segmentation and Recognition Greg Mori, Xiaofeng Ren, and Jitentendra Malik (UC Berkeley) Alexei A. Efros.
Image Analysis Phases Image pre-processing –Noise suppression, linear and non-linear filters, deconvolution, etc. Image segmentation –Detection of objects.
Eyes for Relighting Extracting environment maps for use in integrating and relighting scenes (Noshino and Nayar)
Automatic Soccer Video Analysis and Summarization
Image Segmentation Image segmentation (segmentace obrazu) –division or separation of the image into segments (connected regions) of similar properties.
 Any time you half press the shutter button, the light meter activates.  As we know, it measures the light in your scene, and calculates a shutter speed.
COLORCOLOR A SET OF CODES GENERATED BY THE BRAİN How do you quantify? How do you use?
Foreground Modeling The Shape of Things that Came Nathan Jacobs Advisor: Robert Pless Computer Science Washington University in St. Louis.
Different Tracking Techniques  1.Gaussian Mixture Model:  1.Construct the model of the Background.  2.Given sequence of background images find the.
Hue-Grayscale Collaborating Edge Detection & Edge Color Distribution Space Jiqiang Song March 6 th, 2002.
Lecture 5 Template matching
Text Detection in Video Min Cai Background  Video OCR: Text detection, extraction and recognition  Detection Target: Artificial text  Text.
A Study of Approaches for Object Recognition
EE663 Image Processing Edge Detection 2 Dr. Samir H. Abdul-Jauwad Electrical Engineering Department King Fahd University of Petroleum & Minerals.
Processing Digital Images. Filtering Analysis –Recognition Transmission.
Ensemble Tracking Shai Avidan IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE February 2007.
CS223B Assignment 1 Recap. Lots of Solutions! 37 Groups Many different approaches Let’s take a peek at all 37 results on one image from the test set.
Multiple Human Objects Tracking in Crowded Scenes Yao-Te Tsai, Huang-Chia Shih, and Chung-Lin Huang Dept. of EE, NTHU International Conference on Pattern.
A Robust Scene-Change Detection Method for Video Segmentation Chung-Lin Huang and Bing-Yao Liao IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY.
CSSE463: Image Recognition Day 30 Due Friday – Project plan Due Friday – Project plan Evidence that you’ve tried something and what specifically you hope.
CS223B Homework 1 Results. Considered 2 Metrics Raw score –Number of pixels in error Weighted score –Car pixels weighted more heavily than non-car pixels.
CS 223B Assignment 1 Help Session Dan Maynes-Aminzade.
CS 376b Introduction to Computer Vision 04 / 01 / 2008 Instructor: Michael Eckmann.
Fitting a Model to Data Reading: 15.1,
CSE 291 Final Project: Adaptive Multi-Spectral Differencing Andrew Cosand UCSD CVRR.
Shadow Removal Seminar
Facial Features Extraction Amit Pillay Ravi Mattani Amit Pillay Ravi Mattani.
A Novel 2D To 3D Image Technique Based On Object- Oriented Conversion.
Robust estimation Problem: we want to determine the displacement (u,v) between pairs of images. We are given 100 points with a correlation score computed.
A Vision-Based System that Detects the Act of Smoking a Cigarette Xiaoran Zheng, University of Nevada-Reno, Dept. of Computer Science Dr. Mubarak Shah,
Real-Time Face Detection and Tracking Using Multiple Cameras RIT Computer Engineering Senior Design Project John RuppertJustin HnatowJared Holsopple This.
3D Scene Models Object recognition and scene understanding Krista Ehinger.
Facial Recognition CSE 391 Kris Lord.
FEATURE EXTRACTION FOR JAVA CHARACTER RECOGNITION Rudy Adipranata, Liliana, Meiliana Indrawijaya, Gregorius Satia Budhi Informatics Department, Petra Christian.
CSSE463: Image Recognition Day 30 This week This week Today: motion vectors and tracking Today: motion vectors and tracking Friday: Project workday. First.
Kumar Srijan ( ) Syed Ahsan( ). Problem Statement To create a Neural Networks based multiclass object classifier which can do rotation,
Lecture 6-1CS251: Intro to AI/Lisp II I can see clearly now May 4th, 1999.
By Doğaç Başaran & Erdem Yörük
Automated Face Detection Peter Brende David Black-Schaffer Veni Bourakov.
September 23, 2014Computer Vision Lecture 5: Binary Image Processing 1 Binary Images Binary images are grayscale images with only two possible levels of.
September 5, 2013Computer Vision Lecture 2: Digital Images 1 Computer Vision A simple two-stage model of computer vision: Image processing Scene analysis.
Digital Image Processing CCS331 Relationships of Pixel 1.
Joon Hyung Shim, Jinkyu Yang, and Inseong Kim
December 9, 2014Computer Vision Lecture 23: Motion Analysis 1 Now we will talk about… Motion Analysis.
CS654: Digital Image Analysis Lecture 25: Hough Transform Slide credits: Guillermo Sapiro, Mubarak Shah, Derek Hoiem.
Fourier Descriptors For Shape Recognition Applied to Tree Leaf Identification By Tyler Karrels.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
Autonomous Robots Vision © Manfred Huber 2014.
Using Cross-Media Correlation for Scene Detection in Travel Videos.
CS 376b Introduction to Computer Vision 03 / 31 / 2008 Instructor: Michael Eckmann.
1 Mathematic Morphology used to extract image components that are useful in the representation and description of region shape, such as boundaries extraction.
Digital Image Processing CSC331
Computer Vision Computer Vision based Hole Filling Chad Hantak COMP December 9, 2003.
Over the recent years, computer vision has started to play a significant role in the Human Computer Interaction (HCI). With efficient object tracking.
Content Based Coding of Face Images
IMAGE PROCESSING Tadas Rimavičius.
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
Motion and Optical Flow
Fitting Curve Models to Edges
Image Segmentation Image analysis: First step:
CSSE463: Image Recognition Day 30
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
CSSE463: Image Recognition Day 30
Midterm Exam Closed book, notes, computer Similar to test 1 in format:
CSSE463: Image Recognition Day 30
Introduction to Artificial Intelligence Lecture 22: Computer Vision II
Presentation transcript:

Information Extraction from Cricket Videos Syed Ahsan Ishtiaque Kumar Srijan

Problems Statement There can be many challenges like face detection and recognition, boundary detection, separating out the ad and the cricket part and many more. The problems we addressed are – Shot Transition Detection – Crowd Detection – Pitch Detection – Bowling Side Detection – Ball by Ball segmentation – Summary Making

Shot Transition Detection A shot is defined as a sequence of frames captured by the camera in a contiguous way, without interruptions. Shots are of two types – Hard Transition – Soft Transition Following are our solution to this problem

Normalized Cross Correlation Threshold the correlation between every 20 th frames – By checking the correlation between every 20 th frame, and if its greater than certain threshold assuming that some transition has occurred, then checking for correlation between every frame in that interval, if any correlation value is greater than another threshold, then Hard Transition else Soft Transition Results and Inferences – The results were satisfactory for Hard Transition, but got false matches for soft transition in the case when there was actually no transition and the camera was moving very fast, as the correlation value for the K th and K+20 th frame crossed the threshold. – Also there was a problem that the correlation between the frames of crowd was generally low, so generally every frame of crowd was classified as a shot transition

Histogram Based Histogram based Detection – By calculating the Histogram for all the bands or for gray band, and thresholding for sharp change of values we can detect for a transition. Results and Inferences – This gave good results, the problem with this approach was that it was not taking the position of pixels into account

Every frame here is a hard transition, but the Histogram approach will not be able to detect this Hard Transition

Difference of NCC Difference of correlation values between consecutive frames – In this approach we checked the difference of correlation, and if it is greater than a certain threshold we declare it as a hard transition. – Also blurring of the frames improved correlation and hence the results Results and Inferences – This also gave very good results, and by this approach the problem of classification of crowd frames generally as shot was also solved, as although the correlation between the frames of crowd was low, but difference of correlation between the two low correlation frame groups was not that high, so it solved the problem.

Crowd Detection Detecting the scenes in which crowd is present or is in focus.

Histogram Based Histogram based – In a cricket video, we usually get two kinds of frames, one with field where there are very narrow range of colors present and of crowd where many different kinds of colors are present. – The histogram of crowd will be flat and will cover the whole range, whereas other scenes will have histograms which will be concentrated in a narrow range. Results and Inferences – The results were not that encouraging. A single band or combination of bands, and also the range of values within a histogram could not be determined for histogram to be constructed and analyzed.

Edge Based Detection Edge Based – We observed that energy profile of the frame containing crowd is distinctly higher than those of not containing crowd, so we used canny edge detector and build a edge map of all the frames, now comparing the energy profiles of the edge map solved our problem Results and Inferences – The results were very good

Pitch Detection Finding whether the frame is showing a pitch or not and determining its position in the frame.

Template Based solution Template Matching – We took a narrow horizontal strip, from the middle of a frame showing the pitch, as the template. For each frame in the video, we looked whether the strip can match somewhere near the middle of the frame. We used normalized correlation and square difference error as the matching standards. Results and Inferences – The results were not that good, had few false positives in the frame showing the field. – The problem with template matching was that it did not address the variation of color of the pitch under different lighting conditions.

HSI space based As template matching didn’t look into the variation of lighting conditions The idea is that any frame not having the pitch will not be able to match in all the planes, viz. Hue, Saturation and Intensity, even if we take sufficiently large margins so that all the frames showing the pitch are correctly classified For each frame, we broke it into HSI space and applied thresholding on all the planes according to the ranges calculated before. We dilated each of the planes for filling the small holes and then took the intersection of the resulting planes. We again dilated and eroded the resulting plane to remove the holes. After that we took the distance transform of the image and searched for the maximum value in that. Thresholding this maximum value worked as the blob detection

Results The results were very good Intensity SaturationHue IntersectionBlob Original

Bowling Side Detection To detect from which side the bowler bowled. Solution – Erosion of the blob obtained above shows the skeleton of the pitch which is available. – Whenever a bowler bowls the skeleton becomes “L” shaped due to the occlusion caused by the bowler. – By detecting the orientation of “L” we can easily determine the side from which the ball was being bowled

Results This method gave very good results Hue Intersection Intensity SkeletonOriginal Saturation

Difficulties Later in the video as the part of the pitch was covered with shadows, the “L” shape got truncated and it was hard to determine the bowling side. SkeletonOriginal Truncated “L” Bowling side not detected, but pitch was still detected

Ball by Ball Segmentation To segment the video into deliveries Solution – The detection of “L” can also help in knowing when the ball is being bowled. When the skeleton of the pitch changes from rectangle to L shaped then a ball was bowled. We also made sure that there was a gap of a few seconds between the deliveries by setting a timer.

Results This was not classified as a delivery due to the timer Classified as a delivery Again Classified as delivery, as timer got reset

Summary Making To generate a meaningful summary out of the video. Solution – By a simple assumption that crowd is displayed only when there is six, four, wicket or sometimes ads. – So, every time crowd is detected we can go back few seconds and take that part as highlight. – Now sometimes crowd is displayed after an ad break, so now pitch detection helps here, every time we go back we can check for the presence of pitch, if it is detected then it is not from the ad break and we can include that part in our summary and if it is not we can skip that.