Qian Chen, Guangtao Zhai, Xiaokang Yang, and Wenjun Zhang ISCAS,2008.

Slides:



Advertisements
Similar presentations
1 A Spectral-Temporal Method for Pitch Tracking Stephen A. Zahorian*, Princy Dikshit, Hongbing Hu* Department of Electrical and Computer Engineering Old.
Advertisements

Evaluating Color Descriptors for Object and Scene Recognition Koen E.A. van de Sande, Student Member, IEEE, Theo Gevers, Member, IEEE, and Cees G.M. Snoek,
EI San Jose, CA Slide No. 1 Measurement of Ringing Artifacts in JPEG Images* Xiaojun Feng Jan P. Allebach Purdue University - West Lafayette, IN.
Hierarchical Saliency Detection School of Electronic Information Engineering Tianjin University 1 Wang Bingren.
黃文中 Preview 2 3 The Saliency Map is a topographically arranged map that represents visual saliency of a corresponding visual scene. 4.
A Novel Method for Generation of Motion Saliency Yang Xia, Ruimin Hu, Zhenkun Huang, and Yin Su ICIP 2010.
 Introduction  Principles of context – aware saliency  Detection of context – aware saliency  Result  Application  Conclusion.
IEEE TCSVT 2011 Wonjun Kim Chanho Jung Changick Kim
Video Coding with Spatio-temporal Texture Synthesis and Edge-based inpainting Chunbo Zhu, Xiaoyan Sun, Feng Wu, and Houqiang Li ICME 2008.
Natan Jacobson, Yen-Lin Lee, Vijay Mahadevan, Nuno Vasconcelos, Truong Q. Nguyen IEEE, ICME 2010.
Golnaz Abdollahian, Cuneyt M. Taskiran, Zygmunt Pizlo, and Edward J. Delp C AMERA M OTION -B ASED A NALYSIS OF U SER G ENERATED V IDEO IEEE TRANSACTIONS.
Real-time Embedded Face Recognition for Smart Home Fei Zuo, Student Member, IEEE, Peter H. N. de With, Senior Member, IEEE.
Efficient Moving Object Segmentation Algorithm Using Background Registration Technique Shao-Yi Chien, Shyh-Yih Ma, and Liang-Gee Chen, Fellow, IEEE Hsin-Hua.
Text Detection in Video Min Cai Background  Video OCR: Text detection, extraction and recognition  Detection Target: Artificial text  Text.
3. Introduction to Digital Image Analysis
1 Static Sprite Generation Prof ︰ David, Lin Student ︰ Jang-Ta, Jiang
Robust Object Segmentation Using Adaptive Thresholding Xiaxi Huang and Nikolaos V. Boulgouris International Conference on Image Processing 2007.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 6: Low-level features 1 Computational Architectures in Biological.
Pores and Ridges: High- Resolution Fingerprint Matching Using Level 3 Features Anil K. Jain Yi Chen Meltem Demirkus.
Signal Processing Institute Swiss Federal Institute of Technology, Lausanne 1 “OBJECTIVE AND SUBJECTIVE IDENTIFICATION OF INTERESTING AREAS IN VIDEO SEQUENCES”
Face Region Based Conversational Video Coding Bing Xiong, Xiaojiu Fan, Ce Zhu, Senior Member, IEEE, Xuan Jing, and Qiang Peng.
Christian Siagian Laurent Itti Univ. Southern California, CA, USA
Extraction of Salient Contours in Color Images Vonikakis Vasilios, Ioannis Andreadis and Antonios Gasteratos Democritus University of Thrace
A fuzzy video content representation for video summarization and content-based retrieval Anastasios D. Doulamis, Nikolaos D. Doulamis, Stefanos D. Kollias.
Perceived video quality measurement Muhammad Saqib Ilyas CS 584 Spring 2005.
Xinqiao LiuRate constrained conditional replenishment1 Rate-Constrained Conditional Replenishment with Adaptive Change Detection Xinqiao Liu December 8,
An Introduction to H.264/AVC and 3D Video Coding.
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
1 Motivation Video Communication over Heterogeneous Networks –Diverse client devices –Various network connection bandwidths Limitations of Scalable Video.
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis Laurent Itti, Christof Koch, and Ernst Niebur IEEE PAMI, 1998.
CSNDSP’06 Visual Attention based Region of Interest Coding for Video - telephony Applications Nicolas Tsapatsoulis Computer Science Dept. University of.
Robust Hand Tracking with Refined CAMShift Based on Combination of Depth and Image Features Wenhuan Cui, Wenmin Wang, and Hong Liu International Conference.
Motion Object Segmentation, Recognition and Tracking Huiqiong Chen; Yun Zhang; Derek Rivait Faculty of Computer Science Dalhousie University.
REU Presentation Week 3 Nicholas Baker.  What features “pop out” in a scene?  No prior information/goal  Identify areas of large feature contrasts.
Visual Attention Accelerated Vehicle Detection in Low-Altitude Airborne Video of Urban Environment Xianbin Cao, Senior Member, IEEE, Renjun Lin, Pingkun.
黃文中 Introduction The Model Results Conclusion 2.
Joon Hyung Shim, Jinkyu Yang, and Inseong Kim
MDDSP Literature Survey Presentation Eric Heinen
Sadaf Ahamed G/4G Cellular Telephony Figure 1.Typical situation on 3G/4G cellular telephony [8]
Variation of aspect ratio Voice section Correct voice section Voice Activity Detection by Lip Shape Tracking Using EBGM Purpose What is EBGM ? Experimental.
Post-processing of JPEG image using MLP Fall 2003 ECE539 Final Project Report Data Fok.
עיבוד תמונות Image Processing. Images and Data Image Image Processing Image Image Computer Vision Data Data Computer Graphics Image.
Compression of Real-Time Cardiac MRI Video Sequences EE 368B Final Project December 8, 2000 Neal K. Bangerter and Julie C. Sabataitis.
Just Noticeable Difference Estimation For Images with Structural Uncertainty WU Jinjian Xidian University.
Creating Better Thumbnails Chris Waclawik. Project Motivation Thumbnails used to quickly select a specific a specific image from a set (when lacking appropriate.
1 Perception and VR MONT 104S, Fall 2008 Lecture 4 Lightness, Brightness and Edges.
An Improved Method Of Content Based Image Watermarking Arvind Kumar Parthasarathy and Subhash Kak 黃阡廷 2008/12/3.
Segmentation of Vehicles in Traffic Video Tun-Yu Chiang Wilson Lau.
Face Detection Using Skin Color and Gabor Wavelet Representation Information and Communication Theory Group Faculty of Information Technology and System.
Stas Goferman Lihi Zelnik-Manor Ayellet Tal Technion.
Spatio-temporal saliency model to predict eye movements in video free viewing Gipsa-lab, Grenoble Département Images et Signal CNRS, UMR 5216 S. Marat,
Journal of Visual Communication and Image Representation
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis
Flow Control in Compressed Video Communications #2 Multimedia Systems and Standards S2 IF ITTelkom.
COMPARATIVE STUDY OF HEVC and H.264 INTRA FRAME CODING AND JPEG2000 BY Under the Guidance of Harshdeep Brahmasury Jain Dr. K. R. RAO ID MS Electrical.
1 Yu Liu 1, Feng Wu 2 and King Ngi Ngan 1 1 Department of Electronic Engineering, The Chinese University of Hong Kong 2 Microsoft Research Asia, Beijing,
Learning video saliency from human gaze using candidate selection CVPR2013 Poster.
Li-Wei Kang and Chun-Shien Lu Institute of Information Science, Academia Sinica Taipei, Taiwan, ROC {lwkang, April IEEE.
Effects of Grayscale Window/Level on Breast Lesion Detectability Jeffrey Johnson, PhD a John Nafziger, PhD a Elizabeth Krupinski, PhD b Hans Roehrig, PhD.
Digital Image Processing CSC331
 Mentor : Prof. Amitabha Mukerjee Learning to Detect Salient Objects Team Members - Avinash Koyya Diwakar Chauhan.
A Fast Video Noise Reduction Method by Using Object-Based Temporal Filtering Thou-Ho (Chao-Ho) Chen, Zhi-Hong Lin, Chin-Hsing Chen and Cheng-Liang Kao.
2. Skin - color filtering.
Robustness Evaluation of Perceptual Watermarks
Face recognition using improved local texture pattern
User-Oriented Approach in Spatial and Temporal Domain Video Coding
Pei Qi ECE at UW-Madison
Reduction of blocking artifacts in DCT-coded images
Novel Multiple Spatial Watermarking Technique in Color Images
Presentation transcript:

Qian Chen, Guangtao Zhai, Xiaokang Yang, and Wenjun Zhang ISCAS,2008

 Introduction  Scalable visual sensitivity profile (SVSP)  SVSP in noise-shaping  SVSP in ROI coding of JPEG2000  SVSP in ROI scalable video coding  Conclusion

 Computational visual attention models have been developed over the last 20 years and have already facilitated various aspects of the evolution in visual communication systems.  Its important applications is to enhance the image and video compression algorithms perceptually.

Feature extraction down-sampling filter Center surround receptive field simulation Cross level addition and normalize Non-linear feature combination

 Low-level Feature Detection  Intensity channel :  Color channels :  Orientation channel :  motion channel : optical flow Gabor filter

 By iteratively down-sampling for L times of these channels,we can create pyramids for each of these channels of the frame i  Center-surround Receptive Field Simulation c ∈ [0, 8], s = c + δ, δ ∈ [−3,−2,−1, 1, 2, 3] and s is thrown away if s ∈ [0, 8].

 Cross level addition and normalize  Non-linear Feature Combination

Skin & caption detection Down-sampling filter SVSP integration Post-processing

 Skin Color Detection  The skin color area indicates the appearance of people and often attracts human attention.  Hsu’s [5] skin model  Caption Detection  Luo’s [6]

 SVSP integration  Considering the fact that human face by its nature attracts more low-level human attention, we emphasize skin map more and α = 1.5, β = 1.2 Ref.G. T. Zhai, Q. Chen, X. K. Yang, W. J. Zhang,”Scalable Visual Significance Profile Estimation”, submitted to International Conference on Acoustics, Speech, and Signal Processing, April, 2008, Las Vegas, US.

 To validate the effectiveness of the proposed model.  JND (Just-noticeable distortion/difference) :refers to the visibility threshold below which changes cannot be perceived by human.  Noise shaping is a popular way to evaluate the correctness of JND models.

 Noise-injection process is :  The proposed VSP-based JND model is :  We will compare it with Chou’s JND model [8] JNDC and the JND model we previously proposed [9] JNDY

(a)Luminance of frame 51 in president debate.(b)Chou’s JND model, PNSR=25.99 dB. (c)Yang’s JND model, PNSR=25.99 dB. (d) proposed VSP-based JND model, PNSR=25.99 dB.

 We define the arbitrary ROI a in an image as areas that take half the top values in.  To generate a rectangular ROI r, we explore a seeded region growing algorithm, seed is placed at the most saliency point in and then expands to surroundings. The stopping criterion is that the pixel value on region borders falls below 60% of the starting seed-value.

(a) Details of the most sensitive of frame 51 in president debate. (b) Details of image coded at 0.1bpp with arbitrary ROI defined in VSP, PSNR-Y=27.2dB. (c) Details of image coded at 0.1bpp with rectangular-shaped ROI defined in SVP, PSNR- Y=32.6dB. (d)Details of image coded at 0.1bpp without ROI, PSNR-Y=24.0dB.

SVSP Filter out isolated Most saliency point Sensitive region

(a) Average PSNR-Y vs. bit rate of president debate.(b) Average PSNR-Y vs. bit rate of foreman. ( c) Average PSNR-Y vs. bit rate of crew. (d)Average PSNR-Y vs. bit rate of coastguard.

 Visual comparison in saliency area of frame 60 in president debate, CIF size coded at 900 kbps. (a)without ROI (b)with SVSP defined ROI

 This paper applies the proposed computational model for scalable visual sensitivity profile (SVSP) to image/video processing.  Extensive experimental results have justified the effectiveness of the proposed SVSP model.