Sound Applications Advanced Multimedia Tamara Berg.

Slides:



Advertisements
Similar presentations
Matthias Gruhne, Page 1 Fraunhofer Institut Integrierte Schaltungen Robust Audio Identification for Commercial Applications Matthias.
Advertisements

Tamara Berg Advanced Multimedia
ECE 8443 – Pattern Recognition EE 3512 – Signals: Continuous and Discrete Objectives: Response to a Sinusoidal Input Frequency Analysis of an RC Circuit.
A Musical Data Mining Primer CS235 – Spring ’03 Dan Berger
LECTURE Copyright  1998, Texas Instruments Incorporated All Rights Reserved Use of Frequency Domain Telecommunication Channel |A| f fcfc Frequency.
Franz de Leon, Kirk Martinez Web and Internet Science Group  School of Electronics and Computer Science  University of Southampton {fadl1d09,
1 Copyright 2011 G.Tzanetakis Music Information Retrieval George Tzanetakis Associate Professor, IEEE Senior Member.
Typical transducer Microphone Microprocessor Typical signal generator circuit Signal intensity Time Sound Parameter Signal intensity Time Signal intensity.
Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification.
SIMS-201 Characteristics of Audio Signals Sampling of Audio Signals Introduction to Audio Information.
1 / 22 Issues in Text Similarity and Categorization Jordan Smith – MUMT 611 – 27 March 2008.
LYRIC-BASED ARTIST NETWORK METHODOLOGY Derek Gossi CS 765 Fall 2014.
Rhythmic Similarity Carmine Casciato MUMT 611 Thursday, March 13, 2005.
Content-Based Classification, Search & Retrieval of Audio Erling Wold, Thom Blum, Douglas Keislar, James Wheaton Presented By: Adelle C. Knight.
Information Retrieval in Practice
Classification of Music According to Genres Using Neural Networks, Genetic Algorithms and Fuzzy Systems.
Image Fourier Transform Faisal Farooq Q: How many signal processing engineers does it take to change a light bulb? A: Three. One to Fourier transform the.
Classification of Music According to Genres Using Neural Networks, Genetic Algorithms and Fuzzy Systems.
Overview of Search Engines
CELLULAR COMMUNICATIONS DSP Intro. Signals: quantization and sampling.
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING MARCH 2010 Lan-Ying Yeh
Goals For This Class Quickly review of the main results from last class Convolution and Cross-correlation Discrete Fourier Analysis: Important Considerations.
Advanced Multimedia Music Information Retrieval Tamara Berg.
GCT731 Fall 2014 Topics in Music Technology - Music Information Retrieval Overview of MIR Systems Audio and Music Representations (Part 1) 1.
Lecture 1 Signals in the Time and Frequency Domains
TEMPORAL VIDEO BOUNDARIES -PART ONE- SNUEE KIM KYUNGMIN.
Beats and Tuning Pitch recognition Physics of Music PHY103.
R ESEARCH BY E LAINE C HEW AND C HING -H UA C HUAN U NIVERSITY OF S OUTHERN C ALIFORNIA P RESENTATION BY S EAN S WEENEY D IGI P EN I NSTITUTE OF T ECHNOLOGY.
Image recognition using analysis of the frequency domain features 1.
Learning Object Metadata Mining Masoud Makrehchi Supervisor: Prof. Mohamed Kamel.
Chapter 7 Web Content Mining Xxxxxx. Introduction Web-content mining techniques are used to discover useful information from content on the web – textual.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
Acoustic Analysis of Speech Robert A. Prosek, Ph.D. CSD 301 Robert A. Prosek, Ph.D. CSD 301.
Wireless and Mobile Computing Transmission Fundamentals Lecture 2.
Student: Mike Jiang Advisor: Dr. Ras, Zbigniew W. Music Information Retrieval.
Music Information Retrieval -or- how to search for (and maybe find) music and do away with incipits Michael Fingerhut Multimedia Library and Engineering.
Harvestworks Part 3 : Audio analysis & machine learning Rebecca Fiebrink Princeton University 1.
Complex Variables & Transforms 232 Presentation No.1 Fourier Series & Transforms Group A Uzair Akbar Hamza Saeed Khan Muhammad Hammad Saad Mahmood Asim.
Multimodal Information Analysis for Emotion Recognition
Fundamentals of Music Processing
MUMT611: Music Information Acquisition, Preservation, and Retrieval Presentation on Timbre Similarity Alexandre Savard March 2006.
Music Information Retrieval Information Universe Seongmin Lim Dept. of Industrial Engineering Seoul National University.
Spatial Frequencies Spatial Frequencies. Why are Spatial Frequencies important? Efficient data representation Provides a means for modeling and removing.
Speech Recognition Feature Extraction. Speech recognition simplified block diagram Speech Capture Speech Capture Feature Extraction Feature Extraction.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Fourier Series Fourier Transform Discrete Fourier Transform ISAT 300 Instrumentation and Measurement Spring 2000.
GG313 Lecture 24 11/17/05 Power Spectrum, Phase Spectrum, and Aliasing.
MSc Project Musical Instrument Identification System MIIS Xiang LI ee05m216 Supervisor: Mark Plumbley.
Issues in Automatic Musical Genre Classification Cory McKay.
MPEG-7 Audio Overview Ichiro Fujinaga MUMT 611 McGill University.
Fourier Transform.
1 Hidden Markov Model: Overview and Applications in MIR MUMT 611, March 2005 Paul Kolesnik MUMT 611, March 2005 Paul Kolesnik.
BASS TRACK SELECTION IN MIDI FILES AND MULTIMODAL IMPLICATIONS TO MELODY gPRAI Pattern Recognition and Artificial Intelligence Group Computer Music Laboratory.
A content-based System for Music Recommendation and Visualization of User Preference Working on Semantic Notions Dmitry Bogdanov, Martin Haro, Ferdinand.
 Carrier signal is strong and stable sinusoidal signal x(t) = A cos(  c t +  )  Carrier transports information (audio, video, text, ) across.
1 Tempo Induction and Beat Tracking for Audio Signals MUMT 611, February 2005 Assignment 3 Paul Kolesnik.
Information Retrieval in Practice
CS 445/656 Computer & New Media
Discrete Fourier Transform (DFT)
Spectrum Analysis and Processing
SIGNALS PROCESSING AND ANALYSIS
ARTIFICIAL NEURAL NETWORKS
Introduction to Music Information Retrieval (MIR)
Time domain & frequency domain
Brian Whitman Paris Smaragdis MIT Media Lab
Audio and Speech Computers & New Media.
LECTURE 18: FOURIER ANALYSIS OF CT SYSTEMS
Lec.6:Discrete Fourier Transform and Signal Spectrum
Measuring the Similarity of Rhythmic Patterns
Practical Hidden Voice Attacks against Speech and Speaker Recognition Systems NDSS 2019 Hadi Abdullah, Washington Garcia, Christian Peeters, Patrick.
Presentation transcript:

Sound Applications Advanced Multimedia Tamara Berg

Reminder HW2 due March 13, 11:59pm Questions?

Howard Leung

Audio Indexing and Retrieval Features for representing audio: – Metadata – low level features – high level audio features Example usage cases: Audio classification Music retrieval

Content Based Music Retrieval Extract music descriptions from a database of music documents. Extract music description from a query music document. Compute similarity between query and database descriptions. Retrieve similar music documents to query. Casey et al IEEE 2008

MIR tasks H: high level specificity – match specific instances of audio content. M: mid-level specificity – match high level audio features like melody, but do not match audio content. L: low specificity – match global (statistical) properties of the query Different usage cases require different descriptions and matching schema. Casey et al IEEE 2008

Metadata Most common method of accessing music Can be rich and expressive When catalogues become very large, difficult to maintain consistent metadata Useful for low specificity queries Casey et al IEEE 2008

Metadata Pandora.com – Uses metadata to estimate artist similarity and track similarity and creates personalized radio stations. Experts entered metadata of musical- cultural properties (20-30 minutes per track of an expert’s time – 50 person-years for 1 million tracks). Pandora.com Crowd sourced metadata repositories (gracenote, musicbrainz). Factual metadata (artist, album, year, title, duration). Cultural metadata (mood, emotion, genre, style).gracenote musicbrainz Automatic metadata methods – generate descriptions from community metadata automatically. Language analysis to associate noun and verb phrases with musical features (Whitman & Rifkin). Casey et al IEEE 2008

Content features Low level or high level Want features to be robust to certain changes in the audio signal – Noise – Volume – Sampling High level features will be more robust to changes, low level features will be less robust. Low level features will be easy to compute, high level difficult

Content features Low level or high level Want features to be robust to certain changes in the audio signal – Noise – Volume – Sampling High level features will be more robust to changes, low level features will be less robust. Low level features will be easy to compute, high level difficult

Content features Low level or high level Want features to be robust to certain changes in the audio signal – Noise – Volume – Sampling High level features will be more robust to changes, low level features will be less robust. Low level features will be easy to compute, high level difficult

Low level audio features Low level measurements of audio signal that contain information about a musical work. Can be computed periodically ( ms intervals) or beat synchronous. Casey et al IEEE 2008 In text analysis we had words, here we have to come up with our own set of features to compute from audio signal!

Example Low-Level Audio Features Howard Leung

Average number of times signal crosses zero amplitude value.

Howard Leung Average number of times signal crosses zero amplitude value.

Howard Leung Average number of times signal crosses zero amplitude value. 1 if true O o.w.

Howard Leung

Example Low-Level Audio Features Howard Leung

Frequency Domain Reminder How much of each describes the frequency spectrum of a signal. Li & Drew Signals can be decomposed into a weighted sum of sinusoids

Frequency domain features How do we get to frequency domain? TimeFrequency

DFT Discrete Fourier Transform (DFT) of the audio Converts to a frequency representation DFT analysis occurs in terms of number of equally spaced ‘bins’ Each bin represents a particular frequency range DFT analysis gives the amount of energy in the audio signal that is present within the frequency range for each bin Inverse Discrete Fourier Transform (IDFT) Converts from frequency representation back to audio signal.

DFT Discrete Fourier Transform (DFT) of the audio Converts to a frequency representation DFT analysis occurs in terms of number of equally spaced ‘bins’ Each bin represents a particular frequency range DFT analysis gives the amount of energy in the audio signal that is present within the frequency range for each bin Inverse Discrete Fourier Transform (IDFT) Converts from frequency representation back to audio signal.

DFT Discrete Fourier Transform (DFT) of the audio Converts to a frequency representation DFT analysis occurs in terms of number of equally spaced ‘bins’ Each bin represents a particular frequency range DFT analysis gives the amount of energy in the audio signal that is present within the frequency range for each bin Inverse Discrete Fourier Transform (IDFT) Converts from frequency representation back to audio signal.

Howard Leung

Filtering Removes frequency components from some part of the spectrum Low pass filter – removes high frequency components from input and leaves only low in the output signal. High pass filter – removes low frequency components from input and leaves only high in the output signal. Band pass filter – removes some part of the frequency spectrum.

How could you do this using the FT and IFT? Compute FT spectrum of input. Zero out the part of the frequency spectrum that you want to filter out. Compute the IFT of this modified spectrum -> output will be input with some frequency components removed.

How could you do this using the FT and IFT? f = input

How could you do this using the FT and IFT? f = input FT(f)

How could you do this using the FT and IFT? 1 0.* f = input FT(f)

How could you do this using the FT and IFT? 1 0.* f = input FT(f) Zero out some freq components

How could you do this using the FT and IFT? 1 0.* = f = input FT(f) Zero out some freq components x xxx xxxxx xxxx x

How could you do this using the FT and IFT? 1 0.* = f = input FT(f) Zero out some freq components IFT o = Frequency limited output x xxx xxxxx xxxx x

How could you do this using the FT and IFT? 1 0.* = f = input FT(f) Zero out some freq components IFT o = Frequency limited output x xxx xxxxx xxxx x What kind of filter is this?

How could you do this using the FT and IFT? f = input

How could you do this using the FT and IFT? f = input FT(f)

How could you do this using the FT and IFT? 1 0.* f = input FT(f)

How could you do this using the FT and IFT? 1 0.* f = input FT(f) Zero out some freq components

How could you do this using the FT and IFT? 1 0.* = f = input FT(f) Zero out some freq components x x x x xx xxxxxx x

How could you do this using the FT and IFT? 1 0.* = f = input FT(f) Zero out some freq components IFT o = Frequency limited output x x x x xx xxxxxx x

How could you do this using the FT and IFT? 1 0.* = f = input FT(f) Zero out some freq components IFT o = Frequency limited output x x x x xx xxxxxx x What kind of filter is this?

Howard Leung

Frequency Domain Reminder How much of each describes the frequency spectrum of a signal. Li & Drew Signals can be decomposed into a weighted sum of sinusoids

Pitch-Class Profile (PCP) Represent the energy due to each pitch class Integrates the energy in all octaves into a single band There are 12 equally spaced pitch classes in western tonal music. So, typically 12 bands in the PCP.

Pitch-Class Profile (PCP) Represent the energy due to each pitch class Integrates the energy in all octaves into a single band There are 12 equally spaced pitch classes in western tonal music. So, typically 12 bands in the PCP. How might we calculate this using the DFT?

High level music features High level intuitive information about a piece of music (melody, harmony etc). “It is melody that enables us to distinguish one work from another. It is melody that human beings are innately able to reproduce by singing, humming, and whistling. It is melody that makes music memorable: we are likely to recall a tune long after we have forgotten its text.” -Selfridge-Field Intuitive features, but hard to extract and ongoing areas of research. Casey et al IEEE 2008

Melody & Bass Estimation Melody and bass lines represented as continuous temporal trajectory of fundamental frequency, F0, (a series of musical notes). PreFEst ( Goto 1999 ) – Estimate the F0 trajectory in mid-high freq range of input -> melody. – Estimate the F0 trajectory in low freq range-> bass. Casey et al IEEE 2008

Chord Recognition Recognize chord progressions based on: - Estimated PCPs - Statistics of transitions between PCPs Casey et al IEEE 2008

Chord Recognition

Music as vector of features Once again we represent (music) documents as a vector of numbers – Each entry (or set of entries) in this vector is a different feature

Music as vector of features Once again we represent (music) documents as a vector of numbers – Each entry (or set of entries) in this vector is a different feature To retrieve music documents given a query we can: – Find exact matches – Find nearest match – Find nearby matches – Train a classifier to recognize a given category (genre, style etc).

Audio Similarity We have a description of a music document based on some set of features, now how do we compare two descriptions? Casey et al IEEE 2008

Usage examples

Howard Leung

Query by humming Requires robustness to variation because matches will not be exact Extract melody from dataset of songs Extract melody from hum Match by comparing similarities of melodies (nearby matches)

Copyright monitoring Compute fingerprints from database examples Compute fingerprint from query example Find exact matches

Best performing systems on MIREX 2007 Casey et al IEEE 2008

Music Browsing Musicream – UI for discovering and managing musical pieces. User can select a disc and listen to it. By dragging a disc in the flow, the user can easily pick out other similar pieces (attach similar discs). This interaction allows a user to unexpectedly come across various pieces similar to other pieces the user likes. Link to demo Casey et al IEEE 2008

Music Browsing Musicrainbow – UI for discovering unknown artists. Artists are mapped on a circular rainbow where colors represent different styles of music. Similar artists are mapped near each other. User rotates rainbow by turning a knob. Link to demo Casey et al IEEE 2008

Howard Leung