Automatic synchronisation between audio and score musical description layers Antonello D’Aguanno, Giancarlo Vercellesi Laboratorio di Informatica Musicale.

Slides:



Advertisements
Similar presentations
Patient information extraction in digitized X-ray imagery Hsien-Huang P. Wu Department of Electrical Engineering, National Yunlin University of Science.
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Audio-based Emotion Recognition for Advanced Information Retrieval in Judicial Domain ICT4JUSTICE 2008 – Thessaloniki,October 24 G. Arosio, E. Fersini,
Search in Source Code Based on Identifying Popular Fragments Eduard Kuric and Mária Bieliková Faculty of Informatics and Information.
Multimedia Database Systems
Toward Automatic Music Audio Summary Generation from Signal Analysis Seminar „Communications Engineering“ 11. December 2007 Patricia Signé.
Adapted representations of audio signals for music instrument recognition Pierre Leveau Laboratoire d’Acoustique Musicale, Paris - France GET - ENST (Télécom.
Evaluation of the Audio Beat Tracking System BeatRoot By Simon Dixon (JNMR 2007) Presentation by Yading Song Centre for Digital Music
Feature Vector Selection and Use With Hidden Markov Models to Identify Frequency-Modulated Bioacoustic Signals Amidst Noise T. Scott Brandes IEEE Transactions.
A Novel Scheme for Video Similarity Detection Chu-Hong Hoi, Steven March 5, 2003.
Chapter 11 Beyond Bag of Words. Question Answering n Providing answers instead of ranked lists of documents n Older QA systems generated answers n Current.
Fifth International Conference on Information
Aki Hecht Seminar in Databases (236826) January 2009
ADVISE: Advanced Digital Video Information Segmentation Engine
From Midi to Traditional Musical Notation by Emilios Cambouropoulos.
MPEG Audio Compression by V. Loumos. Introduction Motion Picture Experts Group (MPEG) International Standards Organization (ISO) First High Fidelity Audio.
Classification of Music According to Genres Using Neural Networks, Genetic Algorithms and Fuzzy Systems.
Processor Frequency Setting for Energy Minimization of Streaming Multimedia Application by A. Acquaviva, L. Benini, and B. Riccò, in Proc. 9th Internation.
Object Tracking for Retrieval Application in MPEG-2 Lorenzo Favalli, Alessandro Mecocci, Fulvio Moschetti IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR.
Introduction & Overview CS4533 from Cooper & Torczon.
IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING MARCH 2010 Lan-Ying Yeh
/14 Automated Transcription of Polyphonic Piano Music A Brief Literature Review Catherine Lai MUMT-611 MIR February 17,
Adaptive Signal Processing Class Project Adaptive Interacting Multiple Model Technique for Tracking Maneuvering Targets Viji Paul, Sahay Shishir Brijendra,
1 Presenter: Ming-Shiun Yang Sah, A., Balakrishnan, M., Panda, P.R. Design, Automation & Test in Europe Conference & Exhibition, DATE ‘09. A Generic.
Supporting the Automatic Construction of Entity Aware Search Engines Lorenzo Blanco, Valter Crescenzi, Paolo Merialdo, Paolo Papotti Dipartimento di Informatica.
CONTI’2008, 5-6 June 2008, TIMISOARA 1 Towards a digital content management system Gheorghe Sebestyen-Pal, Tünde Bálint, Bogdan Moscaliuc, Agnes Sebestyen-Pal.
1 Yolanda Gil Information Sciences InstituteJanuary 10, 2010 Requirements for caBIG Infrastructure to Support Semantic Workflows Yolanda.
Theoretic and artistic research studying opportunities of symbiosis of Western and non-Western musical idiom Olmo Cornelis.
Patent Liability Analysis Team 14: Faris Issa Sean King David Record Prateek Singhal.
Ian Addison Primary School Teacher. Didn’t have PCs at school Owned my first PC at 16 Studied BTEC IT at college – this included breaking (and fixing)
Audio Compression Usha Sree CMSC 691M 10/12/04. Motivation Efficient Storage Streaming Interactive Multimedia Applications.
Data Publication 101 for PhD students, starting their academic career [2014] CC-BY: 3TU.Datacentre more
Signal Digitization Analog vs Digital Signals An Analog Signal A Digital Signal What type of signal do we encounter in nature?
Aspects of Music Information Retrieval Will Meurer School of Information University of Texas.
Polyphonic Music Transcription Using A Dynamic Graphical Model Barry Rafkind E6820 Speech and Audio Signal Processing Wednesday, March 9th, 2005.
By Chung-Hong Lee ( 李俊宏 ) Assistant Professor Dept. of Information Management Chang Jung Christian University 資料庫與資訊檢索系統的整合 - 一個文件資料庫系統的開發研究.
Document Clustering for Forensic Analysis: An Approach for Improving Computer Inspection.
Feature Vector Selection and Use With Hidden Markov Models to Identify Frequency-Modulated Bioacoustic Signals Amidst Noise T. Scott Brandes IEEE Transactions.
Semantic Wordfication of Document Collections Presenter: Yingyu Wu.
COMMUNICATION SYSTEM EEEB453 Chapter 5 (Part V) DIGITAL TRANSMISSION-LINE ENCODING Intan Shafinaz Mustafa Dept of Electrical Engineering Universiti Tenaga.
Dynamic Captioning: Video Accessibility Enhancement for Hearing Impairment Richang Hong, Meng Wang, Mengdi Xuy Shuicheng Yany and Tat-Seng Chua School.
Event retrieval in large video collections with circulant temporal encoding CVPR 2013 Oral.
2005/12/021 Fast Image Retrieval Using Low Frequency DCT Coefficients Dept. of Computer Engineering Tatung University Presenter: Yo-Ping Huang ( 黃有評 )
Mining Document Collections to Facilitate Accurate Approximate Entity Matching Presented By Harshda Vabale.
Search Engine using Web Mining COMS E Web Enhanced Information Mgmt Prof. Gail Kaiser Presented By: Rupal Shah (UNI: rrs2146)
Ask a Librarian: The Role of Librarians in the Music Information Retrieval Community Jenn Riley, Indiana University Constance A. Mayer, University of Maryland.
Data Structure Introduction Dr. Bernard Chen Ph.D. University of Central Arkansas Fall 2010.
MPEG-7 Audio Overview Ichiro Fujinaga MUMT 611 McGill University.
Reporter: Shau-Shiang Hung( 洪紹祥 ) Adviser:Shu-Chen Cheng( 鄭淑真 ) Date:99/06/15.
Content-Based MP3 Information Retrieval Chueh-Chih Liu Department of Accounting Information Systems Chihlee Institute of Technology 2005/06/16.
1 Automatic Music Style Recognition Arturo Camacho.
DYNAMIC TIME WARPING IN KEY WORD SPOTTING. OUTLINE KWS and role of DTW in it. Brief outline of DTW What is training and why is it needed? DTW training.
Preliminary Transformations Presented By: -Mona Saudagar Under Guidance of: - Prof. S. V. Jain Multi Oriented Text Recognition In Digital Images.
A shallow description framework for musical style recognition Pedro J. Ponce de León, Carlos Pérez-Sancho and José Manuel Iñesta Departamento de Lenguajes.
1 Tempo Induction and Beat Tracking for Audio Signals MUMT 611, February 2005 Assignment 3 Paul Kolesnik.
A G3-PLC Network Simulator with Enhanced Link Level Modeling
Catherine Lai MUMT-611 MIR February 17, 2005
Robust Data Hiding for MCLT Based Acoustic Data Transmission
Spread Spectrum Audio Steganography using Sub-band Phase Shifting
Posture Monitoring System for Context Awareness in Mobile Computing
Neural Networks Dr. Peter Phillips.
Aspects of Music Information Retrieval
An XML Multi-layer Framework for Music Information Description
Block Matching for Ontologies
Authors: Wai Lam and Kon Fan Low Announcer: Kyu-Baek Hwang
Ms Jennifer - Senior 4 - Data Representation Introduction
M. Kezunovic (P.I.) S. S. Luo D. Ristanovic Texas A&M University
Communication Theory as Applied to Wireless Sensor Networks
♪ Embedded System Design: Synthesizing Music Using Programmable Logic
Music Signal Processing
Presentation transcript:

Automatic synchronisation between audio and score musical description layers Antonello D’Aguanno, Giancarlo Vercellesi Laboratorio di Informatica Musicale Università degli Studi di Milano Speaker: Antonello D’Aguanno

Introduction Contemporary digital music archives consist of huge collections of heterogeneous documents The heterogeneity of music information makes retrieval hard to accomplish Synchronization is one of the open problems in Music Information Retrieval (MIR) Automatic synchronisation between audio and score musical description layers 2 /15

Synchronisation 3 /15 Stream Audio Score Synchronized score Synchronization Synchronisation means that for a given event in a music score we can determine the timing of the corresponding audio events Automatic synchronisation between audio and score musical description layers

Related Works Many algorithms have been proposed in literature that deal with synchronisation All the algorithms proposed use MIDI score representation and PCM audio 4 /15 Automatic synchronisation between audio and score musical description layers

Related Works The unsolved problems of synchronisation are not limited to finding a suitable algorithm How can this result generalised to use it in other applications? No answers in literature… MX can solve the problem 5 /15 Automatic synchronisation between audio and score musical description layers

The linking structure in MX 6 /15 Notational Layer Spine Sub-Layer Performance Layer Audio Layer Automatic synchronisation between audio and score musical description layers

The COMSI algorithm Three phases: Score Analysis Audio Analysis Recursive Decisional Matching 7 /15 Automatic synchronisation between audio and score musical description layers

The three phases Score Analysis: The MX score is read in order to extract all relevant musical events Audio Analysis: The PCM audio signal is analysed to identify all possible musical notes Recursive Decisional Matching: Relates the event at the score level with the same event at the audio level 8 /15 Automatic synchronisation between audio and score musical description layers

MX score analysis 1.For each measure, only the notes with strong accent are selected 2.For every strong accent, the verticalisation of the score is computed 9 /15 Observation: COMSI Algorithm can manage conveniently even MIDI score Automatic synchronisation between audio and score musical description layers

Audio Analysis 1.The audio signal is filtered with a notch filter centred on the frequency of the note examined 2.The audio signal in time-windows is split at 100ms 3.For each note, a possible attack-time is the audio segment which has an energy value above a threshold, obtained from the average energy of the filtered signal 10 /15 Automatic synchronisation between audio and score musical description layers

Execution Events The execution events are all the audio events having an energy above the threshold The set of these execution events is named pseudo-score 11 /15 Automatic synchronisation between audio and score musical description layers

The Decisional Matching 12 /15 Automatic synchronisation between audio and score musical description layers Measure Synchronisation Recursive research of attack-time related to the first event of each measure in the score Musical Event Synchronization Sequential research of each musical event contained in a measure

Results 13 /15 TrackTime Signature Correct Measure One Error Measure Event Twice Error Measure Event Chopin 3/466%24%10% Beethoven 4/471%18%11% Beethoven Faster 4/459%36%15% Fiore di Maggio 4/480%17%13% Cavatina 6/843%48%9% Aria 6/861%31%8% Automatic synchronisation between audio and score musical description layers

Conclusions In this work we have described algorithm dedicated to score and audio alignment using the MX / IEEE P1599 format This algorithm allows alignment of an MX score and its execution, coded in PCM format It produces an output for the MX Spine that contains synchronisation between notes and audio signal 14 /15 Automatic synchronisation between audio and score musical description layers

Future works A new test infrastructure is being developed, which will use the MX capabilities More tests in a shorter time! Will take into account every musical event Automatic synchronisation between audio and score musical description layers 15 /15

Thanks! Titolo della presentazione16 di 15