Overview What : Stroke type Transformation: Timbre Rhythm When: Stroke timing Resynthesis.

Slides:



Advertisements
Similar presentations
Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Advertisements

Speaker Associate Professor Ning-Han Liu. What’s MIR  Music information retrieval (MIR) is the interdisciplinary science of retrieving information from.
Content-based retrieval of audio Francois Thibault MUMT 614B McGill University.
Outline Introduction Music Information Retrieval Classification Process Steps Pitch Histograms Multiple Pitch Detection Algorithm Musical Genre Classification.
Evaluation of the Audio Beat Tracking System BeatRoot By Simon Dixon (JNMR 2007) Presentation by Yading Song Centre for Digital Music
Content-Based Classification, Search & Retrieval of Audio Erling Wold, Thom Blum, Douglas Keislar, James Wheaton Presented By: Adelle C. Knight.
Digital audio and computer music COS 116: 2/26/2008.
Presentation Outline  Project Aims  Introduction of Digital Video Library  Introduction of Our Work  Considerations and Approach  Design and Implementation.
Classification of Music According to Genres Using Neural Networks, Genetic Algorithms and Fuzzy Systems.
Rowing Motion Capture System Simon Fothergill Ph.D. student, Digital Technology Group, Computer Laboratory Jesus College graduate conference May 2009.
Using Sound in Games Alex Baumann Outline 3D Spatialization Getting and Editing Sounds Using Sounds in Games Music in Games Example Videos.
SmarTrader Analytical/research software system © ITC Software.
Digital audio and computer music COS 116, Spring 2012 Guest lecture: Rebecca Fiebrink.
Team Members: Mohammed Hoque Troy Tancraitor Jonathan Lobaugh Lee Stein Joseph Mallozi Pennsylvania State University.
Carnegie Mellon Music Understanding and the Future of Music Performance Roger B. Dannenberg Professor of Computer Science, Art, and Music Carnegie Mellon.
Knowledge Systems Lab JN 8/24/2015 A Method for Temporal Hand Gesture Recognition Joshua R. New Knowledge Systems Laboratory Jacksonville State University.
 The Weka The Weka is an well known bird of New Zealand..  W(aikato) E(nvironment) for K(nowlegde) A(nalysis)  Developed by the University of Waikato.
JSymbolic and ELVIS Cory McKay Marianopolis College Montreal, Canada.
GCT731 Fall 2014 Topics in Music Technology - Music Information Retrieval Overview of MIR Systems Audio and Music Representations (Part 1) 1.
Polyphonic Queries A Review of Recent Research by Cory Mckay.
SoundSense by Andrius Andrijauskas. Introduction  Today’s mobile phones come with various embedded sensors such as GPS, WiFi, compass, etc.  Arguably,
TINONS1 Nonlinear SP and Pattern recognition
Educational Software using Audio to Score Alignment Antoine Gomas supervised by Dr. Tim Collins & Pr. Corinne Mailhes 7 th of September, 2007.
Music Emotion Recognition 許博智 謝承諺.
KYLE PATTERSON Automatic Age Estimation and Interactive Museum Exhibits Advisors: Prof. Cass and Prof. Lawson.
Harvestworks Part 3 : Audio analysis & machine learning Rebecca Fiebrink Princeton University 1.
Multimodal Information Analysis for Emotion Recognition
Dan Rosenbaum Nir Muchtar Yoav Yosipovich Faculty member : Prof. Daniel LehmannIndustry Representative : Music Genome.
Recognition, Analysis and Synthesis of Gesture Expressivity George Caridakis IVML-ICCS.
MUMT611: Music Information Acquisition, Preservation, and Retrieval Presentation on Timbre Similarity Alexandre Savard March 2006.
TEMPLATE DESIGN © Zhiyao Duan 1,2, Lie Lu 1, and Changshui Zhang 2 1. Microsoft Research Asia (MSRA), Beijing, China.2.
Automatic music classification and the importance of instrument identification Cory McKay and Ichiro Fujinaga Music Technology Area Faculty of Music McGill.
 Detecting system  Training system Human Emotions Estimation by Adaboost based on Jinhui Chen, Tetsuya Takiguchi, Yasuo Ariki ( Kobe University ) User's.
Virtual Instruments and Sequencing. Proses Pembuatan Lagu/Musik TrackingMixingMastering.
It sure is smart but can it swing? (Digital audio and computer music)
Advanced Analytics on Hadoop Spring 2014 WPI, Mohamed Eltabakh 1.
MSc Project Musical Instrument Identification System MIIS Xiang LI ee05m216 Supervisor: Mark Plumbley.
Predicting Voice Elicited Emotions
Objectives: Terminology Components The Design Cycle Resources: DHS Slides – Chapter 1 Glossary Java Applet URL:.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.../publications/courses/ece_8443/lectures/current/lecture_02.ppt.
Realtime Recognition of Orchestral Instruments Ichiro Fujinaga McGill University.
Issues in Automatic Musical Genre Classification Cory McKay.
Faris Issa David Record Sean King Prateek Singhal.
BOOTSTRAPPING INFORMATION EXTRACTION FROM SEMI-STRUCTURED WEB PAGES Andrew Carson and Charles Schafer.
ENTERFACE’08 Multimodal Communication with Robots and Virtual Agents mid-term presentation.
Message Source Linguistic Channel Articulatory Channel Acoustic Channel Observable: MessageWordsSounds Features Bayesian formulation for speech recognition:
Extraction of Individual Tracks from Polyphonic Music Nick Starr.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
A content-based System for Music Recommendation and Visualization of User Preference Working on Semantic Notions Dmitry Bogdanov, Martin Haro, Ferdinand.
Automatic Classification of Audio Data by Carlos H. L. Costa, Jaime D. Valle, Ro L. Koerich IEEE International Conference on Systems, Man, and Cybernetics.
1 Tempo Induction and Beat Tracking for Audio Signals MUMT 611, February 2005 Assignment 3 Paul Kolesnik.
Course Outline (6 Weeks) for Professor K.H Wong
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
The Greek Audio Dataset
Audio to Score Alignment for Educational Software
ECE 417 Lecture 1: Multimedia Signal Processing
Machine Learning overview Chapter 18, 21
Deep Learning Amin Sobhani.
Machine Learning overview Chapter 18, 21
ASAP and Deep ASAP: End-to-End Audio Sentiment Analysis Pipelines
Machine Learning Week 1.
Presented by Steven Lewis
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Realtime Recognition of Orchestral Instruments
Realtime Recognition of Orchestral Instruments
Machine Learning overview Chapter 18, 21
Modeling IDS using hybrid intelligent systems
Measuring the Similarity of Rhythmic Patterns
BeatBox Hero Ankit Gupta Rohan Jain.
Music Signal Processing
Presentation transcript:

Overview What : Stroke type Transformation: Timbre Rhythm When: Stroke timing Resynthesis

Video Demo

The Drum Dayan – treble drum Bayan – bass drum

DhaGeTeTeDhaGeDhaGeTeTeTe Tabla Language

Recognition Architecture Onset detection Statistical Model SVM Bayesian Neural Net Training data ke tun dhe ge dha te Input music Stroke Label Rhythm

Build Model: Training Data Several Datasets Professional musician Home recording Audio recordings manually edited and labeled

Build Model: Target Mapping Standardize idiosyncratic traditional naming conventions Map timbrally similar (or identical) strokes to the same category

Build Model: Feature Extraction Spectral Features MFCCs (24) Centroid Variance Skewness Kurtosis Slope Roll-off Variance F1 F2 F3. Fn Spectral centroid Kurtosis Feature Vector

Build Model: Trained Model WEKA machine learning package Support Vector Machine Models trained on different datasets can be saved for future use

Audio: Input Live audio is taken from a close-mic’d tabla Stereo signal provides partial separation of drums

Audio: Segmentation Onset detection done in Max using bonk~ More recent parallel project uses spectral flux algorithm in Java End of stroke marked by next onset (1 sec buffer size) Onset times stored

Audio: Feature Extraction Variance F1 F2 F3. Fn Spectral centroid Kurtosis Feature Vector

Output: Classification Feature vector is fed to previously trained model Single category label returned SVMlabelfeature vector

Output: Symbolic Score Stroke label combined with timing and amplitude information Score stored in temporary buffer in Max patch

Output: Timbre Remapping Stroke labels can be flexibly remapped

Output: Conditional Repetition

Output: User Interface

Dangum

Future Directions Beat tracking Modeling specific types of improvisational forms (e.g. qaida, tihai …) Automate transformations Improve interface so it can be “played” Tracking of expressive parameters (e.g. bayan pitch modulation)

Conclusions Shown a realtime tabla interaction system Implemented as Max java external using machine learning to identify strokes Supports flexible transformations Foundation for more general improvisation system