Making Sense of Sounds? Towards Reusable Software for Audio & Music Research Mark Plumbley Centre for Digital Music Queen Mary, University of London SANE.

Slides:



Advertisements
Similar presentations
Tips for Training (module 6.2).
Advertisements

SoundSoftware.ac.uk Prizes for Reproducibility in Audio & Music Research Chris Cannam, Luís Figueira, Mark Plumbley Centre for Digital Music Queen Mary,
1 Copyright © 2010 AQA and its licensors. All rights reserved. Introduction to the new specification GCSE Computer Science Paul Varey.
Session Objective To understand how to set high quality learning objectives and learning outcomes Session Outcomes By the end of this session you will.
Web services for Improving the development of automatic generalisation solutions Nicolas Regnauld Research & Innovarion Ordnance Survey 07 th March 2006,
Build VIVO in the Cloud NIH Workshop on Value Added Services for VIVO Brand Niemann Semantic Community March 25-26,
Measuring and reporting outcomes for your BTOP grant 1Measuring and Reporting Outcomes.
1 CCLI Proposal Writing Strategies Tim Fossum Program Director Division of Undergraduate Education National Science Foundation Vermont.
Object-Oriented Software Engineering Practical Software Development using UML and Java Chapter 3: Basing Software Development on Reusable Technology.
August 14, 2015 Research data management – an introduction Slides provided by the DaMaRO Project, University of Oxford Research Services.
Open and Shared Information System OaSIS. SUNCOM’s Standard Business Process Centralized ordering for the enterprise Maintenance of an enterprise inventory.
The Community Café project: language teachers creating and sharing resources online Alison Dickens Subject Centre for Languages, Linguistics and Area Studies.
SYNAT - the Polish National Research Content Infrastructure Wojtek Sylwestrzak, ICM Tomasz Rosiek, ICM Tomasz Krassowski, ICM Tartu, Estonia June 27, 2012.
Project Proposal: Academic Job Market and Application Tracker Website Project designed by: Cengiz Gunay Client: Cengiz Gunay Audience: PhD candidates and.
“From Diary Room to Board Room” Using the Student Voice to Inform Institutional Practice and Policy to Enhance the Student Experience Nicola Poole - UWIC.
Peer Assisted Learning (PAL)
Management, marketing and population of repositories Morag Greig, University of Glasgow.
Pointers to more effective software & data in audio research Mark D Plumbley, Chris Cannam, Steve Welburn Centre for Digital Music Queen Mary, University.
WIKI IN EDUCATION Giti Javidi. W HAT IS WIKI ? A Wiki can be thought of as a combination of a Web site and a Word document. At its simplest, it can be.
Mendeley Institutional Edition Hazman Aziz, eProduct Manager (APAC) University Kebangsaan Malaysia.
SoundSoftware.ac.uk: Software sustainability for the audio and music researcher Chris Cannam, Mark Plumbley, Luís Figueira Centre for Digital Music Queen.
Software Sustainability Institute Linking software: Citations, roles, references,and more
Queen Mary, University of London: Discipline Bridging Initiative 1 Queen Mary, University of London Discipline Bridging Initiative Funded via the MRC/EPSRC.
Pointers to more effective software & data in audio research Mark D Plumbley, Chris Cannam, Steve Welburn Centre for Digital Music Queen Mary, University.
The DSpace Course Module – An introduction to DSpace.
Introduction CSE 1310 – Introduction to Computers and Programming Vassilis Athitsos University of Texas at Arlington 1.
Learning how to learn, with Software Carpentry Luis Figueira, Chris Cannam, Mark D Plumbley Centre for Digital Music Queen Mary, University of London.
We are the 92% Valuing the contribution of research software Neil Chue Hong, FORCE2015 Research Communications and e-Scholarship.
Per Møldrup-Dalum State and University Library SCAPE Information Day State and University Library, Denmark, SCAPE Scalable Preservation Environments.
©Ian Sommerville 2000 Software Engineering, 6th edition. Chapter 10Slide 1 Architectural Design l Establishing the overall structure of a software system.
Introduction CSE 1310 – Introduction to Computers and Programming Vassilis Athitsos University of Texas at Arlington 1.
The ILA in Detail. OBJECTIVES 1.To understand the purpose of the ILA 2.To have a clear understanding of how to complete the ILA 3.To have considered the.
EGEE is a project funded by the European Union under contract IST Testing processes Leanne Guy Testing activity manager JRA1 All hands meeting,
Software Sustainability Institute Dealing with software: the research data issues 26 August.
1 G4MICE Design Iteration Malcolm Ellis MICE Video Conference 21 st April 2004.
Can an Authentic Interest for Research Integrity be Fostered? From a Government Driven Promotion to a Voluntary Self- education and Consensus Forming on.
©Ian Sommerville 2000 Software Engineering, 6th edition. Chapter 10Slide 1 Architectural Design l Establishing the overall structure of a software system.
Architectural Design Yonsei University 2 nd Semester, 2014 Sanghyun Park.
Software Sustainability Institute Software Attribution can we improve the reusability and sustainability of scientific software?
SoundSoftware.ac.uk: Sustainable Software for Audio & Music Research Mark Plumbley, Chris Cannam and Luis Figueira Centre for Digital Music Queen Mary,
May 2, 2013 An introduction to DSpace. Module 1 – An Introduction By the end of this module, you will … Understand what DSpace is, and what it can be.
SoundSoftware.ac.uk: Towards Reusable Software for Audio & Music Research Mark Plumbley, Chris Cannam and Luis Figueira Centre for Digital Music Queen.
We are the 92% 16 November 2014, WSSSPE2, SC14, New Orleans, USA Neil Chue Hong Software Sustainability.
Corporate Data Vault Data Warehousing Workshop Sept Data Warehousing Workshop Sept
Institutional Repositories: the DSpace Experience Ann J. Wolpert Director of Libraries Massachusetts Institute of Technology.
Code.soundsoftware.ac.uk: A software repository for sustainable collaborative research Mark Plumbley, Chris Cannam, Luis Figueira Centre for Digital Music.
Software Sustainability Institute Building sustainable software for science … why good code is only the beginning 10 April 2013, EGI.
Pointers to more effective software & data in audio research Mark D Plumbley, Chris Cannam, Steve Welburn Centre for Digital Music Queen Mary, University.
Working with your archive organization: Broadening your user community Robert R. Downs, PhD Socioeconomic Data and Applications Center (SEDAC) Center for.
Filling institutional repositories: considering copyright issues Susan Veldsman eIFL Content Manager
Overview In this tutorial you will: learn what an e-portfolio is learn about the different things e-portfolios may be used for identify some options for.
Cloud-based e-science drivers for ESAs Sentinel Collaborative Ground Segment Kostas Koumandaros Greek Research & Technology Network Open Science retreat.
SciencePAD Open Software for Open Science Alberto Di Meglio – CERN.
RDA-WDS Publishing Data IG Data Bibliometrics Working Group.
Working with Your Archive : Broadening Your User Community Robert R. Downs, PhD NASA Socioeconomic Data and Applications Center (SEDAC) Center for International.
LAMS: Yesterday, Today, Tomorrow James Dalziel Professor of Learning Technology & Director, Macquarie University E-learning Centre of Excellence (MELCOE)
Talking about the Scholarship Repository June 21, 2016 Charlotte Roh, University of San Francisco.
Reference Management Module I: Introduction By Rehema Chande-Mallya(PhD)
CS 501: Software Engineering Fall 1999 Lecture 23 Design for Usability I.
Advanced Higher Computing Science
Introduction to SHERPA RoMEO and its Significance for Publishers
IS301 – Software Engineering Dept of Computer Information Systems
Tim Smith CERN Geneva, Switzerland
Pointers to more effective software & data in audio research
Part 3 Design What does design mean in different fields?
Publishing software and data
KIOS Open Knowledge: A pillar for excellence
Get thinking: What is the problem, issue or challenge you want to take on? 
Open Access to your Research Papers and Data
Open access in REF – Planning Workshop
Presentation transcript:

Making Sense of Sounds? Towards Reusable Software for Audio & Music Research Mark Plumbley Centre for Digital Music Queen Mary, University of London SANE Workshop 24 Oct 2013

Overview Centre for Digital Music at Queen Mary Reproducible research Survey Research Pipeline problems Barriers and approaches Software Carpentry Boot Camps Reproducible Research Prizes D-CASE Data Challenge Conclusions

Centre for Digital Music (C4DM) Digital Music Research began at QM 2001, C4DM formed 2003 World-leading research into digital technologies for new understanding and innovation in music and audio. 55+ people: 11 academics, 35 PhD students, 12 researchers Research funding: over £17M ($27M) since 2007 Conferences: ISMIR 2005, ICA 2007, MPEG 2009, CMMR 2012 Teaching: BEng Audio Systems Eng, MSc Digital Music Proc Regular international visitors Software: Sonic Visualiser, SoundBite,... Partners: BBC, last.fm, FXpansion, Yamaha,...

C4DM Research Areas

Centre for Digital Music: People Dec 2011

Project Spotlights

Rhythm Transformation Rhythm Tracking on model (top) and original (bottom) Time-scale segments of original to rhythm of model Licenced for new LickMixeR App Original Result

Live Beat Tracking: B-Keeper [Video]

Sonic Visualiser Visualise and edit content-derived metadata (low-level audio features and semantic descriptors) Open source “VAMP” plug-in API for creating new feature extractors Plug-ins for onset, beat, structural segmentation, key, transcription, etc ‘Speaks’ RDF, contribute/consume Linked Open Data Used by music researchers, musicologists, etc. (> 200k downloads)

Reproducible Research

Idea: researchers should be able to reproduce the work of others. Research used to be “reproducible” from the paper alone. In audio & music research, methods are now too complex. The paper is not enough: need algorithm, parameters, datasets,... So, we need The paper (ideally Open Access) The code(ideally Open Source) The data(ideally Open Data) Well-known example: WaveLab (Buckheit & Donoho, 1995) B ut in audio & music research, few people do this. Why? (Buckheit & Donoho, 1995; Vandewalle et al, 2009)

Survey 2010– % develop code * - Oct 2010-Apr 2011, 54 complete + 23 partial responses. For these figures we considered 72 responses. We asked some researchers:*

Survey 2010–2011 of whom 39% report taking steps to reproducibility

Survey 2010–2011 of whom 35% report publishing any code

Survey 2010–2011 That's 11% of the whole

Summary 82% of respondents reported developing software, but only 39% of those said they took steps to reproducibility only 35% of those reported ever publishing any code i.e. only 11% tried to be reproducible and published the code. Also: 51% said their code never left their own computer

Why don't people publish code? We found: Lack of time Copyright restrictions Potential for future commercial use Other factors (UK Research Information Network, 2010): Lack of evidence of benefits Culture of independence or competition Quality concerns (self-taught programmers) Also: it takes effort early in the research cycle; hard to find time/motivation after the paper is published

So instead of this Research Pipeline, Researcher A (“Producer”) Read background papers Do own research Publish paper X Researcher B (“Consumer-Producer”) Read paper X Understand/reproduce results in paper X Do more research building on X Publish paper Y that cites X / produce product that uses X... and so on.

... we have: Real Research Pipeline Researcher A (“Producer”) Read background papers Do own research (including lots of coding) Publish paper X (not enough space for all the code) Researcher B (“Consumer-Producer”) Read paper X Can’t reproduce or use results in paper X Tear out hair Give up / do something else NB: A and B may be in same group (or same person later!)

Sustainability Can still be problems, even if software is published… Software designed for legacy platforms / not maintained (Sun SPARC, NeXTSTEP, Dec Alpha) PhD students graduate, staff move: -> web pages containing original software/data lost Software/datasets in slightly different versions (error corrections, enhancements) Other issues Copyright issues of datasets (e.g. audio of Beatles tracks cannot be placed on the web)

Example: Beat Tracking Classic algorithm (Scheirer, 1998) in legacy C++ for DEC Alpha, original website lost, various modified versions “passed around” Another early algorithm (Goto, ) written for a parallel architecture, computer no longer exists, code never released UK algorithm (Hainsworth, 2005) in Matlab but with Windows-only DLL component (limits its portability) Another key algorithm (Klapuri et al, 2006) in Matlab, available from authors, requires contract preventing redistribution Several other algorithms never released, but researchers may help individually (Cemgil, 2000; Laroche 2003; Peeters 2005). So; hard for new researchers to compare with those algortihms!

Example of potential: Beat Tracking Onset detection (Bello, 2003) in beat tracking (Davies, 2005) Davies (Matlab), ported to SuperCollider by Collins (2006) Davies alg ported into cross-platform C++ Vamp plugin for Sonic Visualiser / Sonic Annotator (on EPSRC OMRAS2 project) Inspired Max/MSP beat tracking system (Robertson, 2007) Used in beat-synchronous audio effects (Stark, 2008) - developed in Matlab, ported to real-time VST plugins. Rhythm morphing (Hockman & Davies, 2008) originally Matlab, ported into a C++ library (on EPSRC Platform Grant) Helped by (a) personal continuity, (b) funding for “extra step”

How can we tackle these issues? We’re taking a bottom-up approach: Make incremental improvements to development practice by Identifying specific barriers to publication and reuse, that are relatively straightforward to address So we hope to: Increase perception among researchers that code is something you can work on together, that can be reused Prepare the ground for reproducible publication

Barriers to publication and reuse Lack of education and confidence with code Lack of facilities and tools Lack of incentive for publication (Also: Platform incompatibilities)

Barrier: Lack of confidence in code Issue: Researchers largely self-trained in software development Our approach: Training in research software development Relatively small amounts of training can pay off Autumn School (Nov 2010, 5 days) based on Software Carpentry Version control systems Unit testing, test-driven development Python syntax and structure Managing experimental datasets

Software Carpentry Boot Camps Pre-conference Boot Camp (DAFx 2012, York, UK, pictured) Tutorials at conferences (DAFx 2012, ISMIR 2012) Integration into PhD training programme

Barrier: Lack of facilities and tools Issue: Researchers don’t use code hosting / version control Research groups / institutions often do not provide any Researchers often unaware of them Our approach: code site: Focus on audio and music research Public and private projects Link publications with code Also: User interfaces for version control Existing ones are surprisingly difficult EasyMercurial: and tutorials and videos

Over 350 projects from over 200 users

Barrier: Lack of incentive Issue: Software not recognised as valued research output Lacks publication conventions for authorship, academic rewards unclear Approach (1): Link publications to code on the code site Increase likelihood of code users discovering your papers Ensure users know how to cite your work Increase take-up / impact of your research Approach (2): Other encouragements Reproducible Research Prizes Data challenges

Reproducible Research Prizes

RR Prizes: Motivation Promote development and release of sustainable and reusable software associated with published research Recognise researchers who take the extra step, or whose work which will enable others to do so Offer a really clear short-term incentive

RR Prizes: What we did (and why) Broad call for submissions (April 2013) journal or conference papers, published or pending Very little idea what response we might get “If you have published your software or datasets as part of your audio or music research output, so that other UK researchers can reproduce your results, you could win a prize!”

RR Prizes: Categories and prizes Journal paper: New submission Conference paper: New submission Journal paper: Already published Conference paper: Already published Prizes: Article Processing Charge for open access publication; travel bursaries; other options

RR Prizes: Categories and prizes Journal paper: New submission – one entry! Conference paper: New submission – 3 entries Journal paper: Already published – 3 entries Conference paper: Already published – 5 entries 13 entries total (7 from UK) across 10 institutions 5 MATLAB; 3 C/C++; 2 Python; 1 Lisp Two with no software (datasets only)

RR Prizes: Work and meta-work Five traditional research papers Two papers presenting “challenges” Two presenting software applications Two reviewing reproducibility of other work One presenting a newly-compiled test dataset One presenting a data interchange format

RR Prizes: Judging criteria – Ease of reproducibility of the results → assessed by SoundSoftware (that’s us) – Quality of sustainability planning → assessed by the Software Sustainability Institute – Potential to enable high quality research in the UK audio and music research community → assessed by external reviewers

How reproducible were they? Somewhat… with a number of fiddly details! Hard-coded paths for dependency files and scripts (/home/myname/test1/data.csv) MATLAB version incompatibilities, missing Python modules Public datasets gone missing (e.g. Magnatagatune) Randomised test datasets, random matrix initialisers, etc Good practices: Publishing via e.g. github or our own code site (5 submissions) Script to test the environment is set up correctly (1 submission) Scripts as used when assembling the actual paper!

Prize winners Winners announced at SSW workshop June 2013 Majdak, P., et al, Spatially Oriented Format for Acoustics: A Data Exchange Format Representing Head-Related Transfer Functions Sturm, B. L., et al, Comments on “Automatic Classification of Musical Genres Using Inter-Genre Similarity” – and two other papers Giannoulis, D., et al, A Database and Challenge for Acoustic Scene Classification and Event Detection Raffel, C., and Ellis, D., Reproducing Pitch Experiments in “Measuring the Evolution of Contemporary Western Popular Music”

Next round Linked to Audio Engineering Society (AES) 53 rd Conference on Semantic Audio (London, January 2014) Prize submission deadlines coordinated with the AES Hints for reproducible publication available Encourage people to think about software as they write paper *** Watch this space *** Future: “RR Prize Kit” for conference organizers

Data Challenge: Detection and Classification of Audio Scenes & Events [WASPAA slides]

D-CASE Challenge Summary Outcome from WASPAA Special Session (12 posters) Lots of interest in the Challenge Many discussions about evaluation! Code available for 4 submitted systems: Scene Classification: Chum et al, Olivetti Event Detection: Gemmeke et al, Vuegen et al Also: Our own baseline systems More at:

Many thanks to: Sound Software and Reproducible Research: QM: Chris Cannam, Luis Figueira, Daniele Barchiesi, Steve Welburn Software Carpentry: Greg Wilson SSI: Tim Parkinson, Arno Proeme, Neil Chue Hong D-CASE Challenge: QM: Dimitrios Giannoulis, Dan Stowell; City: Emmanouil Benetos; IRCAM: Mathias Rossignol, Mathieu Lagrange

Conclusions Research too complex to be reproducible from paper alone Reproducible Research: Paper + Software + Data. Not Easy! Identify barriers and simple approaches to overcome them: Lack of education / confidence with code -> Training Lack of facilities and tools -> Repository, GUI Lack of incentive for publication -> Link papers to code, Prizes, Data Challenge Do want help to run a … Software Carpentry Boot Camp? … Reproducible Research Prize? … Data Challenge? -> Talk to us!