Connect:  SPS Facebook Page  SPS Twitter  SPS LinkedIn  SPS YouTube Channel  SPS Google+ Page     Share: Share

Audio and Acoustic Signal Processing Technical Committee

AASP Challenges

--> Click here for Call for participation for active challenges -->>

The AASP Technical Committee runs a series of ‘Challenges’ in order to encourage research and development with comparable and repeatable results, and to stimulate new ground-breaking approaches to specific problems in the AASP technical scope. This activity is coordinated by the Challenges Subcommittee listed below.

Call for Challenges

Proposals to organize an AASP Challenge are invited. This is an open call with no deadline. Please email a Stage 1 - Statement of Interest (see below) to the Challenges subcommittee chair (

  1. Stage 1 - Statement of Interest
  2. To propose to organise a Challenge, please send a statement of interest in about 2 pages outlining the aim of the challenge and its value to the community, giving also a preliminary perspective of the practical elements including the planned test data and evaluation methodology.

  3. Stage 2 - Full Proposal
    • a textual description of the challenge and its context (1 to 2 pages);
    • a clear formulation of the problem to be addressed;
    • a specification of the evaluation methodology leading to an objective figure of merit (FoM) and, where appropriate, a software tool to compute the FoM;
    • a development dataset which represents the challenge and which will be made public (a public training dataset may also be needed in some challenges);
    • a test dataset which also represents the challenge but which will remain private during the challenge;
    • a commitment to provide a website to disseminate the challenge itself and, eventually, the results;
    • a commitment to evaluate the submitted results and publish the comparison on the website and elsewhere as appropriate;
    • a proposed schedule for the challenge (date of publication of the challenge, deadline of results submission, deadline of comparative results publication).
  4. If supported by the Subcommittee, a full proposal would be invited. The full proposal should include the following items:

    Please send the Statement of Interest and the Full Proposal to the AASP Challenges Subcommittee chair at the address below. All proposals will be considered by the Challenges Subcommittee. The Subcommittee may request modifications to the challenge as a condition of acceptance.


Researchers entering the challenge are invited to sign up at the challenge website. Participants will address the challenge specification and employ the evaluation methodology and the development dataset to develop their algorithm. Participation is open to all.


At the end of the challenge, the organizers will coordinate a comparative evaluation employing the defined FoM. Evaluation may be done for example by releasing the test dataset and asking the participants to return their results on the test data within a short period of time, typically two weeks. Participants would be honour-bound not to use the test dataset for tuning.



The evaluation results will then be published by the organizers. The Challenges Subcommittee will work with the challenge organizers towards publication of the challenge and its outcome in the IEEE Transactions and appropriate conferences, ICASSP in particular. In addition, the challenge organizers’ website will be linked from the TC webpage. Participants can choose to remain anonymous in publications.

The 'AASP Challenges' Subcommittee

The current membership of the subcommittee is as follows:

Emmanuel Vincent INRIA, France (Chair)
Laurent Daudet Paris Diderot University, France
Jean-Marc Jot DTS Inc., USA
Patrick Naylor Imperial College London, UK
Bryan Pardo Northwestern University, USA
Mark Plumbley Queen Mary University of London, UK
Gael Richard TELECOM ParisTech, France
Ivan Tashev Microsoft Research, USA
Ono Nobutaka National Institute of Informatics, Japan


DCASE 2016

Sounds carry a large amount of information about our everyday environment and physical events that take place in it. Humans can perceive the sound scene we are within (busy street, office, etc.), and recognize individual sound sources (car passing by, footsteps, etc.). The scope of this evaluation is to advance the development of computational scene and event analysis methods by comparing different approaches using a common publicly available dataset and similar metrics, and anchor the current performance for further reference. This challenge follows the success of first edition of the challenge, DCASE 2013.

Everyone is welcome to participate to the challenge, whether it is for only one task or multiple tasks.

On the website of DCASE 2016 you will find everything you need, including:

  • Overview of the tasks providing details of the motivation
  • Detailed descriptions of the tasks, including description of the training and development datasets, metrics and baseline systems for each
  • Detailed description of the challenge rules
  • Instructions on how to submit your results
  • Download page with links to all tools and data packages

DCASE 2016 is an official IEEE Audio and Acoustic Signal Processing (AASP) challenge. Participants will be invited to present their work during a one day workshop that will be organized as a satellite workhop (to be confirmed) to the 2016 European Signal Processing Conference (EUSIPCO), to be held in Budapest, Hungary.

If you need any additional information or have questions about the challenge, do not hesitate to contact us at

If you are considering participating or just want to learn more then please join the Google group DCASE discussions to be up to date with all related information.


DCASE 2016 is a challenge organized by IEEE Signal Processing Society, Tampere University of Technology, Queen Mary University of London, Institut de Recherche et Communication et Cibernétique de Nantes and University of Surrey.




Call for Participation:

Several established parameters and metrics have been used to characterize the acoustics of a room. The most important are the Direct-To-Reverberant Ratio (DRR), the Reverberation Time (T60) and the reflection coefficient. The acoustic characteristics of a room based on such parameters can be used to predict the quality and intelligibility of speech signals in that room. Recently, several important methods in speech enhancement and speech recognition have been developed that show an increase in performance compared to the predecessors but do require knowledge of one or more fundamental acoustical parameters such as the T60. Traditionally, these parameters have been estimated using carefully measured Acoustic Impulse Responses (AIRs). However, in most applications it is not practical or even possible to measure the acoustic impulse response. Consequently, there has been a growing research activity in the estimation of such parameters directly from speech and audio signals.

ACE Challenge Overview

The ACE Challenge is part of the programme of Challenges organised by the IEEE Audio and Acoustic Signal Processing Technical Committee

The aim of this challenge is to evaluate state-of-the-art algorithms for blind acoustic parameter estimation from speech and to promote the emerging area of research in this field. Participants will evaluate their algorithms for T60 and DRR estimation against the ‘ground truth’ values provided with the data-sets. Furthermore, they are expected to present the results in a paper describing the method used.

Full details can be found at

  • Data: A data-set specifically designed for the challenge tasks will be provided. This will include speech from male and female talkers in different sized rooms and different noise conditions for a single microphone and for microphone arrays with three (mobile), five (cruciform), eight (linear), and thirty-two (spherical) microphones

  • Task 1: Single-microphone T60 and DRR estimation

  • Task 2: Multi-microphone T60 and DRR estimation

  • Evaluation: The evaluation metrics will be based on the ground truth values. This will consist of full-band and 1/3-octave band values for T60 and full-band values for the DRR


    ACE Challenge Schedule

    Please see for the key dates of the ACE Challenge.

    ACE Challenge Workshop

    The challenge participants will be invited to present their results at the ACE Challenge workshop, which is planned to be held as a satellite event in conjunction with WASPAA 2015 in New Paltz, NY, USA

    Organizing Committee
    Patrick A. Naylor and James Eaton (Imperial College London), Nikolay D. Gaubitch (Delft University of Technology)





    Call for Participation:

    Recently, substantial progress has been made in the field of reverberant speech signal processing, including both single- and multi-channel de-reverberation techniques, and automatic speech recognition (ASR) techniques robust to reverberation. To evaluate state-of-the-art algorithms and draw new insights regarding potential future research directions, we call for participation in the REVERB (REverberant Voice Enhancement and Recognition Benchmark) challenge that will provide an opportunity to the researchers in the field to carry out comprehensive evaluation of their methods based on a common database and evaluation metrics. The challenge aims at bringing together researchers from a broad range of disciplines to discuss novel and established approaches to handle reverberant speech. This challenge is part of the IEEE SPS AASP challenge series. The following is an outline description of the challenge.

    REVERB Challenge Overview

  • Data: Real and simulated 1-, 2-, and 8-channel recordings in reverberant meeting rooms based on the Wall Street Journal Corpus. This data is common for the following 2 tasks*.

  • Task 1: Enhancement of reverberant speech with single-/multi-channel de-reverberation techniques

    (Evaluation metrics: objective and subjective measures)

  • Task 2: Robust recognition of reverberant speech

    (Evaluation metric: word error rate)

    * Participants are invited to take part in either or both of the above tasks.

  • REVERB Challenge Workshop

    The results will be presented by the participants at the REVERB challenge workshop, which will be held in conjunction with ICASSP2014 and HSCMA2014.

    Important Dates




    Release of development dataset and scripts for evaluation




    Release of evaluation dataset




    Deadline for submission of results




    Deadline for submission of papers




    Notification of acceptance




    Workshop in conjunction with ICASSP2014 (Florence, Italy)

    Further details at

    Organizing Committee
    Marc Delcroix (NTT), Sharon Gannot (Bar-Ilan Univ.), Emannuel Habets (International Audio Labs Erlangen), Reinhold Haeb-Umbach (Paderborn Univ.), Walter Kellermann (Univ. of Erlangen-Nuremberg), Keisuke Kinoshita (NTT), Volker Leutnant (Paderborn Univ.), Roland Maas (Univ. of Erlangen-Nuremberg), Tomohiro Nakatani (NTT), Bhiksha Raj (Carnegie Mellon Univ.), Armin Sehr (Beuth Univ. of Applied Sciences Berlin), Takuya Yoshioka (NTT)



    Detection and Classification of Acoustic Scenes and Events


    Call for Participation:

    On behalf of the IEEE AASP Technical Committee, I am happy to announce a new challenge entitled: "Detection and Classification of Acoustic Scenes and Events". The challenge has the form of a public contest for the evaluation of the performance of systems for the detection and classification of acoustic events and audio scenes.

    The challenge includes a set of tasks for the detection and classification of acoustic scenes and events and its goal is to provide a focus of attention for the scientific community in developing systems for computational auditory scene analysis (CASA) that will encourage sharing of ideas and improve the state of the art, potentially leading to the development of systems that achieve performance closer to that of humans.

    This challenge will help the research community move forward by providing a focus for better defining the specific tasks, and will also provide incentive for researchers to actively pursue research on this field. Finally, it will offer a reference point for future systems developed to perform similar tasks and it will provide the community with a high quality database for future research.

    There will be a discussion phase where potential participants are invited to contribute their ideas, ending on 30th September 2012. The deadline for code submission is 31st March 2013. Results will be presented at a special session in WASPAA 2013 (; participants are invited to present a poster at a special session. Also, authors of novel work are encouraged to submit their work as a regular paper at WASPAA 2013.

    For more details as well as a copy of the full proposal of the challenge please visit:

    The challenge organisers,
    Dimitrios Giannoulis (QMUL), Emmanouil Benetos (QMUL), Dan Stowell (QMUL), Mathieu Lagrange (IRCAM) and Mark Plumbley (QMUL)

    2nd CHiME Speech Separation and Recognition Challenge


    Deadline: January 15, 2013
    Workshop: June 1, 2013, Vancouver, Canada


    Following the success of the 1st PASCAL CHiME Speech Separation and
    Recognition Challenge, we are happy to announce a new challenge
    dedicated to speech recognition in real-world reverberant, noisy conditions,
    that will culminate in a dedicated satellite workshop of ICASSP 2013.

    The challenge is supported by several IEEE Technical Committees and by
    an Industrial Board.


    The challenge consists of recognising distant-microphone speech mixed in
    two-channel nonstationary noise recorded over a period of several weeks
    in a real family house. Entrants may address either one or both of the
    following tracks:

    Medium vocabulary track: WSJ 5k sentences uttered by a static speaker

    Small vocabulary track: simpler commands but small head movements


    You will find everything you need to get started (and even more) on the
    challenge website:
    - a full description of the challenge,
    - clean, reverberated and multi-condition training and development data,
    - baseline training, decoding and scoring software tools based on HTK.

    Submission consists of a 2- to 8-page paper describing your system and
    reporting its performance on the development and the test set. In
    addition, you are welcome to submit an earlier paper to ICASSP 2013,
    which will tentatively be grouped with other papers into a dedicated

    Any approach is welcome, whether emerging or established.

    If you are interested in participating, please email us so we can
    monitor interest and send you further updates about the challenge.


    The best challenge paper will distinguished by an award from the
    Industrial Board.


    July 2012 Launch
    October 2012 Test set release
    January 15, 2013 Challenge & workshop submission deadline
    February 18, 2013 Paper notification & release of the challenge results
    June 1, 2013 ICASSP satellite workshop


    Masami Akamine, Toshiba
    Carlos Avendano, Audience
    Li Deng, Microsoft
    Erik McDermott, Google
    Gautham Mysore, Adobe
    Atsushi Nakamura, NTT
    Peder A. Olsen, IBM
    Trausti Thormundsson, Conexant
    Daniel Willett, Nuance


    Conexant Systems Inc.
    Audience Inc.
    Mitsubishi Electric Research Laboratories


    Emmanuel Vincent, INRIA
    Jon Barker, University of Sheffield
    Shinji Watanabe & Jonathan Le Roux, MERL
    Francesco Nesta & Marco Matassoni, FBK-IRST