--

The research project focusses on how sound data can be converted into understandable and actionable information by humans and machines. It started on 14 March 2016 and will run until 13 March 2019. The project is funded by the Engineering and Physical Sciences Research Council (EPSRC) with a funding value of £1,275,401. This is a joint project between the Centre for Vision, Speech and Signal Processing (CVSSP) at the University of Surrey, and the Acoustics Research Centre at the University of Salford.

A project overview can be found here.

Contact info.

MSoS Challenge: Results published

The Making Sense of Sounds challenge results have been published.

More...

MSoS Challenge: Extension of submission deadline

The deadline for submissions to the Making Sense of Sounds challenge has been extended to 5. November 2018.

More...

MSoS Challenge: Baseline added

We added a strong state-of-the-art deep learning baseline in the Making Sense of Sounds. Details can be found here.

More...

In the media: TASS interview - 'Can AI/robots create art?'

Christian Kroos gave an interview to the Russian news agency TASS on the question whether AI/robots could create art, sparked by his upcoming invited talk at the VII St. Petersburg International Cultural Forum (Saint Petersburg, Russia).

More...

MSoS Challenge: Evaluation data released and submission opened

The Evaluation data set of the Making Sense of Sounds challenge has been released and the submission system is now open.

More...

MSoS Data Challenge submission updated

More information about the submission to the Making Sense of Sounds data challenge was provided with this update.

More...

Making Sense of Sounds Data Challenge opened

Today the Making Sense of Sounds data challenge was opened with the release of the development data set.

More...

New publication: 'Sound Categories: Category Formation and Evidence-Based Taxonomies '

Bones, O., Cox, T. J., & Davies, W. J. (2018). Sound categories: category formation and evidence-based taxonomies. Frontiers in Psychology, 9, 1277.

More...

New publication: 'Supporting audiography: Design of a system for sentimental sound recording, classification and playback'

Duel, Tijs, Frohlich, David M., Kroos, Christian, Xu, Yong, Jackson, Philip J. B. and Plumbley, Mark D. (2018) Supporting audiography: Design of a system for sentimental sound recording, classification and playback In: HCI International: 20th International Conference on Human-Computer Interaction, 15-20 Jul 2018, Las Vegas, Nevada.

More...

New publication: 'Sonic mnemonic'

Kroos, Christian, Duel, Tijs and Frohlich, David (2018) Sonic mnemonic. In: International Conference: Human-Technology Relations, July 11-13th, 2018, University of Twente, The Netherlands.

More...

New publication: 'Large-scale weakly supervised audio classification using gated convolutional neural network'

Xu, Yong, Kong, Qiuqiang, Wang, Wenwu and Plumbley, Mark D. (2017) Large-scale weakly supervised audio classification using gated convolutional neural network. In: ICASSP 2018 - 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 15 - 20 Apr 2018, Calgary, Canada.

More...

Talk at Research Software Engineer (RSE) event: 'Here be dragons'

The project’s research software developer Christian Kroos gave a presentation at the RSE event ‘What is a Research Software Engineer?’ at the University of Surrey.

More...

New publication: 'A joint separation-classification model for sound event detection of weakly labelled data'

Kong, Qiuqiang, Xu, Yong, Wang, Wenwu and Plumbley, Mark D (2018) A joint separation-classification model for sound event detection of weakly labelled data In: ICASSP 2018 - 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 15 - 20 Apr 2018, Calgary, Canada.

More...

New publication: 'Orthogonality-regularized masked NMF for learning on weakly labeled audio data'

Sobieraj, Iwona, Rencker, Lucas and Plumbley, Mark D (2018) Orthogonality-regularized masked NMF for learning on weakly labeled audio data In: IEEE ICASSP 2018, 15 - 20 April 2018, Calgary, Alberta, Canada.

More...

New publication: 'Audio set classification with attention model'

Kong, Qiuqiang, Xu, Yong, Wang, Wenwu and Plumbley, Mark D. (2018) Audio set classification with attention model: a probabilistic perspective In: ICASSP 2018 - 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 15 - 20 Apr 2018, Calgary, Canada.

More...

New publication: 'Synthesis of images by two-stage generative adversarial networks'

Huang Qiang, Jackson, Philip, Plumbley, Mark D. and Wang, Wenwu (2018) Synthesis of images by two-stage generative adversarial networks In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing, 15–20 Apr 2018, Calgary, Alberta, Canada.

More...

New publication: 'An Evidence-Based Soundscape Taxonomy'

Bones, Oliver, Cox, Trevor J. and Davies, William J. (2017) An Evidence-Based Soundscape Taxonomy. In: 24th International Congress on Sound and Vibration, 23 -27 July 2017 London, UK.

More...

New publication: 'Clang, chitter, crunch: Perceptual organisation of onomatopoeia'

Bones, Oliver, Davies, William J. and Cox, Trevor J. (2017) Clang, chitter, crunch: Perceptual organisation of onomatopoeia The Journal of the Acoustical Society of America 141 (5), 3694-3694

More...

New publication: 'Neuroevolution for sound event detection in real life audio: A pilot study'

Kroos, Christian and Plumbley, Mark D. (2017) Neuroevolution for sound event detection in real life audio: A pilot study. In: DCASE 2017, 16 - 17 November 2017, Munich, Germany.

More...

DCASE 2017 challenge success

Yong Xu, Qiuqiang Kong, Wenwu Wang and Mark Plumbley won the 1st prize in Task 4, ‘large-scale weakly supervised sound event detection for smart cars’, Subtask A, ‘audio tagging’ in the IEEE AASP Challenge on Detection and Classification of Acoustic Scenes and Events (DCASE 2017). The DCASE challenge constitutes the most important challenge in the non-speech audio domain. It is organised by Tampere University of Technology, Carnegie Mellon University and INRIA and sponsored by Google and Audio Analytic. Because of its unique standing, the best players in the field participate such as CMU, New York University, Bosch, USC, TUT, Singapore A*Star, Korean Advanced Institute of Science and Technology, Seoul National University, National Taiwan University and CVSSP.

More...