--

The research project focusses on how sound data can be converted into understandable and actionable information by humans and machines. It started on 14 March 2016 and will run until 13 March 2019. The project is funded by the Engineering and Physical Sciences Research Council (EPSRC) with a funding value of £1,275,401. This is a joint project between the Centre for Vision, Speech and Signal Processing (CVSSP) at the University of Surrey, and the Acoustics Research Centre at the University of Salford.

A project overview can be found here.

Contact info.

Making Sense of Sounds Data Challenge opened

Today the Making Sense of Sounds data challenge was opened with the release of the development data set.

More...

New publication: 'Sound Categories: Category Formation and Evidence-Based Taxonomies '

Bones, O., Cox, T. J., & Davies, W. J. (2018). Sound categories: category formation and evidence-based taxonomies. Frontiers in Psychology, 9, 1277.

More...

New publication: 'Supporting audiography: Design of a system for sentimental sound recording, classification and playback'

Duel, Tijs, Frohlich, David M., Kroos, Christian, Xu, Yong, Jackson, Philip J. B. and Plumbley, Mark D. (2018) Supporting audiography: Design of a system for sentimental sound recording, classification and playback In: HCI International: 20th International Conference on Human-Computer Interaction, 15-20 Jul 2018, Las Vegas, Nevada.

More...

New publication: 'Sonic mnemonic'

Kroos, Christian, Duel, Tijs and Frohlich, David (2018) Sonic mnemonic. In: International Conference: Human-Technology Relations, July 11-13th, 2018, University of Twente, The Netherlands.

More...

New publication: 'Large-scale weakly supervised audio classification using gated convolutional neural network'

Xu, Yong, Kong, Qiuqiang, Wang, Wenwu and Plumbley, Mark D. (2017) Large-scale weakly supervised audio classification using gated convolutional neural network. In: ICASSP 2018 - 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 15 - 20 Apr 2018, Calgary, Canada.

More...

Talk at Research Software Engineer (RSE) event: 'Here be dragons'

The project’s research software developer Christian Kroos gave a presentation at the RSE event ‘What is a Research Software Engineer?’ at the University of Surrey.

More...

New publication: 'A joint separation-classification model for sound event detection of weakly labelled data'

Kong, Qiuqiang, Xu, Yong, Wang, Wenwu and Plumbley, Mark D (2018) A joint separation-classification model for sound event detection of weakly labelled data In: ICASSP 2018 - 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 15 - 20 Apr 2018, Calgary, Canada.

More...

New publication: 'Orthogonality-regularized masked NMF for learning on weakly labeled audio data'

Sobieraj, Iwona, Rencker, Lucas and Plumbley, Mark D (2018) Orthogonality-regularized masked NMF for learning on weakly labeled audio data In: IEEE ICASSP 2018, 15 - 20 April 2018, Calgary, Alberta, Canada.

More...

New publication: 'Audio set classification with attention model'

Kong, Qiuqiang, Xu, Yong, Wang, Wenwu and Plumbley, Mark D. (2018) Audio set classification with attention model: a probabilistic perspective In: ICASSP 2018 - 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 15 - 20 Apr 2018, Calgary, Canada.

More...

New publication: 'Synthesis of images by two-stage generative adversarial networks'

Huang Qiang, Jackson, Philip, Plumbley, Mark D. and Wang, Wenwu (2018) Synthesis of images by two-stage generative adversarial networks In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing, 15–20 Apr 2018, Calgary, Alberta, Canada.

More...

New publication: 'An Evidence-Based Soundscape Taxonomy'

Bones, Oliver, Cox, Trevor J. and Davies, William J. (2017) An Evidence-Based Soundscape Taxonomy. In: 24th International Congress on Sound and Vibration, 23 -27 July 2017 London, UK.

More...

New publication: 'Clang, chitter, crunch: Perceptual organisation of onomatopoeia'

Bones, Oliver, Davies, William J. and Cox, Trevor J. (2017) Clang, chitter, crunch: Perceptual organisation of onomatopoeia The Journal of the Acoustical Society of America 141 (5), 3694-3694

More...

New publication: 'Neuroevolution for sound event detection in real life audio: A pilot study'

Kroos, Christian and Plumbley, Mark D. (2017) Neuroevolution for sound event detection in real life audio: A pilot study. In: DCASE 2017, 16 - 17 November 2017, Munich, Germany.

More...

DCASE 2017 challenge success

Yong Xu, Qiuqiang Kong, Wenwu Wang and Mark Plumbley won the 1st prize in Task 4, ‘large-scale weakly supervised sound event detection for smart cars’, Subtask A, ‘audio tagging’ in the IEEE AASP Challenge on Detection and Classification of Acoustic Scenes and Events (DCASE 2017). The DCASE challenge constitutes the most important challenge in the non-speech audio domain. It is organised by Tampere University of Technology, Carnegie Mellon University and INRIA and sponsored by Google and Audio Analytic. Because of its unique standing, the best players in the field participate such as CMU, New York University, Bosch, USC, TUT, Singapore A*Star, Korean Advanced Institute of Science and Technology, Seoul National University, National Taiwan University and CVSSP.

More...

New publication: 'Masked Non-negative Matrix Factorization for Bird Detection Using Weakly Labelled Data'

Sobieraj, Iwona, Kong, Qiuqiang and Plumbley, Mark D. (2017) Masked Non-negative Matrix Factorization for Bird Detection Using Weakly Labelled Data. In: 25th European Signal Processing Conference, 2017 (EUSIPCO-2017), 28 Aug- 2 Sep 2017, Kos Island, Greece.

More...

New publication: 'Joint Detection and Classification Convolutional Neural Network on Weakly Labelled Bird Audio Detection'

Kong, Qiuqiang, Xu, Yong and Plumbley, Mark D. (2017) Joint Detection and Classification Convolutional Neural Network on Weakly Labelled Bird Audio Detection. In: 25th European Signal Processing Conference (EUSIPCO) 2017, Aug 28 - Sep 2 2017, Kos Island, Greece.

More...

New publication: 'Using deep neural networks to estimate tongue movements from speech face motion'

Kroos, Christian, Bundgaard-Nielsen, RL, Best, CT and Plumbley, Mark D. (2017) Using deep neural networks to estimate tongue movements from speech face motion. In: 14th International Conference on Auditory-Visual Speech Processing (AVSP2017), 25 - 26 August 2017, Stockholm, Sweden.

More...

New publication: 'Learning the Mapping Function from Voltage Amplitudes to Sensor Positions in 3D-EMA Using Deep Neural Networks'

Kroos, Christian and Plumbley, Mark D. (2017) Learning the mapping function from voltage amplitudes to sensor positions in 3D-EMA using deep neural networks. In: Interspeech 2017, 20 - 24 August 2017, Stockholm, Sweden.

More...

New publication: 'Attention and Localization based on a Deep Convolutional Recurrent Model for Weakly Supervised Audio Tagging'

Xu, Yong, Kong, Qiuqiang, Huang, Qiang, Wang, Wenwu and Plumbley, Mark D. (2017) Attention and Localization based on a Deep Convolutional Recurrent Model for Weakly Supervised Audio Tagging. In: Interspeech 2017, 20 - 24 August 2017, Stockholm, Sweden.

More...

In the media: Cheltenham Science Festival - 'Is Your Tech Listening To You?'

Mark Plumbley was joint by security researcher Justin Nurse (Oxford) for a panel session at the Cheltenham Science Festival examining whether intelligent virtual assistants (like Siri, Alexa or Contana) leave us vulnerable to oversharing private information. The panel took place on 8th June 2017 and was chaired by BBC broadcaster Rory Cellan-Jones.

More...