In the next Machine Learning Group seminar, we have a talk by Siddharth Sigtia who is a 2nd year Ph.D. student at the Centre for Digital Music, Queen Mary University of London (and a visiting research student at City).
Date & Time: Wednesday 30th April, 13:00-14:30
Location: Room AG06 (College Building)
Title: Improved Music Feature Learning with Deep Neural Networks
Abstract: Recent advances in neural network training provide a way to efficiently learn representations from raw data. Good representations are an important requirement for Music Information Retrieval (MIR) tasks to be performed successfully. However, a major problem with neural networks is that training time becomes prohibitive for very large datasets and the learning algorithm can get stuck in local minima for very deep and wide network architectures. In this paper we examine 3 ways to improve feature learning for audio data using neural networks: 1.using Rectified Linear Units (ReLUs) instead of standard sigmoid units; 2.using a powerful regularisation technique called Dropout; 3.using Hessian-Free (HF) optimisation to improve training of sigmoid nets. We show that these methods provide significant improvements in training time and the features learnt are better than state of the art hand- crafted features, with a genre classification accuracy of 83 ± 1.1% on the Tzanetakis (GTZAN) dataset. We found that the rectifier networks learnt better features than the sigmoid networks. We also demonstrate the capacity of the features to capture relevant information from audio data by applying them to genre classification on the ISMIR 2004 dataset.
Link to paper: http://www.eecs.qmul.ac.uk/~sss31/Pubs/ICASSP-2014.pdf