Music Emotion Recognition using Deep Neural Networks

dc.contributor.authorBathigama, T.H.
dc.contributor.authorMadushika, M.K.S.
dc.date.accessioned2022-09-02T05:00:12Z
dc.date.available2022-09-02T05:00:12Z
dc.date.issued2021
dc.description.abstractEmotion is an integral part of music and a complex aspect of music that is not easily understood by machines. The emotional aspect of music is further complicated by the fact that it is a subjective experience that cannot be easily conveyed to machine. Although it is a complex problem, some progress has been made in this area suggesting that it might be feasible to develop computation models that can be used in real-world applications. Real-world applications of music emotion recognition systems range from entertainment to healthcare. In this paper we introduce a deep learning model that recognizes emotion in music from the audio signal. 1d and 2d convolution layers with different kernel sizes have been tested. Adaptive pooling layers have also been used to extract a fixed feature representation for the dense layers. We have also used trainable spectrogram extractors to learn different representations of the audio. To address the lack of data for the task of music emotion recognition we have also used the latest trends in audio data augmentation and converted it for music data. Till now we have been able to achieve an accuracy of about 0.92 for the PMEmo dataset and about 0.6 F-1 score from using the raw audio signal and 1D convolution layers to extract features. Preliminary experiments show that using 1d convolutions with the combination of learnable spectrograms performs satisfactorily. Further experiments are to be conducted using different combinations of raw audio and calculated features. Different model architectures using recurrent networks are also to be tested considering that audio has temporal relationship between each unit of time. Finally, the work done in this study is mainly to explore the high dimensional feature space of raw audio to extract features which can contribute to the recognition of emotion in music using automated methods such as convolution and recurrent layers. Keywords: Music Emotion Recognition; Deep Neural Networks; Music Data Augmentation; Arousal and Valence Predictionen_US
dc.identifier.isbn978-624-5856-04-6
dc.identifier.urihttp://www.erepo.lib.uwu.ac.lk/bitstream/handle/123456789/9597/Page%20132%20-%20IRCUWU2021-457%20-Bathigama-Music%20Emotion%20Recognition%20Using%20Deep%20Neural%20Networks.pdf?sequence=1&isAllowed=y
dc.language.isoenen_US
dc.publisherUva Wellassa University of Sri Lankaen_US
dc.subjectComputing and Information Scienceen_US
dc.subjectMusicen_US
dc.subjectNeural Networksen_US
dc.subjectComputer Scienceen_US
dc.titleMusic Emotion Recognition using Deep Neural Networksen_US
dc.title.alternativeInternational Research Conference 2021en_US
dc.typeOtheren_US
Files
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
Page 132 - IRCUWU2021-457 -Bathigama-Music Emotion Recognition Using Deep Neural Networks.pdf
Size:
145.12 KB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: