Dataset augmentation techniques have been widely used to achieve state-of-the-art results in Music Information Retrieval tasks. However, their application in music emotion recognition (MER) remains underexplored. MER methods are particularly relevant to the design of smart musical instruments (SMIs), as emotionally aware SMIs have the potential to enrich musical interaction by providing feedback to musicians or dynamically adjusting their sound properties. In this study, we analyze the effect of 11 augmentation techniques on emotion classification in guitar recordings using a convolutional neural network. Our dataset consists of approximately 400 guitar recordings labeled with four emotions: aggressiveness, relaxation, happiness, and sadness. Results indicate that time shift, time stretch, and pitch shift provide the most significant improvements in classification accuracy. Further analysis combining these techniques under different settings yielded similar performance outcomes. A listening test confirmed that the applied augmentations did not significantly alter the perceived emotional content of the recordings. These findings support the development of emotionally aware SMIs by enhancing MER accuracy through data augmentation, ultimately enabling more expressive and interactive music-making experiences.
Advancing guitar emotion recognition through audio data augmentation to enhance smart musical instruments / Rossi, Michele; Iacca, Giovanni; Turchet, Luca. - In: EURASIP JOURNAL ON AUDIO, SPEECH AND MUSIC PROCESSING. - ISSN 1687-4722. - 1:38(2025). [10.1186/s13636-025-00426-1]
Advancing guitar emotion recognition through audio data augmentation to enhance smart musical instruments
Rossi, Michele;Iacca, Giovanni;Turchet, Luca
2025-01-01
Abstract
Dataset augmentation techniques have been widely used to achieve state-of-the-art results in Music Information Retrieval tasks. However, their application in music emotion recognition (MER) remains underexplored. MER methods are particularly relevant to the design of smart musical instruments (SMIs), as emotionally aware SMIs have the potential to enrich musical interaction by providing feedback to musicians or dynamically adjusting their sound properties. In this study, we analyze the effect of 11 augmentation techniques on emotion classification in guitar recordings using a convolutional neural network. Our dataset consists of approximately 400 guitar recordings labeled with four emotions: aggressiveness, relaxation, happiness, and sadness. Results indicate that time shift, time stretch, and pitch shift provide the most significant improvements in classification accuracy. Further analysis combining these techniques under different settings yielded similar performance outcomes. A listening test confirmed that the applied augmentations did not significantly alter the perceived emotional content of the recordings. These findings support the development of emotionally aware SMIs by enhancing MER accuracy through data augmentation, ultimately enabling more expressive and interactive music-making experiences.| File | Dimensione | Formato | |
|---|---|---|---|
|
Rossi_et_al-2025-EURASIP_Journal_on_Audio,_Speech,_and_Music_Processing.pdf
accesso aperto
Tipologia:
Versione editoriale (Publisher’s layout)
Licenza:
Creative commons
Dimensione
1.84 MB
Formato
Adobe PDF
|
1.84 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione



