Dataset augmentation techniques have been widely used to achieve state-of-the-art results in Music Information Retrieval tasks. However, their application in music emotion recognition (MER) remains underexplored. MER methods are particularly relevant to the design of smart musical instruments (SMIs), as emotionally aware SMIs have the potential to enrich musical interaction by providing feedback to musicians or dynamically adjusting their sound properties. In this study, we analyze the effect of 11 augmentation techniques on emotion classification in guitar recordings using a convolutional neural network. Our dataset consists of approximately 400 guitar recordings labeled with four emotions: aggressiveness, relaxation, happiness, and sadness. Results indicate that time shift, time stretch, and pitch shift provide the most significant improvements in classification accuracy. Further analysis combining these techniques under different settings yielded similar performance outcomes. A listening test confirmed that the applied augmentations did not significantly alter the perceived emotional content of the recordings. These findings support the development of emotionally aware SMIs by enhancing MER accuracy through data augmentation, ultimately enabling more expressive and interactive music-making experiences.

Advancing guitar emotion recognition through audio data augmentation to enhance smart musical instruments / Rossi, Michele; Iacca, Giovanni; Turchet, Luca. - In: EURASIP JOURNAL ON AUDIO, SPEECH AND MUSIC PROCESSING. - ISSN 1687-4722. - 1:38(2025). [10.1186/s13636-025-00426-1]

Advancing guitar emotion recognition through audio data augmentation to enhance smart musical instruments

Rossi, Michele;Iacca, Giovanni;Turchet, Luca
2025-01-01

Abstract

Dataset augmentation techniques have been widely used to achieve state-of-the-art results in Music Information Retrieval tasks. However, their application in music emotion recognition (MER) remains underexplored. MER methods are particularly relevant to the design of smart musical instruments (SMIs), as emotionally aware SMIs have the potential to enrich musical interaction by providing feedback to musicians or dynamically adjusting their sound properties. In this study, we analyze the effect of 11 augmentation techniques on emotion classification in guitar recordings using a convolutional neural network. Our dataset consists of approximately 400 guitar recordings labeled with four emotions: aggressiveness, relaxation, happiness, and sadness. Results indicate that time shift, time stretch, and pitch shift provide the most significant improvements in classification accuracy. Further analysis combining these techniques under different settings yielded similar performance outcomes. A listening test confirmed that the applied augmentations did not significantly alter the perceived emotional content of the recordings. These findings support the development of emotionally aware SMIs by enhancing MER accuracy through data augmentation, ultimately enabling more expressive and interactive music-making experiences.
2025
38
Rossi, Michele; Iacca, Giovanni; Turchet, Luca
Advancing guitar emotion recognition through audio data augmentation to enhance smart musical instruments / Rossi, Michele; Iacca, Giovanni; Turchet, Luca. - In: EURASIP JOURNAL ON AUDIO, SPEECH AND MUSIC PROCESSING. - ISSN 1687-4722. - 1:38(2025). [10.1186/s13636-025-00426-1]
File in questo prodotto:
File Dimensione Formato  
Rossi_et_al-2025-EURASIP_Journal_on_Audio,_Speech,_and_Music_Processing.pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Creative commons
Dimensione 1.84 MB
Formato Adobe PDF
1.84 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/467490
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
  • OpenAlex 0
social impact