This paper investigates the real-time detection of predefined monophonic patterns from the MIDI output of a digital musical instrument. This enables the development of instruments and systems for live music, which can recognize when a musician plays a certain phrase and repurpose such information to trigger external peripherals connected to the instrument. Specifically, we compare the recognition performance of Dynamic Time Warping and Recurrent Neural Network-based approaches. We employ different representation formats of musical data to optimize the efficiency of each computational method. To evaluate the algorithms, a novel dataset is introduced which includes recordings from 20 keyboard players and 20 guitar players. The evaluation focuses on the algorithms' ability to recognize patterns amid variations that impede a straightforward one-to-one comparison. The results reveal that both methods perform well in detecting up to 3 distinct patterns. However, as the number of different patterns increases up to 10, dynamic time warping exhibits a negative correlation with the recognition performance, while the recurrent neural network maintains high detection accuracy of approximately 98%. Taken together, our findings demonstrate the potential of machine learning in handling complex musical patterns in real-time, paving the way for novel applications involving smart musical instruments.

Real-Time Pattern Recognition of Symbolic Monophonic Music / Silva, Nishal Stanislaus; Turchet, Luca. - (2024), pp. 308-317. ( 19th International Audio Mostly Conference, Audio Mostly 2024 Milano 2024) [10.1145/3678299.3678329].

Real-Time Pattern Recognition of Symbolic Monophonic Music

Nishal Stanislaus Silva;Luca Turchet
2024-01-01

Abstract

This paper investigates the real-time detection of predefined monophonic patterns from the MIDI output of a digital musical instrument. This enables the development of instruments and systems for live music, which can recognize when a musician plays a certain phrase and repurpose such information to trigger external peripherals connected to the instrument. Specifically, we compare the recognition performance of Dynamic Time Warping and Recurrent Neural Network-based approaches. We employ different representation formats of musical data to optimize the efficiency of each computational method. To evaluate the algorithms, a novel dataset is introduced which includes recordings from 20 keyboard players and 20 guitar players. The evaluation focuses on the algorithms' ability to recognize patterns amid variations that impede a straightforward one-to-one comparison. The results reveal that both methods perform well in detecting up to 3 distinct patterns. However, as the number of different patterns increases up to 10, dynamic time warping exhibits a negative correlation with the recognition performance, while the recurrent neural network maintains high detection accuracy of approximately 98%. Taken together, our findings demonstrate the potential of machine learning in handling complex musical patterns in real-time, paving the way for novel applications involving smart musical instruments.
2024
AM '24: Proceedings of the 19th International Audio Mostly Conference: Explorations in Sonic Cultures
1601 Broadway, 10th Floor, NEW YORK, NY, UNITED STATES
Association for Computing Machinery
9798400709685
Silva, Nishal Stanislaus; Turchet, Luca
Real-Time Pattern Recognition of Symbolic Monophonic Music / Silva, Nishal Stanislaus; Turchet, Luca. - (2024), pp. 308-317. ( 19th International Audio Mostly Conference, Audio Mostly 2024 Milano 2024) [10.1145/3678299.3678329].
File in questo prodotto:
File Dimensione Formato  
Real-time pattern recognition of symbolic monophonic music.pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Creative commons
Dimensione 895.72 kB
Formato Adobe PDF
895.72 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/443134
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? 1
  • OpenAlex ND
social impact