Skip to main navigation Skip to search Skip to main content

음고 개수 정보 활용을 통한 기계학습 기반 자동악보전사 모델의 성능 개선 연구

Translated title of the contribution: A study on improving the performance of the machine-learning based automatic music transcription model by utilizing pitch number information

Research output: Contribution to journalArticlepeer-review

Abstract

In this paper, we study how to improve the performance of a machine learning-based automatic music transcription model by adding musical information to the input data. Where, the added musical information is information on the number of pitches that occur in each time frame, and which is obtained by counting the number of notes activated in the answer sheet. The obtained information on the number of pitches was used by concatenating it to the log mel-spectrogram, which is the input of the existing model. In this study, we use the automatic music transcription model included the four types of block predicting four types of musical information, we demonstrate that a simple method of adding pitch number information corresponding to the music information to be predicted by each block to the existing input was helpful in training the model. In order to evaluate the performance improvement proceed with an experiment using MIDI Aligned Piano Sounds (MAPS) data, as a result, when using all pitch number information, performance improvement was confirmed by 9.7 % in frame-based F1 score and 21.8 % in note-based F1 score including offset.

Translated title of the contributionA study on improving the performance of the machine-learning based automatic music transcription model by utilizing pitch number information
Original languageKorean
Pages (from-to)207-213
Number of pages7
JournalJournal of the Acoustical Society of Korea
Volume43
Issue number2
DOIs
StatePublished - 2024

Keywords

  • Automatic music transcription
  • Machine learning
  • Pitch number information
  • Polyphonic traanscription

Fingerprint

Dive into the research topics of 'A study on improving the performance of the machine-learning based automatic music transcription model by utilizing pitch number information'. Together they form a unique fingerprint.

Cite this