Seolhee Lee


2024

pdf bib
PIAST: A Multimodal Piano Dataset with Audio, Symbolic and Text
Hayeon Bang | Eunjin Choi | Megan Finch | Seungheon Doh | Seolhee Lee | Gyeong-Hoon Lee | Juhan Nam
Proceedings of the 3rd Workshop on NLP for Music and Audio (NLP4MusA)

While piano music has become a significant area of study in Music Information Retrieval (MIR), there is a notable lack of datasets for piano solo music with text labels. To address this gap, we present PIAST (PIano dataset with Audio, Symbolic, and Text), a piano music dataset. Utilizing a piano-specific taxonomy of semantic tags, we collected 9,673 tracks from YouTube and added human annotations for 2,023 tracks by music experts, resulting in two subsets: PIAST-YT and PIAST-AT. Both include audio, text, tag annotations, and transcribed MIDI utilizing state-of-the-art piano transcription and beat tracking models. Among many possible tasks with the multimodal dataset, we conduct music tagging and retrieval using both audio and MIDI data and report baseline performances to demonstrate its potential as a valuable resource for MIR research.