@dasaemjeong.bsky.social
150 followers 240 following 7 posts
MIR / Assistant Prof. @ Sogang University, Seoul /
Posts Media Videos Starter Packs
dasaemjeong.bsky.social
By training a model to generate audio tokens from given score image, the model learn how to read notes from the score image. This led our model to break SOTA for OMR! Vice versa for AMT can work, while the gain was not significant enough compared to the OMR.
dasaemjeong.bsky.social
Score videos are slideshow of audio-aligned score image. Although they does not include any machine-readable symbolic data, we thought these score image - audio pairs can be used for understand each modality, because they share same semantic in (hidden) symbolic music domain.
dasaemjeong.bsky.social
Can we unify these tasks into a single framework? And what would be the benefit of that unification?

Answer: We can exploit tons of Score Video from YouTube!
We collected about 2k hours of score video from YouTube and used 1.3k hours after filtering.
dasaemjeong.bsky.social
Music exists in various modal, and the translation between modality is important MIR Tasks.
Score Image→Symbolic Music: OMR
Audio → MIDI: AMT
MIDI → Audio: Synthesis
Score → Performance MIDI: Performance Rendering
Audio → Music Notation: Complete AMT
dasaemjeong.bsky.social
🎶Now a neural network can read scanned score image and generate performance audio in end-to-end😎
I'm super excited to introduce our work on Unified Cross-modal translation between Score Image, Symbolic Music, and Audio.
Why does it matter and how to make it? Check the thread🧵