CultureMERT: Continual Pre-Training for Cross-Cultural Music Representation Learning
Angelos-Nikolaos Kanatas, Charilaos Papaioannou, Alexandros Potamianos
2025-06-24
Summary
This paper talks about CultureMERT-95M, a foundation model that learns to understand and represent music from many different cultures, including Greek, Turkish, and Indian music, by using a special two-step training process.
What's the problem?
The problem is that most music models are mainly trained on Western music and don’t work well with music from other cultures, which limits their ability to understand the diversity and unique features in world music.
What's the solution?
The researchers used a two-stage continual pre-training strategy where they kept training an existing music model on a mix of non-Western music while carefully adjusting the training speed to avoid forgetting what it learned before. This helped the model adapt and perform better on diverse music styles without losing its knowledge of Western music.
Why it matters?
This matters because it makes AI better at recognizing and understanding music from different cultures, which can support music research, improve music recommendation systems, and help preserve and celebrate global musical diversity.
Abstract
CultureMERT-95M, a multi-culturally adapted foundation model, enhances cross-cultural music representation learning with a two-stage continual pre-training strategy, demonstrating superior performance in diverse non-Western music tasks.