A fundamental and general representation of audio and music which integrates multi-modal data sources is important for both application and basic research purposes. In this paper we address this challenge by proposing a multi-modal version of the Latent Dirichlet Allocation model which provides a joint latent representation. We evaluate this representation on the Million Song Dataset by integrating three fundamentally different modalities, namely tags, lyrics, and audio features. We show how the resulting representation is aligned with common 'cognitive' variables such as tags, and provide some evidence for the common assumption that genres form an acceptable categorization when evaluating latent representations of music. We furthermore quantify the model by its predictive performance in terms of genre and style, providing benchmark results for the Million Song Dataset.
I E E E International Conference on Acoustics, Speech and Signal Processing. Proceedings, 2013, p. 3168-3172
Signal Processing and Analysis
Main Research Area:
IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP 2013)International Conference on Acoustics, Speech and Signal Processing