Yuexuan Kong ICASSP 2025

S-KEY: Self-Supervised Learning of Major and Minor Keys from Audio

Communications dans un congrès

Auteurs : Yuexuan Kong, Gabriel Meseguer-Brocal, Vincent Lostanlen, Mathieu Lagrange, Romain Hennequin.

Conférence : IEEE International Conference on Acoustics, Speech and Signal Processing

Date de publication : 2025

Music key estimationSelf-supervised learningMusic information retrieval
Lien vers le dépot HAL

Abstract


STONE, the current method in self-supervised learning for tonality estimation in music signals, cannot distinguish relative keys, such as C major versus A minor. In this article, we extend the neural network architecture and learning objective of STONE to perform self-supervised learning of major and minor keys (S-KEY). Our main contribution is an auxiliary pretext task to STONE, formulated using transposition-invariant chroma features as a source of pseudo-labels. S-KEY matches the supervised state of the art in tonality estimation on FMAKv2 and GTZAN datasets while requiring no human annotation and having the same parameter budget as STONE. We build upon this result and expand the training set of S-KEY to a million songs, thus showing the potential of large-scale self-supervised learning in music information retrieval.