Have a personal or library account? Click to login
The AI Music Arms Race: On the Detection of AI-Generated Music Cover

The AI Music Arms Race: On the Detection of AI-Generated Music

Open Access
|Jun 2025

Figures & Tables

Table 1

Mean and standard deviation values of selected Essentia descriptors across sources: Suno, Udio, and MSD.

FeatureSunoUdioMSD
Average Loudness
Spectral Centroid
BPM
Duration (s)
Pitch Salience
Dynamic Complexity
Tuning Diatonic Strength
Chord Change Rate
tismir-8-1-254-g1.png
Figure 1

Distributions of selected Essentia descriptors across sources. The mean and standard deviations of these descriptors are provided in Table 1

tismir-8-1-254-g2.png
Figure 2

Two‑dimensional UMAP projection of Contrastive Language–Audio Pretraining audio embeddings from our dataset, with color denoting source: Suno (blue), Udio (red), and MSD (green).

Table 2

Parent‑level classification results (AI vs. non‑AI) on the sample set.

ClassifierClassPrecisionRecallF1 ScoreSupport
SVMNon‑AI0.9580.9130.935150
AI0.9580.9800.969300
RFNon‑AI0.9550.8400.894150
AI0.9250.9800.951300
5‑NNNon‑AI0.9610.8200.885150
AI0.9160.9830.949300
IRCAM AmplifyNon‑AI1.0000.9530.976150
AI0.9771.0000.988300
SpecTTTraNon‑AI0.5280.8930.663150
AI0.9180.6000.726300
Table 3

Normalized confusion matrices for various classifiers on the sample set, classified at the parent level.

Predicted TrueSVMRF5‑NNIRCAM AmplifySpecTTTra
Non‑AIAINon‑AIAINon‑AIAINon‑AIAINon‑AIAI
MSD0.9130.0870.8400.1600.8200.1800.9530.0470.8930.107
Suno0.0070.9930.0001.0000.0001.0000.0001.0000.0470.953
Udio0.0330.9670.0400.9600.0330.9670.0001.0000.7530.247
Table 4

Child‑level classification results (MSD, Suno, Udio) on the sample set.

ClassifierClassPrecisionRecallF1 ScoreSupport
SVMMSD0.9510.9070.928150
Suno0.9700.8670.915150
Udio0.8150.9400.873150
RFMSD0.9470.8330.887150
Suno0.9430.7670.846150
Udio0.6990.9130.792150
5‑NNMSD0.9530.8130.878150
Suno0.7140.9800.826150
Udio0.7760.6000.677150
Table 5

Cross‑domain classification results showing performance when training on one AI source plus MSD and testing on another AI source. Results are grouped by classifier type and ordered by F1 score within each group.

ClassifierTrain TestPrecisionRecallF1 Score
SVMSuno  Suno0.9950.9950.995
Udio  Suno0.9720.9720.972
Udio  Udio0.9710.9710.971
Suno  Udio0.7950.6730.629
RFSuno  Suno0.9880.9880.988
Udio  Suno0.9560.9550.955
Udio  Udio0.9490.9480.948
Suno  Udio0.8150.7350.713
5‑NNSuno  Suno0.9850.9850.985
Udio  Suno0.9430.9400.940
Udio  Udio0.9360.9340.934
Suno  Udio0.8340.7880.778
tismir-8-1-254-g3.png
Figure 3

Impact of low‑pass (left) and high‑pass (right) filtering on the test set. This figure shows the F1 scores for three classifiers (support vector machine, random forest, 5‑NN) as a function of cut‑off frequency.

tismir-8-1-254-g4.png
Figure 4

Impact of low‑pass (left) and high‑pass (right) filtering on the Udio test set. This figure shows the F1 scores for three classifiers (support vector machine, random forest, 5‑NN) as a function of cut‑off frequency.

tismir-8-1-254-g5.png
Figure 5

Impact of low‑pass (left) and high‑pass (right) filtering on the Suno test set. This figure shows the F1 scores for three classifiers (support vector machine, random forest, 5‑NN) as a function of cut‑off frequency.

tismir-8-1-254-g6.png
Figure 6

Impact of low‑pass (left) and high‑pass (right) filtering on the Million Song Dataset test set. This figure shows the false‑positive rate for three classifiers (support vector machine, random forest, 5‑NN) as a function of cut‑off frequency.

DOI: https://doi.org/10.5334/tismir.254 | Journal eISSN: 2514-3298
Language: English
Submitted on: Jan 27, 2025
Accepted on: May 22, 2025
Published on: Jun 25, 2025
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2025 Laura Cros Vila, Bob L. T. Sturm, Luca Casini, David Dalmazzo, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.