Twenty-five years ago, the Music Genome Project® was conceived as a solution to the cold start problem in music recommendation. A dedicated team of trained musicologists has spent a quarter century painstakingly analyzing millions of songs — categorizing instrumentation, genre, and mood, mapping rhythmic, harmonic, and melodic structures, and even conducting lyrical analysis.
To date, we’ve analyzed over 2.2 million songs. While this is an impressive achievement, it represents only a small fraction of all the music that exists. So, what’s next for the Music Genome Project?
The rich, detailed metadata we’ve curated over the decades allows us to extrapolate insights about the deep catalog—the tens of millions of songs we haven’t manually analyzed yet. With sophisticated machine learning techniques, we can now derive valuable metadata from these vast, untouched portions of our catalog.
Here’s how it works:
Inferring Labels for Unanalyzed Songs
Once our model can accurately predict tags for these audio embeddings, we apply it to tracks that haven’t yet been analyzed by human experts, equipping us with new metadata at scale so we can deliver the best recommendations to our listeners.
A few things to point out:
Our mission is to understand the DNA of the music we have—not just the hits, but the hidden gems waiting to be discovered. While there are some very popular songs that millions of people listen to, music taste is deeply personal. Some of the most meaningful discoveries happen in the long tail—the vast catalog of rarely played tracks that might be perfect for one particular listener.
By combining expert human analysis with AI-driven content understanding, we can surface those undiscovered songs—the ones that might never be mainstream, but could be exactly what you’re looking for.
This is how we bring the perfect song to the perfect listener—one track at a time.
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.