Free Republic
Browse · Search
General/Chat
Topics · Post Article

Skip to comments.

Understanding Song Popularity with Automated Machine Learning (short, coarse pop makes money)
Data Robot ^ | October 24, 2019 | Taylor Larkin

Posted on 11/12/2019 6:35:15 PM PST by DoodleBob

Spotify made waves after it introduced its music service to the masses in 2008, allowing users to stream music as opposed to buying a record or purchasing a song on iTunes. Hailing nearly twice as many subscribers as competitor Apple music. Spotify continues to be a primary source for those who want to listen to music without having to buy every song or album desired.

One of the cool aspects about the former startup is that much of their music information (artists, albums, tracks, etc) is exposed via a web API. This includes attributes for each song such as its tempo, how much energy it has, what key it’s in, etc. Personally, I’ve always been fascinated about the mechanics behind a popular song. What are their general characteristics? Do they normally have more high-energy beats? A certain length? Catchy names? I set out to investigate these things empirically using DataRobot, an automated machine learning platform.

 

Data Collection

Using the R package spotifyr, I pulled the discography from artists that appeared in the Billboard Year End Chart - Hot 100 Songs list dating back to 2010. This chart ranks artists based on radio airplay, sales data, and streaming activity across all genres. Naturally, while these artists will have popular songs, not all of their music are hits. Hence, modeling the popularity of all their songs from 2010-2017 will allow us to investigate what separates popular tracks from not-so-popular songs from the most well-known names in music.

For my target, I use the “popularity” measure calculated by Spotify. This is a score from 0 to 100, with more popular tracks having a higher value. As for my features, I include the artist’s name as well as various attributes about the track such as its genre, lyrics, title, what day of the week it was released, and its audio features.

 

Exploratory Data Analysis

An obvious area to explore first is the genre. Each song in the dataset has a list of several categories that it falls into, as determined by Spotify. Using one of DataRobot’s text modelers, I generated a word cloud to identify which genres  are most related to popularity. The larger the word, the more frequently it occurs. The more red the word, the more likely that it’ll be a popular song while the bluer words, are less likely to be popular.

 

 

Based on the data, it appears that “rap” and “pop” music generally denote a more popular track while hybrid genres such as “pop rap” or “dance pop” do not seem to garner as much interest. While it may seem confusing that “trap music” is so blue due to its prevalence in the hip-hop community, a possible reasoning for this lies in another word cloud: the song lyrics. Due to explicit content, the word cloud isn’t shown here, but it does reveal an interesting finding: explicit words are very polarizing, being either really red or really blue. Because of this insight, I decided to add some additional features related to the amount of profanity in a song and the general sentiment.

 

Modeling and Insights

With the help of DataRobot R package, I built over 400 models, each automatically tuned in less than an hour. To create a more powerful approach, I combined the predictions from some of these models to build what is known as a “blender” in DataRobot, which served as my final model for the subsequent analysis.

 

Using a technique called feature impact, my blender  tells me which factors are most important when predicting a song’s popularity. Unsurprisingly, the genre, track lyrics, day of release, and artist are driving forces for predicting popularity. However, we see other attributes also play a role such as the duration and number of profane words present.

 

 

Let’s drill down as to why these other factors are important. Using feature effects, we can analyze the relationship that these particular features have on estimating the popularity of songs, given all the other information in the dataset.

 

 

The above graph indicates a strong, non-linear relationship in the duration of the track, where the least popular songs are either less than two minutes or over five minutes. This peaks right around the infamous 3-minute rule, which is known at the optimal length for a hit song. 

Of the sentiment and profanity-based features, the number of profane words in a song is most important. Its effect is listed below.

 

 

Interpreting this plot, it appears that the more profanity a song has, the more likely it will be considered popular. This pattern is accordant with the recent uptick in explicit content in popular songs. However, it does seem to have a diminishing law of return, beginning to plateau off at around 30 profane words (looking at another effect yields that some of the least popular songs have over 20% of its lyrics considered profanity). Ultimately, based on the historical data, one of the keys to a popular song seems to be striking a balance between explicit content and being totally G-rated.

 

Kendrick Lamar: A Case Study

Now having established some general trends in regards to what makes a popular song, I wanted to apply the blender on one of my favorite albums, DAMN, by Kendrick Lamar. This album not only won a Grammy for Best Rap Album, but also won a Pulitzer Prize for music, making Kendrick the first non-classical or jazz musician to win the award.

Given its notoriety, I was curious if there existed similar types of tracks on the album. One way to accomplish this would be via some kind of clustering algorithm on the raw data I collected; however, one of the issues with clustering is being unable to apply these techniques on more complex datasets such as those with mixed data types (numeric, categorical, and text features), which mine has.

Fortunately, with the help of prediction explanations, we can overcome this issue. Prediction explanations is an interpretive tool in DataRobot that provides insights as to why a model makes a prediction at the row level, shining light onto the “black-box” nature of machine learning algorithms.

These are available for any model built in DataRobot, even for my complicated blender. The beauty behind this approach is that it can numerically quantify the most relevant factors and their effect, for any song I want. After generating this insight for every track on DAMN, I visualize the similarity of the prediction explanations using this approach. 

 

 

The above graph pictures which songs are most similar songs and how many distinct groups (3 in this case) exist based on their prediction explanations. What’s really cool about this result is that the red songs are four of the top five most popular tracks from the album, including “HUMBLE.” and “DNA.”, both of which were in the Billboard Year End Chart - Hot 100 Songs list for 2017. In addition, it puts some of the shortest and least profane tracks in the same group (see like “YAH.”, “BLOOD.”, and “LOVE. FEAT. ZACARI.”). With the help of prediction explanations, the music industry could empirically identify which songs are most likely to be popular and why, allowing for better insights and collaboration with new and established artists.

 

So What?

While this is a fun application of machine learning, it demonstrates a potential role it can have in the music industry. Music producers and radio executives could study what makes a successful song from a data perspective, helping them more fully understand why people listen to the songs that they do. Considering the wide diversity of music available, having easy ways to gain insights like this is paramount. While there’s no replacement for creativity and subject matter expertise, adopting machine learning approaches can aid in these decisions, and ultimately, lead to better music for everyone.

 


TOPICS: Business/Economy; Computers/Internet; Music/Entertainment
KEYWORDS: ai; machinelearning; popmusic
This is an interesting use of 'machine learning' techniques to decompose petabytes of streamed music data into what makes up successful songs.

What seems likely, however, is that these analytics are being used by record companies to continue feeding the public more of what sells at this time: short, trashy pop songs.

This also means, that there is likely a vast, untapped audience waiting for something good to come along. A fortune awaits the next band that cracks that code.

1 posted on 11/12/2019 6:35:16 PM PST by DoodleBob
[ Post Reply | Private Reply | View Replies]

To: ShadowAce

For your perusal.


2 posted on 11/12/2019 6:41:43 PM PST by DoodleBob (Gravity's waiting period is about 9.8 m/s^s)
[ Post Reply | Private Reply | To 1 | View Replies]

To: sauropod

.


3 posted on 11/12/2019 6:48:53 PM PST by sauropod (I am His and He is mine)
[ Post Reply | Private Reply | To 1 | View Replies]

To: DoodleBob

When Neil Sedaka, who was a serious musician as well as pop artist song-writer, started out, he spent a lot of time analyzing the hits of the late ‘50s, looking for common traits, and churning out more of the same. He was very good at it, and kept writing songs that scored big well into the ‘70s (Bad Blood, Love Will Keep Us Together).

I find the word cloud approach is misleading for all but the youngest demographic. We don’t all listen to Spotify.


4 posted on 11/12/2019 6:51:16 PM PST by Dr. Sivana (There is no salvation in politics.)
[ Post Reply | Private Reply | To 1 | View Replies]

To: DoodleBob
Music producers and radio executives could study what makes a successful song from a data perspective, helping them more fully understand why people listen to the songs that they do.

As opposed to actually listening to song to hear if is good. That's why radio has gone downhill in the last decade.

5 posted on 11/12/2019 7:04:41 PM PST by Inyo-Mono
[ Post Reply | Private Reply | To 1 | View Replies]

To: DoodleBob
There are some interesting YouTube videos showing that a lot of pop tunes have the same beat and melodies. So much so that they can be played on top of each other and sync up well.

Giving execs more commonality factors between hit songs will only tend to make modern songs even more indistinguishable.

6 posted on 11/12/2019 7:09:21 PM PST by who_would_fardels_bear
[ Post Reply | Private Reply | To 1 | View Replies]

To: Dr. Sivana
I never understood the value of word clouds. Just the big words stick out.

A list in ascending or descending order is far more clear.

WCs are a fad past the expiration date.

7 posted on 11/12/2019 7:19:15 PM PST by yesthatjallen
[ Post Reply | Private Reply | To 4 | View Replies]

To: DoodleBob

This is a little off topic, but I’ve watched a few “reaction” videos, and have found this “voice coach” fascinating. Here she reacts to Don’t Stop Believing:

https://www.youtube.com/watch?v=kYTPbw_4KfM

Her channel doesn’t have much music I would listen to, but her expressions are fun to watch.


8 posted on 11/13/2019 12:48:58 AM PST by JohnnyP (Thinking is hard work (I stole that from Rush).)
[ Post Reply | Private Reply | To 1 | View Replies]

To: rdb3; Calvinist_Dark_Lord; JosephW; Only1choice____Freedom; martin_fierro; Still Thinking; ...

Tech Ping


9 posted on 11/13/2019 3:52:58 AM PST by ShadowAce (Linux - The Ultimate Windows Service Pack)
[ Post Reply | Private Reply | To 1 | View Replies]

To: DoodleBob

If you like this topic, check out Mark Steyn’s interview with Randy Bachman - fascinating stuff.


10 posted on 11/13/2019 4:00:15 AM PST by Psalm 73 ("I will now proceed to entangle the entire area".)
[ Post Reply | Private Reply | To 1 | View Replies]

To: DoodleBob

I only have two genres. I like it or I don’t.


11 posted on 11/13/2019 4:16:12 AM PST by Lurkina.n.Learnin (If you want a definition of "bullying" just watch the Democrats in the Senate)
[ Post Reply | Private Reply | To 1 | View Replies]

To: JohnnyP
She's great. Very honest and it doesn't require a bot. My fav was her review of Devin Townsend at the Royal Albert Hall. You either love or hate Devin - I'm still deciding, but after one of his vocal runs at 5:30 she says "I need a cigarette break and I don't smoke."
12 posted on 11/13/2019 6:03:01 AM PST by DoodleBob (Gravity's waiting period is about 9.8 m/s^s)
[ Post Reply | Private Reply | To 8 | View Replies]

To: yesthatjallen; Dr. Sivana
Word clouds are simply a tool in the toolbox of visual display approaches for quantitative information. At least in this case they're using size and color to convey information.

This one - covering Napoleon's Russian march and retreat - is generally considered The Godfather of such tools - displaying six dimensions in one picture (and without a computer).


13 posted on 11/13/2019 6:12:35 AM PST by DoodleBob (Gravity's waiting period is about 9.8 m/s^s)
[ Post Reply | Private Reply | To 7 | View Replies]

To: who_would_fardels_bear; Inyo-Mono
This is, in many ways, no different from sports teams using analytics to eeek out an advantage. The music industry is like venture capital - I've read that 85% of all acts lose money, about 10% break even and 5% pay for everyone else. With those stakes, I can't fault the industry for using analytics.

Where they're getting it wrong, however, is focusing on the big words and what sells today. Sure, invest in pop...it pays the bills. But Apple catapulted into where it stands today by selling better computers...it was the iPhone, i.e. Innovation.

This is where these techniques can help - to identify unmet need and then go meet it.

14 posted on 11/13/2019 6:19:22 AM PST by DoodleBob (Gravity's waiting period is about 9.8 m/s^s)
[ Post Reply | Private Reply | To 6 | View Replies]

Disclaimer: Opinions posted on Free Republic are those of the individual posters and do not necessarily represent the opinion of Free Republic or its management. All materials posted herein are protected by copyright law and the exemption for fair use of copyrighted works.

Free Republic
Browse · Search
General/Chat
Topics · Post Article

FreeRepublic, LLC, PO BOX 9771, FRESNO, CA 93794
FreeRepublic.com is powered by software copyright 2000-2008 John Robinson