Summary: While machine learning technology is not too accurate at distinguishing between dance styles, when it comes to recognizing the personal style of dancers, it is 97% effective at determining who is who.
Source: University of Jyvaskyla
Nearly everyone responds to music with movement, whether through subtle toe-tapping or an all-out boogie. A recent discovery shows that our dance style is almost always the same, regardless of the type of music, and a computer can identify the dancer with astounding accuracy.
Studying how people move to music is a powerful tool for researchers looking to understand how and why music affects us the way it does. Over the last few years, researchers at the Centre for Interdisciplinary Music Research at the University of Jyväskylä in Finland have used motion capture technology–the same kind used in Hollywood–to learn that your dance moves say a lot about you, such as how extroverted or neurotic you are, what mood you happen to be in, and even how much you empathize with other people.
Recently, however, they discovered something that surprised them. “We actually weren’t looking for this result, as we set out to study something completely different,” explains Dr. Emily Carlson, the first author of the study.
“Our original idea was to see if we could use machine learning to identify which genre of music our participants were dancing to, based on their movements.”
The 73 participants in the study were motion captured dancing to eight different genres: Blues, Country, Dance/Electronica, Jazz, Metal, Pop, Reggae and Rap. The only instruction they received was to listen to the music and move any way that felt natural. “We think it’s important to study phenomena as they occur in the real world, which is why we employ a naturalistic research paradigm,” says Professor Petri Toiviainen, the senior author of the study.
The researchers analysed participants’ movements using machine learning, trying to distinguish between the musical genres. Unfortunately, their computer algorithm was able to identify the correct genre less that 30% of the time. They were shocked to discover, however, that the computer could correctly identify which of the 73 individuals was dancing 94% of the time. Left to chance (that is, if the computer had simply guessed without any information to go on), the expected accuracy would be less than 2%. “It seems as though a person’s dance movements are a kind of fingerprint,” says Dr. Pasi Saari, co-author of the study and data analyst. “Each person has a unique movement signature that stays the same no matter what kind of music is playing.”
Some genres, however, had more effect on individual dance movements than others. The computer was less accurate in identifying individuals when they were dancing to Metal music. “There is a strong cultural association between Metal and certain types of movement, like headbanging,” Emily Carlson says. “It’s probable that Metal caused more dancers to move in similar ways, making it harder to tell them apart.”
Does this mean that face-recognition software will soon be joined by dance-recognition software? “We’re less interested in applications like surveillance than in what these results tell us about human musicality,” Carlson explains. “We have a lot of new questions to ask, like whether our movement signatures stay the same across our lifespan, whether we can detect differences between cultures based on these movement signatures, and how well humans are able to recognize individuals from their dance movements compared to computers. Most research raises more questions than answers,” she concludes, “and this study is no exception.”
About this deep learning research article
Source: University of Jyvaskyla Media Contacts: Emily Carlson – University of Jyvaskyla Image Source: The image is in the public domain.
Dance to your own drum: Identification of musical genre and individual dancer from motion capture using machine learning
Machine learning has been used to accurately classify musical genre using features derived from audio signals. Musical genre, as well as lower-level audio features of music, have also been shown to influence music-induced movement, however, the degree to which such movements are genre-specific has not been explored. The current paper addresses this using motion capture data from participants dancing freely to eight genres. Using a Support Vector Machine model, data were classified by genre and by individual dancer. Against expectations, individual classification was notably more accurate than genre classification. Results are discussed in terms of embodied cognition and culture.