Smartsound Cloud isn’t quite AI, but it does use a lot of algorithmic magic under the hood. Here’s the difference.
Mention music and Artificial Intelligence together in a sentence in a crowded room. You might get drastically different reactions depending on who is present. The modern techie might think that AI is the answer to all music. That like with so many other occupations, musicians have their time numbered. AI after all can learn any music genre and repeat it. Musicians on the other hand, would argue that artists are the only ones truly safe from automation, since AI can never truly be a creative source. Repetition is not creation.
There is a third group that argues the two can be combined. It’s mainly composed of marketers like YouTube or Spotify, who see AI as a tool to create playlists or sort genres. There are also composers that see how AI can be useful in augmenting their creativity, through automation or developing and/or finding patterns.
Smartsound seeks to have the most natural sound possible in our stock music. We’ve rejected any influence of Artificial Intelligence. Though what our app does might seem like we use it, it’s really just a lot of hard work and techno wizardry.
Pierre Langer, Smartsound’s chief, says in our latest interview with him, “That’s the main solution for some, to have an artificial intelligence or a neural network, right? Music on the fly and replace composers. And I always thought this idea is not the right choice, because you want to have the creative input and emotional coloring of a composer.”
What is artificial intelligence?
We must understand what AI means in this context. It’s not an army of sentient robots from the future patrolling the streets for hiding musicians, ready to gun them down. Though maybe that is the future, who knows.
What “artificial intelligence” really maens is a concept known as “machine learning”. A computer program can receive data in which it finds patterns, then it repeats those patterns based upon what it has been taught as the best possible output. “Best” being defined by the engineer/programmer’s boss.
Yufeng Guo, a developer for Google Cloud, puts it this way on his easy to understand video: “In machine learning, the goal… is to create an accurate model that answers our questions correctly most of the time.” In Yufeng’s example, they wanted a model that could answer whether a drink was beer or wine. They input a ton of data about both drinks, here color and alcohol percentage. Then the computer can predict whether it’s wine or beer when given that combination in the future.
Another example is in Facebook’s tagging. It’s not magic. It’s not Artificial Intelligence. You’re training the Facebook algorithm what people look like every time you tag someone. The goal of this learning is to develop a model that can, based on the previous data, recognize who is who in a picture. The more data it has – the more tagged pictures – the more accurate it will be. The same goes for deep fake technology.
Can AI create?
In the creative spheres, this has huge drawbacks given the current description of machine learning and artificial intelligence. Artificial intelligence can make things, don’t get me wrong, but it so far can’t really create things.
Take for example Deep-speare, a computer designed to make Shakespearean style poetry. They fed into the AI 2,700 sonnets in an effort to get the AI to be able to craft poetry. “Deep-speare independently learned three sets of rules that pertain to sonnet writing: rhythm, rhyme scheme, and the fundamentals of language,” the creators of the project said in an article for IEEE Spectrum. But upon evaluation with an expert in literature, Adam Hammond, they found that Hammond “could easily tell which poems were generated by Deep-speare.”
For example, one stanza by Deep-speare:
“Yet in a circle pallid as it flow,
by this bright sun, that with his light display,
roll’d from the sands, and half the buds of snow,
And calmly on him shall infold away.”
It’s “nonsensical when you read it closely, but it certainly ‘scans well,’ as an English teacher would say,” the creators admit. “Its rhythm, rhyme scheme, and the basic grammar of its individual lines all seem fine at first glance.”
But what about music?
Music has the same problem as poetry. When AI makes music, it certainly ‘scans well’, and can fool a lot of non-musical people. Music is a much broader and more open form of art than poetry though. That means the ability for an AI to create it is that much more diminished. How can you form meaningful patterns and produce something new when the terms are so wide open?
AIVA and Jukebox are probably the best functioning examples today of an AI producing music. At a TED talk (below), the CEO of AIVA, Pierre Barreau, explains that his team fed into their AI 30,000 pieces from history’s “greatest composers”. The TED talk features an orchestra playing a piece written by AIVA. Upon comparing it to the music available on their site, it’s clear that a composer finessed the piece.
“Music is also a super subjective art,” Barreau says. “And we needed to teach AIVA how to compose the right music for the right person, because people have different preferences.”
Jukebox is probably even more fascinating, as it covers modern genres and singing with surprising alacrity. It uses a similar method, of listening to thousands of other examples and reproducing something that sounds analogous.
It’s all about patterns
It sounds incredible when you first listen to AIVA or Jukebox. But as the creators of Deep-speare point out in their article, there is a “willingness to look past obvious errors in order to marvel at the wonder of AI.” This is known as the Eliza effect, and once listeners are past it, AIVA is much less dazzling.
How are you going to create anything new if you are only studying patterns and writing based on those patterns? And though Barreau claims he’s reviving Beethoven, he’s not. He’s only making something that predicts what sounds acceptable and only makes variations of what’s been written before. When you look at famous classical composers, they’re famous precisely because they broke with their contemporaries and history. They break patterns, they don’t follow them.
Look at Beethoven or Mozart or Chopin. Or at what was happening before them. Look at the modernists who developed the 12-tone technique, and ask yourself if a computer could come up with something that sounds so “bad” in terms of the rules set by predecessors? Or later jazz – something that sounds so different it’s good? If you gave a computer the past twenty years of data, how does it jump to Billie Eilish, Sudan Archives or Ian Chang, based on the data it had from before those musicians?
“You can have a machine compose music,” Pierre says, “but that is like producing forms that it learned, and sometimes it is actually pretty impressive. And in some genres it’s stunning… It sounds like Bach or Beethoven or whoever, but it’s just forms. Music is so rich in all the different styles and influences and then there’s the sound and what instrument is used and why and all these decisions that you have to make that are emotional.”
Does Smartsound use AI?
“We can use tech to make music more ‘intelligent’ and enhance its usability,” Pierre says. “But it’s not AI, it’s software.” All the stock music that Smartsound features has been written by flesh and blood composers. It’s not using any AI in the writing, nor in any of the features that seems like magic.
Of course, the most magical feature on Smartsound Cloud is when a user changes the length of the music, it seems to “rewrite” itself to the appropriate length. But this is due to our composers writing several variations in length and in different parts, called stems. Our software then links the stems together to fit the parameters defined by you, the user. It’s a seriously amazing effect to watch and play with, but it’s not Artificial Intelligence.