Preview Mode Links will not work in preview mode

Feb 11, 2020

AI is being used by music groups, such as our guest this episode Claire Evans, a member of the band YACHT. Their latest album, Chain Tripping, leveraged machine learning solutions for the music, lyrics, and more.

Artists are making the most of machine learning, using the technology both in the creation of their art and as a cultural touchpoint for expression, exploration, and commentary. While the Internet and more modern emerging technologies have long had a negative impact on musicians and others who create using audio, Claire Evans and her band YACHT - Young Americans Challenging High Technology - are at the vanguard of discovering how these technologies will impact art and music in the future.

 

Memorable Quotes

“Since we only really learn by doing things and making things, we figured that the most efficient way for us to get a sort of bodily understanding of what the hell AI is and what it's doing and what it means for artists and for all of us was to try to make something with it.”

“I think when we first started this project, we naively thought we could just kind of hand our back catalog to some algorithm, and the algorithm would analyze that and spit out new songs that would be new YACHT songs. And the project, the art, would be about committing to that, whatever it was. As soon as we started working on this, we realized that we're not there yet, thank God. Algorithms can't just spit out pop songs. If they could, the airwaves would be full of them.”

“If you listen to the record it sounds like an interesting experimental rock or pop record. It doesn't sound like generative, you know, plausible nonsense. It sounds like songs, and that's because there was very much a human in the loop. We used the machine learning model to facilitate the process of generating source material, and then from that source material we built songs the way that we would always build songs as humans in a studio playing music.”

“I was projecting my own meaning onto words that I didn't write. And trying to sort of cobble together some kind of meaning to the songs that made it possible for me to sort of perform and convey them with my voice. And so, it's oddly democratizing, because now the fans, the listeners, and the band, are all trying to figure out what it all means at the same time. And we were going to have as many interpretations of what it means as there are people to listen to it.”

“It also has no consideration of the body, right. It doesn't ‘know’ what it feels like to play any of these melodies on the guitar or on the keyboard. If it's physically challenging to do. All it knows is the MIDI data that it's been fed in the training process. So, a lot of these melodies sounded odd, but simple enough to play. But then when we sat down to actually play them, we found that they were extremely challenging, because they forced us to acknowledge the embodied habits that we bring with us as players into the studio.”

“I like to think of some of these machine learning models being like a camera of their individual disciplines. I mean, a text-generating model that's able to make perfect texts. Maybe that just becomes the camera of writing. And we have to completely step outside of our comfort zone to reinvent what writing means in the 21st century. And what an exciting proposition that is for an artist.”

“There's also something really interesting about the reflective quality of AI as it works today. I mean, you build a machine learning model by feeding it lots of information, trading data. And in the context of music that information is historic. It's the history of music. It's a corpus of millions of notes, or a corpus of millions of words, of song lyrics from musicians and artists that we love. Or ourselves. So this idea that we could use an emerging technology not only learn to understand it, but also maybe learn something about ourselves in the process.”

“Maybe in ten years we won't even be making music for people anymore. Maybe we'll just be making music for other AI's to listen to.”

“Probably we'll get to a place, where machine learning models in some combination are able to generate any song that sounds like a song a human wrote. Or a novel that reads like a novel a human wrote.”

 

“In two or three years, who knows exactly when, we will be at a place where text generating models are able to generate texts that is effectively indistinguishable from human written texts. Arguably we're there already.”

“I think we're in a really interesting moment right now, where some of these tools are just now becoming kind of artist-friendly enough to even be useful or usable to people who don't have hardcore technical backgrounds. And, I think we're going to see an efflorescence of really interesting creative material emerge out of that. And the more sort of democratic these tools get, the more unpredictable it will be.”

“The future doesn't feel vast. The future doesn't feel infinite to me. Like on an individual personal level. The past feels infinite to me. I think that's one of the things that I find kind of comforting about machine learning is, as it's structured, it's not about the future. It's not something that scorches the past and makes something new. It's something that depends on the past.”

“I think in a few years, we will all be nostalgic for the times in which the AI models were not completely perfect. And it will be kind of like the analog of AI. Like, people will be putting on affectations of wonky AI in the same way that artists now record on tape or we have this fetish for vinyl; where we use iPhone camera filters that look like old VHS video.”

“We were really interested in this idea that you could take ‘meaningless material’ and give it meaning through performance.”

“It feels more true to who we are than anything we've ever made, even though this new weird variable is in the mix.”

“The personal computer represented a great deal of freedom for independent artists. Not just in terms of how music is produced but also how music is distributed, and all of the other things that touch that. Like, you know, artwork and messaging and video and text and communication and all of the things which bands have to think about and do, computers simplified that a great deal.”

 

Who You'll Hear

Dirk Knemeyer, Social Futurist and Producer of Creative Next (@dknemeyer)

Jonathan Follett, Writer, Electronic Musician, Emerging Tech Researcher and Producer of Creative Next (@jonfollett)

Claire Evans, Author and Musician (@TheUniverse)

 

Join The Conversation

Website & Newsletter: www.creativenext.org

Twitter: @GoCreativeNext

Facebook: /GoCreativeNext

Instagram: @GoCreativeNext

 

Sponsors

GoInvo, A design practice dedicated to innovation in healthcare whose clients are as varied as AstraZeneca, 3M Health Information Services, and the U.S. Department of Health and Human Services. www.goinvo.com 

Design Museum Foundation, A new kind of museum, they believe design can change the world. They’re online, nomadic, and focused on making design accessible to everyone. Their mission: bring the transformative power of design everywhere. You can learn about their exhibitions, events, magazine, and more. www.designmuseumfoundation.org

BIF, As a purpose-driven firm, BIF is committed to bringing design strategy where it is needed most - health care, education, and public service to create value for our most vulnerable populations. www.bif.is