Recently, I have been seeing many different types of misinformation via media swirling around on the internet, with a new culprit to this societal phenomenon: Artificial Intelligence. Its use has been in place for years, especially in the Hollywood industry, but its access has just recently been given to the general society. UK-based company, ElevenLabs, headed by Zach Silberberg, has been making recent headlines due to its creation of a technology that allows for almost complete voice imitation with little initial reference given. This new development in artificial intelligence has already started off with beat, raising ethical and moral questions due to the many concerning things that many have made with this new technology.
Author John Hendrickson recently wrote an article in The Atlantic in which he interviewed Silberberg about his opinion on his company’s controversial technology. In ElevenLabs’ point of view, their AI audio and video technology was intended to be used for “storytelling”, specifically as an obvious comedic bit. He does acknowledge though, that there has been an unfortunately thin line between storytelling and disinformation/propaganda, especially in the political realm. In recent weeks, these political ai-generated “deepfake”-akin videos center around President Joe Biden. They show him saying that he has not visited East Palestine after that area’s train derailment because he found himself lost on the island from Lost and speaking about the 2011 movie, We Bought a Zoo. While these political figure videos may have comedic intent that could be seen as a simple joke, there is a real worry that this technology could produce dangerous impersonations that are misread. The past few elections were filled with misinformation, culture wars, and propaganda which greatly influenced Americans and gave many outside viewers a poor view on this country. This would likely just be execrated with ElevenLab’s new programs.
When Hendrickson brought this concern to Silberberg, he emphasized that the key to AI deepfakes is that they show the mannerisms of the person that they are misrepresenting. For example, the President’s stuttering due to documented speech disabilities. But, even Silberberg acknowledges that his company’s technology and AI in general are headed in a bad direction:
“My opinion is that, blanket statement, the use of AI technology is pretty bleak. The way that it is headed is scary. And it is already replacing artists, and is already creating really f*****-up, gross scenarios.”
Sure, a joke is fine, but the danger of this technology comes with the intentions behind it. We like to think that people are rational, ethical, and have good intentions, but this naive sentiment puts us between a rock and a hard place. Technology’s evolution is integral to the advancement of society but that evolution may be putting society at risk. As AI technology progresses and gets smoothed out, these videos and audio recordings may become indistinguishable from reality, having us all question each other
These deepfake videos and audios via AI go beyond politics, with some scary results from the use of ElevenLabs’ new technology. Author Kyle Barr from Gizmodo.com writes that this new technology is being used in Hollywood on popular figures. Recently, 4Chan, a popular anonymous English-language image board website, saw its users post various deepfakes on popular stars. Some well-spread examples include Emma Watson was reading Mein Kampf and Rick and Morty creator Justin Roiland assaulting his wife (after his recent criminal charge). In addition, users also posted various manipulated, fake clips of popular animated characters saying offensive and off-putting things.
Author Justin Carter from Gizmodo.com reports that many of the voice actors of these characters being deepfaked have become outspoken about the dangerous effect this technology has on their personal likeness, image, and years of work. As I mentioned earlier, this manipulation of voices has been happening for years in the Hollywood industry, with voice acting being a main focus. Often, there are clauses in the contracts with these actors which basically require them to give away their voice rights which allows for more limited work (and less pay) on their part which production companies can use to create larger forms of media. This “small amount of reference being turned into a large amount of work” aspect is exactly why ElevenLabs’ new AI technology is so potentially dangerous.
ElevenLabs has created such an advanced technology that a 15-30 second clip from TikTok, Instagram, or other media is all that is needed as reference to create longer audio and potential video fakes. Scam calls have been apart of our society for decades, but this new tech has only accelerated their threat to society. According to author Jakob Aylesbury from eTeknix.com, Scammers are actually using ElevenLabs’ capabilities to mimic the voice of loved ones of elderly people. This allows them to easily manipulate and confuse this vulnerable population to being scammed out of money and resources, which is disgusting. While this can be managed like any other scam call from the perspective of phone/service companies, their threat is greatly enhanced and should not be overlooked.
Perhaps the most terrifying thing about ElevenLabs’ capabilities and like technology is that there has been little done to stop the dangers that it brings to society. ElevenLabs is looking to implement account verification or manual checking on these fake videos/audio recordings, but no other action has been considered. I also think there is an argument that social media companies really need to start looking into the affect that this type of misinformation has on their respective platforms. Like much other propaganda, these videos have spread like wildfire on Twitter, Instagram, and TikTok; and while they can be approached similar to other posts with misinformation, that may become harder to differentiate. As this technology becomes more hammered out and loses many of its obvious fake elements, the oversight may become impossible simply due to its nature.
As I have mentioned in my prior two blogs focusing on artificial intelligence, no matter what direction this technology takes, it will have a real and life-changing effect on our society. How we approach it and the level of ethics used may determine the survival of the human race.
*I have not included any of the videos mentioned due to their inappropriate content and intent*
References
- Aylebury, Jakob. “Scammers Using AI Voice Generation to Mimic the Voices of Loved Ones.” ETeknix, 7 Mar. 2023, https://www.eteknix.com/scammers-using-ai-voice-generation-to-mimic-the-voices-of-loved-ones/.
- Barr, Kyle. “Ai Voice Simulator Easily Abused to Deepfake Celebrities Spouting Racism and Homophobia.” Gizmodo, Gizmodo, 30 Jan. 2023, https://gizmodo.com/ai-joe-rogan-4chan-deepfake-elevenlabs-1850050482.
- Carter, Justin. “Voice Actors Are Having Their Voices Stolen by Ai.” Gizmodo, Gizmodo, 12 Feb. 2023, https://gizmodo.com/voice-actors-ai-voices-controversy-1850105561.
- Hendrickson, John. “The next Big Political Scandal Could Be Faked.” The Atlantic, Atlantic Media Company, 3 Mar. 2023, https://www.theatlantic.com/politics/archive/2023/03/politicians-ai-generated-voice-fake-clips/673270/.
- Lajka, Arijeta, and The Associated Press. “Artificial Intelligence Makes Voice Cloning Easy and ‘the Monster Is Already on the Loose’.” Fortune, Fortune, 11 Feb. 2023, https://fortune.com/2023/02/11/artificial-intelligence-makes-voice-cloning-easy-and-the-monster-is-already-on-the-loose/.