Deep Background Presents: Brave New Planet - podcast episode cover

Deep Background Presents: Brave New Planet

Oct 14, 20201 hr 14 min
--:--
--:--
Listen in podcast apps:
Metacast
Spotify
Youtube
RSS
Download Metacast podcast app
Listen to this episode in Metacast mobile app
Don't just listen to podcasts. Learn from them with transcripts, summaries, and chapters for every episode. Skim, search, and bookmark insights. Learn more

Episode description

Introducing Brave New Planet, a seven-part series that delves deep into powerful technologies changing our world. They have amazing potential upsides, but we can’t ignore the serious risks. Hosted by Dr. Eric Lander, Brave New Planet is a partnership between the Broad Institute, Pushkin Industries, and the Boston Globe.


In this episode, we're talking about how it's becoming more common to make convincing -- but false -- videos through artificial intelligence. These “deepfakes” can be useful in art, education, and therapy. But they can also be used to harm the reputation of an ex-partner or a stranger. Could they be weaponized to provoke international conflicts or swing elections? When seeing is not believing, who can we trust, and can democracy and truth survive?

Learn more about your ad-choices at https://www.iheartpodcastnetwork.com

See omnystudio.com/listener for privacy information.

Transcript

Speaker 1

Noah Feldman here. I'm excited to tell you about a special five part Deep Background mini series called Deep Bench. The first episodes will appear in your feed on Saturday, October seventeenth. The battle for the Supreme Court has become a huge issue in the presidential election. In many ways, it's a culmination of a conservative legal revolution spearheaded by

the Federalist Society. Deep Bench is the inside story of how these legal conservatives gained power and how, at the height of their influence, they're actually in danger of splitting apart. But first, we're presenting an episode from Pushkin Industry's newest show, Brave New Planet. Every day we see how powerful technologies are advancing at a breathtaking pace. They have amazing potential upsides, but if we're not careful, some might leave us a

lot worse off. In Brave New Planet, doctor Eric Lander and his guests weigh the pros and cons of a wide range of powerful innovations in science and technology. Doctor Lander directs the Broad Institute of MIT and Harvard. He was a leader of the Human Genome Project for eight years. He served as a science advisor to President Obama's White House.

In this episode, doctor Lander explores deep fakes. Deep fakes can be useful in art, education and therapy, but could they be weaponized to provoke international conflicts or swing elections? And where does the right to free speech fit in? Every episode of Brave New Planet will grapple with opportunities and challenges that are too big to fit in a tweet, but will shape our future. You can subscribe in Apple Podcasts.

Here's the brilliant Eric Lander and Brave New Planet. Your to Brave New Planet, a podcast about amazing new technologies that could dramatically improve our world, or, if we don't make wise choices, could leave us a lot worse off Utopia or dystopia. It's up to us. On July sixteenth, nineteen sixty nine, Apollo eleven blasted off from the Kennedy

Space Center near Cape Canaveral, Florida. Twenty five million Americans watched on television as the spacecraft ascended toward the heavens, carrying Commander Neil Armstrong, Lunar Module pilot Buzz Aldron, and Command Module pilot Michael Collins their mission to be the first humans in history to set foot on the Moon. Four days later, on Sunday, July twentieth, The lunar module separated from the command ship and soon fired its rockets

to begin its lunar descent. Five minutes later, disaster struck about a mile above the Moon's surface. Program alarms twelve oh one and twelve O two sounded loudly, indicating that the mission computer was overloaded, and then well, every American knows what happened next lost date of five good evening, my fellow Americans. President Richard Nixon addressed a grieving nation. Fates has ordained that the men who went to the Moon to explore in peace will stay on the Moon

to rest in peace. These men, Neil Armstrong and Edwin Auburn, know that there is no pope for their recovery, but they also know that there is hook for mankind in their sacrifice. He ended with the now famous words for every human being who looks up at the Moon and the nights to come, will know that there is some pun or another word that is forever mankind. Wait a minute, that never happened. The Moon mission was a historic success. The three astronauts returned safely to ticker tape parades and

a celebratory thirty eight day World Tour. Those alarms actually did sound, but they turned out to be harmless. Nixon never delivered that speech. His speech writer had written it, but it sat in a folder labeled an event of Moon disaster until now. The Nixon you just heard is a deep fake, part of a seven minute film created

by artificial intelligence deep learning algorithms. The fake was made by the Center for Advanced Virtuality at the Massachusetts Institute of Technology as part of an art exhibit to raise awareness about the power of synthesized media. Not long ago, something like this would have taken a lot of time

and money, But now it's getting easy. You can make new paintings in the style of French Impressionism, revived dead movie stars, help patience with ner degenerative disease, or soon maybe take a class on a tour of ancient Rome. But as the technology quickly becomes democratized, we're getting to the point where almost anyone can create a fake video of a friend, an ex lover, a stranger, or a public figure that's embarrassing, pornographic, or perhaps capable of causing

international chaos. Some argue that in a culture where faked news spreads like wildfire, and political leaders deny the veracity of hard facts. Deep faked media may do a lot more harm than good. Today's big question will synthesized media unleash a new wave of creativity or will it erode the already tenuous role of truth in our democracy? And is there anything we can do to keep it in check. My name is Eric Lander. I'm a scientist who works

on ways to improve human health. I helped lead the Human Genome Project, and today I lead the Broad Institute of MIT and Harvard. In the twenty first century, powerful technologies have been peering at a breathtaking pace related to the Internet, artificial intelligence, genetic engineering, and more. They have amazing potential upsides, but we can't ignore the risks that come with them. The decisions aren't just up to scientists

or politicians. Whether we like it or not, we all of us are the stewards of a brave New planet. This generation's choices will shape the future as never before. Coming up on today's episode of Brave New Planet, I speak with some of the leaders behind advances in synthesized media. You could, certainly, by the way, generate stories that could be fresh and interesting and new and personal for every child. We got emails from people who were quadruplegic and they

asked us if we could make them dance. We hear from experts about some of the frightening ways that bad actors can use deep fakes. Creditors would chime in and say, you can absolutely make a deep fake sex video of your ex with thirty pictures. I've done it with twenty. Here's the things that keep me up at night. Right a video of Donald Trump saying I've launched nuclear weapons

against Iran. And before anybody gets around to figuring out whether this is real or not, we have global nuclear mountdown and we explore how we might prevent the worst abuses. It's important that younger people advocate for the Internet that they want. We have to fight for it. We have to ask for different things. Stay with us, Chapter one, Abraham Lincoln's Head. To begin to understand the significance of deep fake technology, I went to San Francisco to speak

with a world expert on synthetic media. My name is alexey Or sometimes called alyosha Eros, and I'm a professor at UC Berkeley and Computer Science and Electrical Engineering Department. My research is on computer vision, computer graphics, machine learning, various aspects of artificial intelligence. Where'd you grow up? I grew up in Saint Petersburg in Russia. I was one of those geeky kids playing around with computers or dreaming

about computers. My first computer was actually the first Soviet personal computer. So you actually are involved in making sort of synthetic content, synthetic media. That's right. Alexei has invented powerful artificial intelligence tools, but his lab also has a wonderful ability to use computers to enhance the human experience. I was struck by a remarkable video on YouTube created

by his team at Berkeley. So this was a project that actually was done by my students who didn't even think of this as anything but a silly little toy project of trying to see if we could get a geeky computer science student to move like a ballerina. In the video, one of the students, Carolyn cham dances with a skill and grace of a professional, despite never having studied ballet. The idea is you take a source actor

like a ballerina. There is a way to detect the limbs of the dancer have a kind of a skeleton extracted and also have my student just move around and do some geeky moves. And now we're basically just going to try to sympathize the appearance of my student driven by the skeleton of the ballerina. Put it all together, and then we have our grad student dancing pirouets like a ballerina through artificial intelligence. Carolyn's body is puppeteered by

the dancer. We weren't even going to publish it, but we just released a video on YouTube called Everybody Dance Now, and somehow it really touched the nerve. Well, there's been an explosion recently a new ways to manipulate media. Alexei notes that the idea itself isn't new, it has a long history. I can't help but ask, given that you come from Russia. One of the premier users of doctoring photographs I think was Stalin, who used the ability to

manipulate images for political effect. How did they do that? Can you think of examples of this and like, what was the technology? Then? The urge to change photographs has been around basically since the invention of photography. For example, there is a photograph of Abraham Lincoln that still hangs in may classrooms. That's fake. It's actually Calhoun with Lincoln's head attached it. Alexei's referring to John C. Calhoun, the

South Carolina senator and champion of slavery. A Civil War portrait artist superimposed a photo of Lincoln's head onto an engraving of Calhoun's body because they thought Lincoln's gangly frame wasn't dignified enough, and so they just said, okay, we can use Calhoun. Let's slap the Lincoln's head on his body. And then, of course, as soon as you go into the twentieth century, as soon as you get to dictatorships, this is a wonderful toy for a dictator to use.

So again, Stalin was a big fan of this. He would get rid of people in photographs once they were out of favor or once they got jailed or killed. He would just basically get them scratched out with reasonably cruetics. Hitler did it, Mao did it, Castro did it, Bresnev did it. I'm sure US agencies have done it also. We have always manipulated images with a desire to change history. This is Honi for Reed. He's also a professor at

Berkeley and a friend of Alexei's. I'm a professor of computer science and I'm an expert in digital forensics, where Alexei works on making synthetic media. Honey has devoted his career to identifying when synthetic media is being used to fool people, that is spotting fakes. He regularly collaborates on this mission with Alexei. So I met Alyosha efros ten

twenty years ago. He is a really incredibly creative and clever guy, and he has done what I consider some of the most interesting work in computer vision and computer graphics over the last two decades. And if you really want to do forensics, well, you have to partner with somebody like Alyosha. You have to partner with world class mind who knows how to think about the synthesiside so that you can synthesize the absolute best content and then

think about how to detect it. I think it's interesting that if you're somebody on the synthesis side and developing the forensic there's a little bit of a jekylin hide there, and I think it's really fascinating. You know, the idea of altering photos, it's not entirely new How far back does this go? So we used to have in the days of Stalin, highly talented, highly skilled, time consuming, difficult process of manipulating images, removing somebody, erasing something from the image,

splicing faces together. And then we moved into the digital age, where now a highly talented digital artist could remove one face and add another face, but it was still a time consuming and required scale. In nineteen ninety four, the makers of the movie Forrest Gump won an Oscar for Visual Effects for their representations of the title character interacting with historical figures like President John F. Kennedy gratulate. How does it feel to be an all Americans? Very good gratulation?

How do you feel? I believe it. Now computers are doing all of the heavy lifting of what used to be relegated to talented artists. The average person now can use sophisticated technology to not just capture the recording, but

also manipulate it and then distribute it. The tools used to create synthetic media have grown by leaps and bounds, especially in the past few years, and so now we have technology broadly called deep fake, but more specifically should be called synthesized content where you point an image or a video or an audio to an AI or machine learning system, and it will replace the face for you. I mean it can do that in an image, it can do that in a video, or it can synthesize

audio for you in a particular person's voice. It's becomes straightforward to swap people's faces. There's a popular YouTube video that features tech pioneer Elon Musk's adult face on a baby's body, and there's a famous meme where actor Nicholas Cage's face replaces those of leading movie actors, both male and female. You can put words into people's mouths and make them jump and dance and run. You can even resurrect powerful figures and have them deliver a fake speech

about a fake tragedy. From an Altered History, Chapter two, Creating Nixon. The text of Nixon's Moon disaster speech that we heard at the top of the show is actually not fake. As I mentioned, it was written for President Nixon as a contingency speech and thankfully never had to be delivered. It's an amazing piece of writing. It was written by Bill Safire, who was one of Nixon's speech writers.

This is artist in journalist Francesco Panetta. She's the co director of the Nixon Fake or MTS Moon Disaster Team. She's also the creative director in MIT's Center for Advanced Virtuality. I was doing experimental journalism at the Guardian newspaper. I ran the Guardians Virtual Reality studio for the last three years. The second half of the Moon Disaster team is sound

artist Halsey Bergund. My name is Halsey Bergund. I'm a sound artist and technologist, and I've had a lot of experience with lots of sorts of audio enhanced with technology, though this is my first experience with synthetic media, especially since I typically focus on authenticity of voice and now I'm kind of doing the opposite. So together Halsey and Francesca chose to automate a tragic moment in history that

never actually happened. I think it all started with it being the fiftieth anniversary of the Moon landing last year, and add on top of that an election cycle in this country, and dealing with this information, which is obviously very important in election cycles. It was like light bulbs went on and we got very excited about pursuing it. It's possible to make mediocre fakes pretty quickly and cheaply,

but Francesca and Halsey wanted high production values. So how does one go about making a first rate fake presidential address. There are two components. There's the visuals and there's the audio, and they are completely different processes. So we decided to go with a video dialogue replacement company called Kenny Ai, who would do the visuals for us, and then we decided to go with Respeecher, who are a dialogue replacement company for the voice of Nixon. They tackled the voice first,

the more challenging of the two mediums. What we were told to do was to get two to three hours worth of Nixon talking. That was pretty easy because the Nixon Library has hours and hours of Nixon, mainly giving Vietnam's speeches. The Communist armies of North Vietnam launched a massive inversion of South Vietnam. That audio was then chopped up into chunks between one and three seconds long. We

found this incredibly patient actor called Lewis D. Wheeler. Lewis would listen to the one second clip and then he would repeat that and do what I believe was right. Speech. Would say to us things like we need to change the diagonal attention, which meant nothing to us. Yes, we have a whole lot of potential band names going forward. Yeah,

Synthetic Nixon is another good one. So once we have our Nixon model made out of these thousands of tiny clips, it means that whatever our act says will come out then in Nixon's voice. So then what we did was record the contingency speech of Nixon, and it meant that we got Lewis's actually performance, but in Nixon's voice. What about the video part, I mean, the video was much easier. We're talking a couple of days here and a tiny

amount of data just with Lewis's iPhone. We filmed him reading the contingency speech once a couple of minutes of him just chatting to camera, and that was it fate that the men who went to the Moon to explore will stay on the moon. You know, we were told by Kenny Ai that everything would be the same in the video apart from just the area around the mouth. So every gesture of the hand, every blink, every time he moved his face, all of that would stay the same,

but just the mouth basically would change. So we used Nixon's resignation speech. To have served in this office, it is to have felt a very personal sense of It was the speech of Nixon that looked to the most somber way. He seemed to have the most emotion in his face. So what actually went on in the computer? Artificial intelligence sometimes sounds inscrutable, but the basic ideas are quite simple. In this case, it uses a type of

computer program called an auto encoder. It's trained to take complicated things, say spoken sentences or pictures, encode them in a much simpler form, and then decode them to recover the original as best it can. The encoder tries to reduce things to their essence, throwing away most of the information but keeping enough to do a good job of reconstructing it to make a deep fake. Here's the trick.

Train a speech auto encode for Nixon to Nixon, and a speech auto encoder for actor to actor, but force them to use the same encoder. Then you can input actor and decoded as Nixon. If you have enough data. It's a piece of cake. Around their carefully created video, the Moon Disaster team created an entire art installation a nineteen sixties living room with a fake vintage newspaper sharing the fake tragic news while a fake Nixon speak solemnly

on a vintage black and white television. Some people, when they were watching the installation, they watched a number of times. You'd see them, they'd watch it once, then they would watch it again, staring at the lips to see if they could see any lack of synchronicity. We had some people who thought that perhaps Nixon had actually recorded this speech as a contingency speech for it to go onto television.

Lots of folks who were listening, viewing, and even press folks just immediately said, oh, the voice is real or whatever, you know, said these things that weren't accurate because they just felt like there wasn't even a question. I suppose that is what we wanted to achieve. But at the same time, it's it was a little bit eye opening and like a little scary, you know that that could happen. Chapter three, Everybody dance. What do you see as just

the wonderful upsides of having technologies like this? Yeah, I mean ai A art is becoming a whole field in itself, so creatively there is enormous potential. One of the potential positive educational uses of deep fake technology would be to bring historical figures back to life to make learning more durable. I think one could do that with bringing Abraham Lincoln back to life and having him deliver speeches. Film companies

and radios cited about re enactments. We're already beginning to see this in films like Star Wars, when we're bringing people like Carrie Fisher back to life. I mean that is at the moment not being done through deep fake technologies. This is using fairy traditional techniques of CGI at the moment. So we still have to see our first deep fake big cinema screen release, but this is just to come.

Like the technology is getting better and better. Not only will we be able to potentially bring back actors and actresses who are no longer alive and have them star movies, but an actor could make a model of their own voice and then sell the use of that voice to anybody to do a voiceover of whatever is wanted, and so they could have twenty of these going on at the same time, and the sort of restriction of their

physical presence is no longer there. And that might mean that you know, Brad Pitt is in everything, or it might just mean that lower budget films can afford to have some of the higher cost talent. That point, you know, the top twenty actors could just do everything. Yes, there's no doubt that there will be winners and losers from these technologies, but the potential of synthetic media goes way beyond the arts. There are possible medical and therapeutic applications.

There are companies that are working very hard to allow people who have either lost their voice or who never had a voice to be able to speak in a way that is either how they used to speak or in a way that isn't a canned voice that everybody has. Alexei ePROs and his students discovered potential uses of synthetic media and medicine quite unintentionally while working on their Everybody Danced Now project that could turn anyone into a ballerina.

We were kind of surprised for all the positive feedback. We god, we've got emails from people who were quadruplegic and they asked us if we could make them dance, and it was very unexpected. So now we are trying to get the soft where to be in a state where people can use that because yeah, it's somehow it did hit a nerve with folks. Chapter four Unicorns in the Andes. The past few years have seen amazing advances

in the creation of synthetic media through artificial intelligence. The technology now goes far beyond fitting one face over another face in a video. A recent breakthrough has made it possible to create entirely new and very convincing content out of thin air. The breakthrough, called generative adversarial networks or GAMS, came from a machine learning researcher at Google named Ian Goodfellow. Like auto encoders, the basic idea is simple but brilliant.

Suppose you want to create amazingly realistic photos of people who don't exist. While you build a GAM consisting of two computer programs, a photo generator that learns to generate fake photos and a photo discriminator that learns to discriminate or identify fake photos from a vast collection of real photos. You then let the two programs compete, continually tweaking their code to outsmart each other. By the time they're done,

the GAN can generate amazingly convincing fakes. You can see for yourself if you go to the website this person does not exist dot com. Every time you refresh the page, you're shown a new uncanny image of a person who, as the website says, does not and never did exist. Francescan I actually tried out the website. This young Asian woman. She's got great complexion, envious of that, neat black hair, with a fringe pink lipstick and a slightly dreamy look

as she's kind of gazing off to her left. Oh, here's a woman who looks like she could be a neighbor of mine in Cambridge, probably about sixty five. She's got nice wire framed glasses, layered hair. Her earrings don't actually match, but that could just be her distinctive style. I mean, of course, she doesn't really exist. It's hard to argue that gans aren't creating original art. In fact, an artist collective recently used a GAM to create a

French Impressionist style portrait. When Christie's sold it at auction, it fetched an eye popping four hundred and thirty two thousand dollars. Alexeiefros, the Berkeley professor, recently pushed gans a step further, creating something called cycle gans by connecting two gans together in a clever way. Cycle. Gans can transform a money painting into what's seemingly a photograph of the same scene, or turn a summer landscape into a winter

landscape of the same view. Alexei's psychogans seem like magic. If you were to add in virtual reality, the possibilities become mind blowing. You maybe reminiscing about walking down Saint German and Paris, and with a few clicks, you are there, and you're walking down the boulevard and you're looking at all the buildings, and maybe you can even switch to a different year. And I think that is I think very exciting as a way to mentally travel to different places.

So if you do this in VR, I mean imagine classes going on a class visit to ancient Rome. That's right. You could imagine from how a particular city like Chrome looks now. Trying to extrapolate we looked into past. It turns out that gans aren't just transforming images. I spoke with a friend who's very familiar with another remarkable application of the technology. My name is Reid Hoffman. I'm a

podcaster of Master's a Scale. I'm a partner at Greylock, which is where we're sitting right now co founder of LinkedIn and then a variety of other eccentric hobbies. Reid is a board member of an unusual organization called open ai. Open a Eyes is highly concerned with artificial general intelligence human level intelligence. I helped Sam Altman and Elon Musk standing up. The basic concern was that if one company created and deployed that that could be disbalancing in all

kinds of ways. And so the thought is, if it could be created, we should make sure that there is essentially a nonprofit that is creating this and that can make that technology available at selective time slices to industry as a whole government sut C. Last year, open ai released a program that uses gams to write language from a short opening prompt. The system, called GPT two, can spin a convincing article or story instead of a deep fake video. It's deep fake text. It's pretty amazing. Actually.

For example, open ai researchers gave the program the following prompt. In a shocking finding, scientists discovered a herd of unicorns living in a remote, previously unexplored valley in the Andes Mountains. Even more surprising to the researchers was the fact that the unicorns spoke perfect English. GPT two took it from there, delivering nine crisp paragraphs on the landmark discovery. I asked

Fran to read a bit from the story. Doctor Jorge Perez, an evolutionary biologist from the University of Lapaz, and several companions were exploring the these mountains when they found a small valley with no other animals or humans. Perez noticed that the valley had what appeared to be a natural fountains surrounded by two peaks of rock and silver snow. Perez and the others then ventured further into the valley.

By the time we reached the top of one peak, the water looked blue with some crystals on top, said Perez. Perez and his friends were astonished to see the unicorn herd. Don't tell me some of the great things you can do with language generation, well, say, for example, entertainment, Generate stories that could be fresh and interesting and new and

personal for every child. Embed educational things in those stories so they're drawn into the fact that the story is involving them and their friends, but also now brings in grammar and math and other kinds of things. As are doing it. Generate explanatory material of this kind of education that works best for this audience, for this kind of people, like we want to have this kind of math or this kind of physics, or this kind of history or

this kind of poetry explained in the right way. And also the sile language right like you know native city x language. When open ai announced its breakthrough program for text generation, it took the unusual step of not releasing the full powered version because it was worried about the possible consequences. Now, part of the open ai decision to say we're going to release a smaller model than the one we did is because we think that the deep

fake problem hasn't been solved. And by the way, some people complained about that because they said, well, you're slowing down our ability to do progress and so far. The answer to say, look, when these are released to the entire public, we cannot control the downside as well as the upsides. Downsides from art to therapy to virtual time travel, personalized stories and education, Synthetic media has amazing upsides. What could possibly go wrong? Chapter five? What could possibly go wrong?

The downsides are actually not hard to find. The ability to reshape reality brings extraordinary power, and people inevitably use power to control other people. It should be no surprise, therefore, that ninety six percent of fake videos posted online are non consensual pornography videos, almost always of women manipulated to depict sex acts that never actually occurred. I spoke with a professor who studies deep fakes, including digital attempts to

control women's bodies. I'm Danielle Citron, and I am a law professor at Boston University School of Law. I write about privacy, technology, automation. My newest work and my next book is going to be about sexual privacy. So I've worked in and around consumer privacy, individual rights, civil rights. I write a lot about free speech and then automated systems. When did you first become aware of deep fakes? Do you remember when this cross your rit? I did so.

There was a Reddit thread devoted to, you know, fake pornography movies of Galjado Emma Watson. But the reddit thread sort of spooled not just from celebrities but ordinary people, and so you had rereditors asking each other, how do I make a deep fake sex video of max girlfriend? I have thirty pictures, and then other reditors would chime in and say, look at this YouTube tutorial. You can absolutely make a deep fake sex video of your ex

with thirty pictures. I've done it with twenty. In November twenty seventeen, an anonymous redditor began posting synthesized porn videos under the pseudonym deep fakes, perhaps a nod to the deep learning technology you used to create them as well as the nineteen seventies porn film throat. The Internet quickly adopted the term deep fakes and broadened its meanings beyond pornography.

To create the videos, he used celebrity faces from Google image search and YouTube videos, and then trains an algorithm on that content together with pornographic videos. Have you seen deep fake pornography videos? Yes, so still pretty crude, so you probably can tell that it's a fake, But for

the person who's inserted into pornography, it's devastating. You use the neural network technology, the artificial intelligence technology to create out of digital whole cloth pornography videos using probably real pornography and then inserting the person in the pornography so they become the female actress. If it's a female, it's usually a female in that video. My name is Noel Martin and I am an activist and Lauraform campaigner in Australia.

Noel is twenty six years old and she lives in Perth, Australia. So the first time that I discovered myself on pornographic sites was when I was eighteen and out of curiosity, decided to Google image reverse search myself. In an instant, like in a less than a millisecond, my life completely changed. At first, it started with photos still images stolen from Noel's social media accounts. They were then doctoring my face from ordinary images and superimposing those onto the bodies of

women depicting me having sexual intercourse. It proved impossible to identify who was manipulating Noel's image in this way. It's still unclear today, which made it difficult for her to seek legal action. I went to the police soon after, I contacted government agencies, tried getting a private investigator. Essentially, there's nothing that they could do. The sites are hosted overseas, the perpetrators are probably overseas. The reaction was at the end of the day, I think you can contact the

webmasters to try and get things deleted. You know, you can adjust your privacy setting so that you know nothing is available to anyone publicly. It was an unwinnable situation. Then things started to escalate. In twenty eighteen, who all saw a synthesized pornographic video of herself And I believe that it was done for the purposes of silencing me because I've been very public about my story and advocating

for change. So I had actually gotten email from a fake email address, and you know, I clicked the link I was actually at work. It was a video of me having sexual intercourse. The title had my name, the face of the woman in it was edited so that it was my face, and you know, all the tags were like Noel Martin Australia, Feminist, and it didn't look real, but the context of everything, with the title my face, with the tags all points to me being depicted in

this video. The fakes were of poor quality, but poor and consumers aren't a discriminating lot, and many people reacted to them as if they were real. The public reaction was horrifying to me. I was a victim, blamed and slut shamed, and it's definitely limited the course of where I can go in terms of career and employment. Noel finished a degree in law and began a peening to

criminalize this sort of content. My advocacy and my activism started off because I had a lived experience of this, and I experienced it at a time where it wasn't criminalized in Australia, the distribution of altered intimate images or altered intimate videos, and so I had to petition, meet

with my politicians in my area. I wrote a number of articles, I spoke to the media, and I was involved in the law reform in Australia in a number of jurisdictions in Western Australia and New South Wales, and I ended up being involved in two press conferences with the Attorney generals of each state at the announcement of

the law that was criminalizing this abuse. Today, in part because of Noel's activism, it is illegal in Australia to distribute intimate images without and scent, including intimate images and videos that have been altered. Although it doesn't encompass all malicious synthetic media, Noel has made a solid start. Chapter six, Scissors and Glue. The videos depicting Noel Martin were nowhere near as sophisticated as those made by the Moon disaster team.

They were more cheap fakes than deep fakes, and yet the point didn't have to be perfect to be devastating. The same turns out to be true in politics. To understand the power of fakes, you have to understand human psychology. It turns out that people are pretty easy to fool. John Kerry I was running for president of the US. His stance on the Vietnam War was controversial. Jane Fond, of course, was a very controversial figure back then because

of her anti war stand. What have we become as a nation if we call the men heroes that were used by the had have gone to try to exterminate an entire people? What business have we to try to exterminate a people? And somebody had created a photo of the two of them sharing a stage and an anti war rally with the hopes of damaging the Carry campaign. The photo was fake. They had never shared a stage together.

They just took two images, probably put it into some standard photo editing software like a photoshop, and just put a headline around it, and out to the world it went. And I will tell you I remember the most fascinating interview I've heard in a long time was right after the election, Carry, of course lost, and a voter was being interviewed and asked how they voted, and he said he couldn't vote for Carry, and the interview said, well

why not? And the gentleman said, I couldn't get that photo of John Carry and Jane Fonda out of my head. And the interview said, well, you know that photo is fake, and the guy said, much to my surprise, yes, but I couldn't get it out of my mind. And this is shows you the power of visual imagery, Like even after I tell you something is fake, it still had an impact on somebody. And I thought, wow, we're in a lot of trouble because it's very very hard to

put the cat back into the bag. Once that content is out there, you can't undo it. So seeing is believing, even above thinking, Yeah, that seems to be the rule. There is very good evidence from the social science literature that it's very very difficult to correct the record after the mistakes are out there. Law professor Danielle Citram also notes that humans tend to pass on information without thinking,

which triggers what she calls information cascades. Information cascades is a phenomenon where we have so much information overload that when someone sends us something, some information, and we trust that person, we pass it on. We don't even check its veracity, and so information can go viral fairly quickly because we're not terribly reflective, because we act on impulse. Danielle says that information cascades have been given new life

in the twenty first century through social media. Think about the twentieth century phenomenon where we get most of our information from trusted sources, trusted newspapers, trusted major couple of TV channels. Growing up, we only had to you know, we didn't have a million, and they were adhering to journalistic ethics and commitments to truth and neutrality and notion that you can't publish something without checking it. Now we are publishing information that most people say. We're lying on

our peers and our friends. Social media platforms are designed to tailor our information diet to what we want and to our pre existing views, so we're locked in a digital echo chamber. We think everybody agrees with us. We pass on that information we haven't checked the veracity, it goes wild. And we're especially likely to pass it on if it's negative and novel. Why's that, It's just like it's one of our weaknesses. We know how gossip goes

like wildfire online. So like Hillary Clinton as running a sex ring. That's crazy. Oh my god, Eric, did you hear about that? I'll post it on Facebook. Eric, you pass it on. We just can't help ourselves. And it is much in the way that we love suits and fats and pizza. You know, we indulge. We don't think not some sense. This phenomenon is an old phenomenon, right, there's the famous observation by Mark Twain about how a lie gets halfway around the world before the truth gets

its pants hang. Yeah, the truth is still in the bedroom getting dressed, and we often will see the lie, but the rebuttal is not seen. It's often lost in the noise of the defamatory statements. That is not new. But what is new is a number of things about our information ecosystem are our force multipliers. Chapter seven, Truth decay. Many experts are worried that the rapid advances in making fakes, combined with a catalyst of information cascades, will undermine democracy.

The biggest concerns have focused on elections. Globally, we are looking at highly polarized situations where this kind of manipulated media can be used as a weapon. One of the main reasons Francesca and Halsey made their Nixon deep fake was to spread awareness about the risks of misinformation campaigns

before the twenty twenty US presidential election. Similarly, a group showcased the power of deep fakes by making videos in the run up to the UK parliamentary election showing the two bitter rivals Boris Johnson and Jeremy Corman, each endorsing the other. I wish to rise above this divide and indorse my worthy opponent, the right Honorable Jeremy Corbyn, to be Prime Minister of our United Kingdom back Boris Johnson to continue as our Prime Minister. But you know what,

don't listen to me. I think I may be one of the thousands of deep fakes on the Internet, using powerful technologies to tell stories that aren't so. This just kind of indicates how candidates and political figures can be misrepresented, and you just need to feed them into people's social media feeds for them to be seeing this. At times when the stakes are pretty high. So far, we haven't yet seen sophisticated deep fakes in US or UK politics.

That might be because fakes will be most effective if they're time for maximum chaos, say close to election day, when newsrooms won't have the time to investigate and debunk them. But another reason might be that well cheap fakes made with basic video editing software or actually pretty effective. Remember the video that surfaced of how Speaker Nancy Pelosi, in which she appeared intoxicated and confused. We want to give this president the opportunity do something historic for our country.

Both President Trump and Rudy Giuliani shared the video as fact on Twitter. The video is just a cheap fake, just slowed down Pelosi's speech to make her seem incompetent. But maybe elections won't be the biggest targets. Some people worry the deep fakes could be weaponized to foment international conflict. Berkeley professor Honey f Reed has been working with US

government's Media Forensics program to address this issue. DARPA, on the Defense Department's Research arm, has been pouring a lot of money over the last five years into this program. They are very concerned about how this technology can be a threat to national security and also how when we get images and videos from around the world in areas of conflict, do we know if they're real or not? Is this really an image of a US soldier who

has been taken hostage? How do we know? So? What do you see as some of the worst case scenarios. Here's the things that keep me up at night right a video of Donald Trump saying I've launched nuclear weapons against Iran, and before anybody gets around to figuring out whether this is real or not, where we have global nuclear mountdown. And here's the thing. I don't think that that's likely, but I also don't think that the probability of that is zero. And that should worry us because

while it's not likely, the consequences are spectacularly bad. Lawyer Danielle Citrom worries about an even more plausible scenario. Imagine a deep fake of a well known American general burning a Koran, and it is timed at a very tense moment in a particular most country, whether it's agha unerstand. It could then lead to physical violence. And you think this could be made. No general, no Quran actually used in the video just programmed. You can use the technology

to mine existing photographs. Kind of easy, especially with someone could take Jim Mattis when he was our defense secretary. Of Jim Mattis, you know, actually taking a Koran and ripping it in half and say all Muslims should die. Imagine the chaos in diplomacy, the chaos of our soldiers abroad in Muslim countries. It would be inciting violence without question. Well, we haven't yet seen spectacular fake videos used to disrupt

elections or create international chaos. We have seen increasingly sophisticated attacks on public policymaking. So we've got an example in twenty seventeen where the FCC solicited public comment on the proposal to repeal net neutrality. Net neutrality is the instable that Internet service providers should be a neutral public utility. They shouldn't discriminate between websites, say slowing down Netflix streaming to encourage you to purchase a different online video service.

As President Barack Obama described in twenty fourteen, there are no gatekeepers deciding which sites you get to access. There are no toll roads on the information super Highway. Federal communications policy had long supported net neutrality, but in twenty seventeen, the Trump administration favored repealing the policy. There were twenty two million comments that the e SEC received, but ninety

six percent of those were actually fake. The interesting thing is the real comments were opposed to repeal, whereas the fake comments were in favor. A Wall Street Journal investigation exposed that the fake public comments were generated by bots. It found similar problems with public comments about payday lending. The bots varied their comments in a combinatorial fashion so that the content wasn't identical. With a little sleuthing, though,

you could see that they were generated by computers. But with a technology increasingly able to generate completely original writing, like open ayes program that wrote the story about unicorns in the ANDES, it's going to become hard to spot the fakes. So there was this Harvest student, Max Weiss, who used GPT two to kind of demonstrate this, And I went on his site yesterday and he's got this little test where you need to decide whether a comment

is real or fake. So you go on and you read it, and you decide whether it's been written by a bot or by a human. So I did this, and the ones that seemed to be really well written and quite narrative and discussive. Generally, I was picking them as human. I was wrong almost all the time. It was amazing and a lot. In our democracy, public commons have been an important way in which citizens can make their voices heard. But now it's becoming easy to drown

out those voices with millions of fake opinions. Now the downfall of truth likely won't come with a bang, but a whimper, a slow, steady erosion that some call truth decay. If you can't believe anything you read, or hear or see anymore, I don't know how you have a democracy, an I don't know, frankly, how we have civilized society if everybody's going to live in an echo chamber, believing their own version of events. How do we have a

dialogue if we can't agree on basic facts. In the end, the most insidious impact of deep fakes may not be the deep fake content itself, but the ability to claim that real content is fake. It's something that Danielle Citron refers to as the liars dividend. The liars dividend is that the more you educate people about the phenomenon of deep fix, the more the wrongdoer can disclaim reality. Think about what President Trump did with the Access Hollywood tape.

You know, I'm automatically attracted to be in the work. I just started kissing them. It's like a magnet. You just I don't even know. And when you were a start, they let you do it. You can do anything whatever you want, grabbed by the I can do anything. Initially, Trump apologized for the remarks. Anyone who knows me knows these words don't reflect who I am. I said it.

I was wrong, and I apologize. But in twenty seventeen, a year after his initial apology and with the idea of deep fake content starting to gain attention, Trump changed his tune. Upon reflection, he said, they're not real. That wasn't me. I don't think that was my voice. That's the liar's dividend. In practice, the Trump commented about Access Hollywood was remarkable. Slightly more subtle than that, He said, I'm not sure that was me. Right. Well, that's the

corrosive gas lighting. Chapter eight, A Life Stored in the Cloud. Deep fakes have the potential to devastate individuals and harms society. The question is can we stop them from spreading before they get out of control. To do so, we need reliable ways to spot deep fakes. So the good news is there are still artifacts in the synthesized content, whether those are images, audio, or a video, that we as

the experts, can tell apart. So when for example, the New York Times wants to run a story with a video, we can help them validate it. What are the real sophisticated experts looking gat So the eyes are really wonderful forensically because they reflect back to you what is in the scene. I'm sitting now right now in a studio. There's maybe about a dozen or so lights around me, and you can see this very complex set of reflections

in my eyes. So we can analyze fairly complex lighting patterns, for example, to determine if this is one person's head spliced onto another person's body, or if the two people standing next to each other were digitally inserted from another photograph. I could spend another hour telling you about the many different forensic techniques that we've developed. There's no silver bullet here.

Really is a sort of a time consuming and deliberate and thoughtful and it requires many many tools, and it requires people with a fair amount of skill to do this. Honey Freed also has quite a few detection techniques that he won't speak about publicly for fear of the deep fake creators will learn how to beat his tests. I don't create a GitHub repository and give my code to all my adversaries. I don't have just one forensic techniques.

I have a couple dozen of them. So that means you, as the person creating this now have to go back and implement twenty different techniques. You have to do it just perfectly, and that makes the landscape a little bit more tricky for you to manage. As technology makes it easier to create deep fakes, a big problem will be

the sheer amounts of content to review. So the average person can download software repositories, and so it's getting to the point now where the average person can just run these as if they're running any standard piece of software. There's also websites that have propped up where you can pay them twenty bucks and you tell them, please put this person's face into this person's video, and they will do that for you. And so it doesn't take a

lot to get access to these tools. Now, I will say that the output of those are not quite as good as what we can create inside the lab. And you just know what the trend is. You just know it's going to get better and cheaper and faster and easier to use. Detecting few fakes will be a never ending cat and mouse game. Remember how generative adversarial networks or gams are built by training a fake generator to outsmart a detector. Well, as detectors get better, fake generators

will be trained to keep pays still. Detectives like Honey and platforms like Facebook are working to develop automated ways to spot deep fakes rapidly and reliably. That's important because more than five hundred additional hours of video are being uploaded to YouTube every minute. I don't mean to sound defeatist about this, but I'm going to lose this war. I know this because it's always going to be easier to create content than it is to detect it. But here's where I will win. I will take it out

of the hands of the average person. So think about, for example, the creation of counterfeit currency. With the latest innovations brought on by the Treasure Department, it is hard for the average person to take their inkjet printer and create compelling fake currency, and I think that's going to be the same trend here is that if you're using some off the shelf tool, if you're paying somebody on the website, we're going to find you, and we're going

to find you quickly. But if you are a deadated highly skilled of the time and the effort to create it, we are going to have to work really hard to detect those. Given the challenges of detecting fake content, some people envision a different kind of techno fix. They proposed developing airtight ways for content creators to mark their own original video as real. That way, we could instantly recognize

an altered version if it wasn't identical. Now, there's ways of authenticating at the point of recording, and these are what it called control capture system. So here's the idea. You use a special app on your mobile device that at the point of capture a cryptographically signs the image of the video or the audio. It puts that signature

onto the blockchain. And the only thing you have to know about the blockchain is that it is an immutable distributed ledger, which means that that signature is essentially impossible to manipulate, and now all of that happened at the

point of recording. If I was running a campaign today and I was worried about candidates likeness being misused, absolutely every public event that they were at, I would record with a control capture system, and I'd be able to prove what they actually said or did at any point in the future. So this approach would shift the burden of authentication to the people creating the videos rather than publishers or consumers. Law professor Danielle Citron has explored how

this solution could quickly become dystopium. We might see the emergence of an essentially an audit trail of everything you do and say all of the time. Danielle refers to the business model as immutable lifelogs in the cloud. In a way we sort of already seen it. There are health plans that if you wear a fitbit all the time and you let yourself be monitored, it lowers your

insurance your health insurance rates. But you can see how if the incentives are there in the market to self surveil, whether it's for health insurance, life insurance, car insurance, we're going to see the unraveling of Brian to say by ourselves.

You know, corporations may very well, because the CEO is so valuable, they may say, you've got to have a log, an immutable audit trail of everything you do and say so when that deep fake comes up the night before the IPO, you can say, look, the CEO wasn't taking the bribe, wasn't having sex with a prostitute. And so we have proof, we have an auto trail, we have

a log. So when we were imagining, we were imagining a business model that hasn't quite come up, but we have gotten a number of requests from insurance companies as well as companies to say we're interested in this idea. So how much has to be in that log? Does this have to be a whole video of your life? That is a great question, one that terrifies us. So it may be that you're logging locate geolocation, you're logging videos.

You see people talking and who they're interacting with, and that might be good enough to prevent the miss Jeff that would hijack the IPM. Your whole life online, yes, stored securely, clock down, protected in the cloud. It is at least for a privacy scholar. There are so many reasons why we ought to have privacy that aren't about hiding things. It's about creating spaces and managing boundaries around

ourselves and our intimates and our loved ones. So I worry that if we entirely unravel privacy a in the wrong hands is very dangerous right B It changes how we think about ourselves and humanity. Chapter nine, Section two thirty. So techno fixes are complicated. What about passing laws to band deep fakes or at least deep fakes that don't disclose their fake So the video and audio is speech?

In our First Amendment doctrine is very much a protective of free speech, and the Supreme Court has explained that lies just lies themselves without harm is protected speech. When lies cause certain kinds of harm, we can regulate it defamation of private people, threats, incitement, fraud, impersonation of government officials.

What about lies concerning public figures like politicians? California and Texas, for instance, recently pass laws making it illegal to publish deep fakes of a candidate in the weeks leading up to an election. It's not clear yet whether the laws will pass constitutional muster. As you're saying in an American content, we are just not going to be able to outlaw fakes. Yeah, we can't have a flat van, and I don't think we should. It would fail on doctrinal grounds, but ultimately

it would prevent the positive uses. Interestingly, in January twenty twenty, China, which has no First Amendment protecting free speech, promulgated regulations banning deep fakes. The use of AI or virtuality now needs to be clearly marked in a prominent manner, and the failure to do so is considered a criminal offense. To explore other options for the US, I went to

speak with a public policy expert. My name is Joan Donovan, and I work at Harvard Kennedy Shorenstein Center, where I lead a team of researchers looking at medium manipulation and disinformation campaigns. Joan is head of the Technology and Social Change Research Project, and her staff studies how social media gives rise to hoaxes and scams. Her team is particularly

interested and precisely how misinformation spreads across the Internet. Ultimately, underneath all of this is the distribution mechanism, which is social media and platforms and platforms have to rethink the open of their design because that has now become a territory for information warfare. In early twenty twenty, Facebook announced

a major policy change about synthesized content. Facebook preissued policies now on deep fakes, saying that if it is an AI generated video and it's misleading in some other contextual way, then they will remove it. Interestingly, Facebook banned the Moon Disaster Team's Nixon video even though it was made for educational purposes, but didn't remove the slowed down version of Nancy Pelosi, which was made to mislead the public. Why

because the Pelosi video wasn't created with artificial intelligence. For now, Facebook is choosing to target deep fakes, but not cheap fakes. One way to push platforms to take a stronger stance might be to remove some of the legal protections that they currently enjoy. Undersection two thirty of the Communication's Decency Act past in nineteen ninety six, platforms aren't legally liable

for content posted by its users. The fact that platforms have no responsibility for the content they host has an upside. It's led to the massive diversity of online content we enjoyed today. But it also allows a dangerous escalation of fake news. Is it time to change section to thirty to create incentives for platforms to police false content? I asked the former head of a major platform, LinkedIn co

founder Reid Hoffman. For example, let's take my view of what the response to the christ Church shooting should be is to say, well, we want you to solve not having terrorism, murderer or murderers displayed to people. So we're simply going to do a fine of ten thousand dollars per view. Two shootings occurred at mosques in Christchurch, New Zealand in March twenty nineteen. Graphic videos of the event were soon posted online. Five people saw it, that's fifty

thousand dollars. But if he becomes a meme and a million people see it, that's ten billion dollars. Yes, right, So what's really trying to do is get you to say, let's make sure that the meme never happens. Okay, So that's a governance mechanism there is you find the channel the platform based on number of views would be a very general way to say, now you guys have to solve. Now you solve, you figure it out. What about other solutions.

If we are to make regulation, it should be about the amount of staff in proportion to the amount of users so that they can get a handle on the content. But can they be fast enough. Maybe the viral spread should be slowed down enough to allow them to moderate. Let's put it this way. The stock market has certain governor is built in when there's massive changes in a stock price. There are decelerators that kick in, breaks that kick in. Should the platforms have breaks that kick in

before something can go fully viral? So in terms of deceleration, there are things that they do already that accelerate the process that they need to think differently about, especially when it comes to something turning into a trending topic. So there needs to be an intervening moment before things get to the homepage and get to trending, where there is a content review. So much to say here, but I want to think particularly about listeners who are in their

twenties and thirties and are very tech savvy. They're going to be part of the solution here. What would you say to them about what they can do? I think it's important that younger people advocate for the Internet that

they want we have to fight for it. We have to ask for different things, and that kind of agitation can come in the form of posting on the platform, writing letters, joining groups like Fight for the Future, and trying to work on getting platforms to do better and to advocate for the kind of content that you want to see more of. The important thing is that our society is shaped by these platforms, and so we're not going to do away with them, but we don't have

to make do with them either. Conclusion, choose your planet. So there you have it, Stewards of the Brave New Planet. Synthetic media or deep fakes, have been manipulating content for more than a hundred years, but recent advances in AI have taken it to a whole new level of verisimilitude.

The technology could transform movies and television, favored actors from years past starring in new narratives, along with actors who never existed, patients regaining the ability to speak in their own voices, personalized stories created on demand for any child around the globe, matching their interests, written in their dialect, representing their communities. But there's also great potential for harm.

The ability to cast anyone in a pornographic video, weaponized media dropping days before an election, or provoking international conflicts. Are we going to be able to tell fact from fiction? Will truth survive? And what does it mean for our democracy? Better? Fake detection may help, but it'll be hard for it to keep up, and logging our lives in blockchain to

protect against misrepresentation doesn't sound like an attractive idea. Outright bands on deep fakes are being tried in some countries, but they're tricky in the US given our constitutional protections for free speech. Maybe the best solution is to put the liability on platforms like Facebook and YouTube. If we can joan Donovan's right to get the future you want, you're going to have to fight for it. You don't have to be an expert, and you don't have to

do it alone. When enough people get engaged, we make wise choices. Deep fakes are a problem that everyone can engage with. Brainstorm with your friends about what should be done. Use social media. Tweet at your elected representatives to ask if they're working on laws, like in California and Texas, And if you work for a tech company, ask yourself and your colleagues if you're doing enough. You can find lots of resources and ideas at our website Brave New

Planet dot org. It's time to choose our planet. The future is up to us. Brave New Planet is a co production of the Broad Institute of MT and Harvard Pushkin Industries in the Boston Globe, with support from the Alfred P. Sloan Foundation. Our show is produced by Rebecca Lee Douglas with Mary Doo theme song composed by Ned Porter, mastering and sound designed by James Garver, fact checking by

Joseph Fridman, and a Stitt and Enchant. Special Thanks to Christine Heenan and Rachel Roberts at Clarendon Communications, to Lee McGuire, Kristen Zarelli and Justine Levin Allerhand at the Broad, to Milobelle and Heather Faine at Pushkin, and to Eli and Edy Brode, who made the Brode Institute possible. This is Brave New Planet. I am her Aclander

Transcript source: Provided by creator in RSS feed: download file
For the best experience, listen in Metacast app for iOS or Android
Open in Metacast