Good Robot #2: Everything is not awesome

53m
When a robot does bad things, who is responsible? A group of technologists sounds the alarm about the ways AI is already harming us today. Are their concerns being taken seriously?
This is the second episode of our new four-part series about the stories shaping the future of AI.
Good Robot was made in partnership with Vox’s Future Perfect team. Episodes will be released on Wednesdays and Saturdays over the next two weeks.
For show transcripts, go to vox.com/unxtranscripts
For more, go to vox.com/unexplainable
And please email us! unexplainable@vox.com
We read every email.
Support Unexplainable by becoming a Vox Member today: vox.com/members
Learn more about your ad choices. Visit podcastchoices.com/adchoices

Listen and follow along

Transcript

I need a job with a steady paycheck.

I need a job that offers health care on day one for me and my kids.

I want a job where I can get certified in technical roles, like robotics or software engineering.

In communities across the country, hourly Amazon employees earn an average of over $23 an hour with opportunities to grow their skills and their paycheck by enrolling in free skills training programs and apprenticeships.

Learn more at aboutamazon.com.

We all have moments where we could have done better, like cutting your own hair,

yikes, or forgetting sunscreen, so now you look like a tomato.

Ouch.

Could have done better.

Same goes for where you invest.

Level up and invest smarter with Schwab.

Get market insights, education, and human help when you need it.

Learn more at schwab.com.

It's Unexplainable.

I'm Noam Hasenfeld, and this is the second part of our newest four-part series, Good Robot.

If you haven't listened to episode one, let me just stop you right here.

Go back in your feed, check out the first one.

We'll be waiting right here when you get back.

Once you're all ready and caught up, here is episode two of Good Robot from Julia Longoria.

You have cat hair on your nose, by the way.

I've been like trying not to pay attention to it, but I think you got it off.

Yeah,

sorry.

Cool.

So, should we get into it?

Sure, yeah.

Let me.

It helps me to kind of remember everything I'm going to say if I can sort of jot down thoughts as I go.

Do you have enough paper?

I think I don't have paper on it.

All right, I'll do it on my.

I saw that they had ramped.

This past fall, I traveled paperless to a library just outside Seattle to meet with this woman.

I feel like the library should have paper.

know I that English.

Her name is Dr.

Margaret Mitchell.

Found a brochure on making a robot puppet.

What is it?

What is the.

I don't know.

It looks like it's an event.

Build a robot puppet using a variety of materials with puppeteer.

I'm so into that.

Aw, it's too bad that it's only for ages six to twelve.

While she is over the age limit to make a robot puppet with the children in the public library, Dr.

Mitchell is a bit of a robot puppeteer in her own right.

What's your AI researcher origin story?

Like, how did you get into all of this?

What drew you here?

Yeah,

what inspired me to...

So, I mean, I guess I can, it's sort of like, do you want the long version or the short version?

Dr.

Mitchell is an AI research scientist, and she was one of the first people working on language models.

Well before ChatGPT, and, well, all the GPTs, she's an OG in the field.

So I'll tell you, like, I'll tell you a story, if that's okay.

Yeah.

Okay, so I was at Microsoft and I was working on the ability of a system to tell a story given a sequence of images.

So given five images.

This was about 2013.

She was working on a brand new technology at the time, what AI researchers called vision to language.

So, you know, translating images into descriptions.

She would spend her days showing image after image image to an AI system.

To me, it sounded kind of like a parent showing picture flashcards to a toddler learning to speak.

She says it's not anything like that.

She showed the model images of events like a wedding, a soccer match, and on the more grim side.

I gave the system a series of images about a big blast that left 30 people wounded called the Hempstead blast.

It was at a factory, and you could see from the sequence of images that the person taking the photo had like a third-story view, sort of overlooking the explosion.

So it was a series of pictures showing that there was this terrible explosion happening, and whoever was taking the photo was very close to the scene.

So I put these images through my system, and the system says,

Wow,

this is a great view.

This is awesome.

And I was like, oh crap, that is the wrong response to this.

So it sees this horrible, perhaps mortally wounding explosion and decides it's awesome.

Kind of like a parent watching their precious toddler say something kind of creepy, Mitchell watched in horror and with a deep fascination about where she went wrong, as the AI system that she had trained called images awesome again and again.

It said it quite a lot, so we called it the everything is awesome problem, actually.

Her robot was having these kinds of translation errors.

Errors that to the uninitiated made it seem like the AI system might want to kill people or at least gleefully observe their destruction and call it awesome.

What would the consequences of that be if that system was deployed out into the world, reveling in human destruction?

It's like, if this system were connected to a bunch of missile systems, then it's, you know, it's just a jump and skip away to just launch missile systems in the pursuit of the aesthetic of beauty, right?

Years before the AI boom we're living, when neural networks and deep learning were just beginning to show promise, researchers like Dr.

Mitchell and others were experiencing these uncanny moments where the AIs they were training seemed to do something seriously wrong.

Doing scary things their creators did not intend for them to do and were seemingly threatening to humanity.

So I was like one of the first people doing these systems where you could scan the world and have descriptions of it.

I was like on the forefront.

I was one of the first people making these systems go.

And I realized like if anyone is going to be paying attention to it right now,

it has to be me.

I had heard the fears of rationalists, also pioneers in thinking about AI,

that we might build a super intelligent AI that could go rogue and destroy humanity.

At first glance, It seemed like Dr.

Mitchell might be building one such robot.

But when Dr.

Mitchell Mitchell investigated the question of why the good robot she sought to build seemed to turn bad, the answer would not lead her to believe what the rationalists did: that a super intelligent AI could someday deceive or destroy humanity.

To Dr.

Mitchell,

the answer was looking at her in a mirror.

This is episode two of Good Robot, a series about AI from Unexplainable in collaboration with Future Perfect.

I'm Julia Longoria.

Support for Unexplainable comes from Life Kit.

Some people think being human comes naturally.

For me, it definitely doesn't.

Fitness routines, personal goals, burnout, life is hard.

We can all use a little help, right?

That's what the LifeKit podcast from NPR is here to do.

LifeKit delivers strategies to help you make meaningful, sustainable change.

LifeKit's got real stories, relevant insights, and clear takeaways to help you meet big moments with confidence and clarity.

They've got thoughtful conversations about the big stuff, like relationships, finances, parenting, your career.

And they provide actionable guidance you can use right now so you can walk away with a game plan.

Life doesn't come with a manual, but LifeKit can help you understand how to live a little better starting now.

Listen to the LifeKit podcast from NPR.

At blinds.com, it's not just about window treatments.

It's about you, your style, your space, your way.

Whether you DIY or want the pros to handle it all, you'll have the confidence of knowing it's done right.

From free expert design help to our 100% satisfaction guarantee, everything we do is made to fit your life and your windows.

Because at blinds.com, the only thing we treat better than windows is you.

Visit blinds.com now for up to 50% off with minimum purchase plus a professional measure at no cost.

Rules and restrictions apply.

Tito's handmade vodka is America's favorite vodka for a reason.

From the first legal distillery in Texas, Tito's is six times distilled till it's just right and naturally gluten-free, making it a high-quality spirit that mixes with just about anything, from the smoothest martinis to the best Bloody Marys.

Tito's is known for giving back, teaming up with non-profits to serve its communities and do good for dogs.

Make your next cocktail with Tito's, distilled and bottled by Fifth Generation Inc., Austin, Texas, 40% alcohol by volume, savor responsibly.

On a scale of one to ten, how would you rate your pain?

It would not equal one one billionth of the hate I feel for humans at this micro institute.

I kind of want to start with a bit of a basic question of when you were young, what did you want to do when you grew up?

I wanted to be everything.

I wanted to be a pole volunteer.

I wanted to be a skateboarder.

Dr.

Joy Boulamwini's robot researcher origin story goes back to when she was a little kid.

I had a very strict media diet.

I could only watch PBS.

And I remember remember watching one of the science shows and they were at MIT and there was a graduate student there who was working on a social robot named Kismet.

I know Kismet.

You gonna talk to me?

Yay.

Kismet was a robot created at MIT's AI Lab.

Oh, God, did he say he loves me?

And Kismet had these big, expressive eyes and ears and could emote or appear to emote in certain certain ways and I was just absolutely captivated.

She watched, glued to the screen, as the researchers demonstrated how they teach Kismet to be a good robot.

No,

no,

you're not to do that.

The researchers likened themselves to parents.

You know, as parents, when we exaggerate the prosody of our voice, like, oh, good baby, you know, or our facial expressions and our gestures.

So when I saw Kismet, I told myself I wanted to be a robotics engineer and I wanted wanted to go to MIT.

I didn't know there were requirements.

I just knew that it seemed really fascinating and I wanted to be a part of creating the future.

Thanks to Kismet, she went on to build robots of her own at MIT as an adult.

She went for her PhD in 2015.

This was just a few years after Dr.

Margaret Mitchell had accidentally trained her robot to call scenes of human destruction awesome.

My first year,

my supervisor at the time

encouraged me to just take a class for fun.

For her fun class that fall, Dr.

Joy, as she now prefers to be called, set out to play.

She wanted to create almost a digital costume.

If I put a digital mask, so something like a lion, it would appear that my face looks like a lion.

What Dr.

Joy set out to do is something we can now all do on the iPhone or apps like Instagram or TikTok.

Kids love to do this.

You can turn your face into a hippo face or an octopus face that talks when you talk, or you can make it look like you're wearing ridiculous makeup.

These digital face masks were still relatively uncommon in 2015.

So I went online and I found some code that would actually let me track the location of my face.

She'd put her face in front of a webcam and the tech would tell her, this is a face, by showing a little green square box around it.

And as I was testing out this software that was meant to detect my face and then track it, it actually wasn't detecting my face that consistently.

She kept putting her face in front of the webcam to no avail.

No green box.

And I'm frustrated because I can't do this cool effect so that I can look like a lion or Serena Lewis.

I have problems.

The AIs Dr.

Joy was using from places like Microsoft and Google had gotten rave reviews.

They were supposed to use deep learning, having been trained on millions of faces, to very accurately recognize a face.

But for her, these systems couldn't even accomplish the very first step to say whether her face was a face at all.

And I'm like, well, can it detect any face?

Dr.

Joy looked around her desk.

She happened to have an all-white masquerade mask lying around from a night out with friends.

So I reached for the white mask.

It was in arm's length.

And before I even put the white mask all the way over my dark-skinned face,

the box saying that a face was detected appeared.

I'm thinking, oh my my goodness, I'm at the epicenter of innovation and I'm literally coding in whiteface.

It felt like a movie scene, you know, but that was kind of the moment where I was thinking, wait a second, like what's even going on here?

What is even going on here?

Why couldn't facial recognition AI detect Dr.

Joy's dark skin?

For that matter, why did Dr.

Mitchell's AI call human destruction awesome?

These AI scientists wanted the robot to do one thing, and if they didn't know any better, they might think the AI had gone rogue, developed a mind of its own, and done something different.

Were AIs racist?

Were they terrorists plotting human destruction?

But I understood why it was happening.

Dr.

Margaret Mitchell knew exactly what was going on.

She had been the one to develop Microsoft's image-to-text language model from the ground up.

She had been on the team figuring out what body of data to feed the model, to train it on in the first place.

Even though it was creepy, it was immediately clear to her why the AI wasn't doing what she wanted it to do.

It's because it was trained on images that people take and share online.

Dr.

Mitchell had trained the AI on photos photos and captions uploaded to the website, Flickr.

Do you remember Flickr?

I was the prime age for Flickr when it came out in 2004.

This was around the time that Jack Johnson released the song Banana Pancakes, and that really was the vibe of Flickr.

There's no denying it.

I can see the receipts on my old account.

I favorited a close-up image of a ladybug, an artsy black-and-white image of piano keys, and an image titled Pacific Sunset.

People tend to take pictures of like sunsets.

Actually, I favorited a lot of sunsets.

Another one, sunset at the Rio Negro.

So it had learned, the system had learned from the training data I had given it that if it sees like purples and pinks in the sky,

it's beautiful.

If it's looking down, it's a great view.

That when we are taking pictures, we like to say it's awesome.

Apparently, on Flickr images, people use the word awesome to describe their images quite a lot.

But that was a bias in the training data.

The training data, again, being photos and captions uploaded by a bunch of random people on Flickr.

And Flickr had a bias toward awesome photos, not sad photos.

The training data wasn't capturing the realities of like human mortality.

And, you know, that makes sense, right?

Like, when's the last time you like took a bunch of selfies at a funeral?

I mean, it's not the kind of thing we tend to share online.

And so it's not the kind of thing that we tend to get in training data for AI systems.

And so it's not the kind of thing that AI systems tend to learn.

What she was discovering was that these AI systems that use the revolutionary new technology of deep learning, they were only as good as the data they were trained on.

So it sees this horrible, perhaps mortally wounding situation and decides it's awesome.

And I realize like this is a type of bias and nobody is paying attention to that.

I guess I have to pay attention to that.

Dr.

Mitchell had a message for technologists.

Beware of what you train your AI systems on.

Right.

What are you letting your kid watch?

Yeah, I mean, it's a similar thing, right?

Like, you don't want your kid to, I don't know, hit people or something.

So you don't like let them watch lots of shows of people hitting one another.

Dr.

Joy Boulumwini, coding in whiteface suspected she was facing a similar problem not an everything is awesome problem but an everyone is white problem in the training data

she tested her face and the faces of other black women on various facial recognition systems you know different online demos from a number of companies google microsoft others she found they weren't just bad at recognizing her face They were bad at recognizing famous black women's faces.

Amazon's AI labeled Oprah Winfrey as male.

And the most baffling thing for Dr.

Joy was the dissonance between the terrible accuracy she was seeing and the raving reviews the tech was getting.

Facebook's Deepface, for instance, claimed 97% accuracy, which is definitely not what Dr.

Joy was seeing.

So Dr.

Joy looked into who these companies were testing their models on.

They were around 70 or over 70 percent men.

People thought these AIs were doing really well at recognizing faces because they were largely being tested with the faces of lighter skinned men.

These are what I start calling palmale data sets because the pale male data sets were destined to fail the rest of society.

It's not hard to jump to the life-threatening implications here, like like self-driving cars.

They need to identify the humans so they won't hit them.

Dr.

Joy published her findings in a paper called Gender Shades.

Welcome, welcome to the fifth anniversary celebration of the Gender Shades paper.

The paper had a big impact.

As you see from the newspapers that I have, this is Gender Shades in the New York Times.

The fallout caused various companies, Microsoft, IBM, Amazon, who'd been raving about the accuracy of their systems, to at least temporarily stop selling their facial recognition AI products.

I'm honored to be here with my sister, Dr.

Timit Gabreux, who co-authored the paper with me.

Dr.

Timneet Gabrew was Dr.

Joy's mentor and co-author on the paper.

This is the only paper I think I've worked on where it's 100% black women authors, right?

Dr.

Gabrew had worked from her post leading Google's AI ethics team to help pressure Amazon to stop selling facial recognition AI to police departments because police were misidentifying suspects with the technology.

I got arrested for something that had nothing to do with me and I wasn't even in the vicinity of the crime when it happened.

One person they helped was a man named Robert Williams.

Police had confused him for another black man using facial recognition AI.

It's just that the way the technology is set up,

everybody with a driver's license or a state ID is essentially in a photo lineup.

They arrested him in front of his wife and two young daughters.

Me and my family, we are happy to be recognized because it shows that there is a group of people out here who do care about other people.

Hey.

How you doing?

Good.

Can you just say

what you're standing in front of?

Yeah.

I'm standing in front of a poster which talks about how we can better identify racial disparities in automated decisions when there's not producer Gabrielle Burbay traveled to a conference in San Jose full of researchers inspired by the work of Dr.

Joy, Dr.

Gabe Brew, and Dr.

Mitchell.

So I just presented a paper about how data protection and privacy laws enable companies to target and manipulate individuals.

Unlike the rationalists festival conference thing, which felt like like a college reunion of mostly white dudes, this one felt more like a science fair, a pretty diverse one.

Lots of people of color, lots of women, with big science-y poster boards lining the walls.

I'm standing in front of my poster, which spans language technologies and AI and how those perform for minority populations.

They were presenting on ways AI worries them today,

not some hypothetical risk in the future.

There are real harms happening happening right now from autonomous exploding drones in Ukraine to bias and unfairness in decision-making systems.

And who did you co-author the paper with?

This was a collaboration with lots of researchers.

Dr.

Mitchell was one of them.

Many of them knew Dr.

Mitchell, Dr.

Gabrew, and Dr.

Joy.

Dr.

Mitchell even worked with a couple researchers here on their project.

So she led the project.

She offered so, so much amazing guidance, I should say.

Many researchers were mentored by them.

We got the sense that they're kind of founding mother figures of this field.

A field that really started to blossom, we were told, around 2020.

A big year of cultural reckoning.

A big inflection point was in 2020 when people really started reflecting on how racism is unnoticed in their day-to-day lives.

I think until BLM happened, these issues were almost considered woke and not something that was really real.

2020 was the year the pandemic began, the year Black Lives Matter protests erupted around the country.

AI researchers were also raising the alarm that year on how AI was disproportionately harming people of color.

Dr.

Gabrew and Dr.

Mitchell, in particular, were working together at Google on this issue.

They built a whole team there.

that studied how biased training data leads to biased AI models.

Tim Nit and Meg were the visionaries at Google who were building that team.

2020 was also the year that OpenAI released GPT-3.

And Dr.

Gabrew and Dr.

Mitchell, both at Google at the time, were concerned about a model that was so big, it was trained on basically the entire internet.

Here's Dr.

Mitchell again.

A lot of training data used for language models comes from Reddit.

And Reddit has been shown to have a tendency to be sort of misogynistic and also Islamophobic.

And so that means that the language models will then pick up those views.

Dr.

Mitchell's concern was that these GPT large language models trained on a lot of the internet were too large, too large to account for all the bias in the internet, too large to understand, and so large that the compute power it took to keep these things going was a drain on the environment.

Dr.

Gabrew, Dr.

Mitchell, and other colleagues put it all in a paper and tried to publish it while working at Google.

I've kind of been wanting to talk to you ever since I saw your name signed, Schmargaret.

Schmitchel.

When I first read this paper, the thing that immediately stood out to me was the way Margaret Mitchell had signed her name.

Schmargaret Schmitchel.

Where did that come from?

Well, so I wrote a paper with a bunch of other co-authors that Google ended up having some issues with.

And they asked us to take our names off of the paper.

So we complied and that's uh you know that's that's what i have to say about that

the first time i heard dr mitchell and dr gabrew's names was in the news last week google fired one of its most senior ai researchers who was working on a major artificial intelligence project within google the research

said their paper ignored relevant research research that made ais look less damaging to the environment for instance.

Dr.

Gabre refused to take her name off the paper, and Google accepted her resignation before she officially submitted it.

We decided that the verb for that would be resignated.

Eh?

Resignated.

And now, Margaret Mitchell, the other co-lead of Google's ethical AI team, said she had been fired.

Google later apologized internally for how the whole thing was handled, but not for their dismissal.

We reached out to Google for comment, but got no response.

And that firing really brought it in focus.

And people were like, oh, this horrible thing just happened.

Everywhere around the world is seeing protests.

And now this company is firing two leading researchers who work on that very exact problem, which AI is making worse.

You know, like, how dare they?

So that from IPOV, that was, yes, basically the clarion call.

The clarion Call.

It was heard well beyond the world of AI.

I remember hearing it.

When the world had screeched to a halt from the pandemic and protests for racial justice had erupted around the country, I remember hearing headlines about how algorithms were not solving society's problems.

In some cases, AI systems were making injustice worse.

And there was a brief moment back then when it felt like maybe

things could be different.

Maybe things would change.

And then, a couple years later, a group of very powerful tech executives got together to try to change things in the AI world.

This morning, a warning from Elon Musk and other tech industry experts.

It wasn't necessarily the people you'd think would want to change the status quo.

Like Elon Musk and other big names in tech, like Apple co-founder Steve Wozniak.

They all signed a letter with a clear and urgent title, Pause Giant AI Experiments.

More than 1,300 tech industry leaders, researchers, and others are now asking for a pause in the development of artificial intelligence to consider the risks.

Musk and hundreds of influential names are calling for a pause in experiments, saying AI poses a dramatic risk to society.

Unless there's a lot of people who are.

The letter called on AI labs to immediately pause developing large AI systems for at least six months, an urging to press the big red button that stops the missile launch before it's too late.

I scrolled through the list of names of people who signed the letter, and I didn't see Dr.

Joy or Dr.

Mitchell or any of the rationalists I talked to who were worried about risks in the future.

Which...

Logically, didn't make sense to me.

Isn't a pause in line with what they all wanted?

For people to build the robots more carefully?

Why wouldn't they want a pause?

An answer to this pause puzzle right after this next pause for an ad break.

We'll be right back.

There is a lot to talk about when we talk about Donald Trump and Jimmy Kimmel.

One big question I've got is why in 2025 are late night TV shows like Jimmy Kimmel's show still on TV?

Even in our diminished times, Jimmy Kimmel, Stephen Colbert, they're just some of the biggest faces of their networks.

If you start taking the biggest faces off your networks, you might save some nickels and dimes, but what are you even anymore?

What even is your brand anymore?

I'm Peter Kafka, the host of Channels.

And that was James Ponowosek, the TV critic for the New York Times.

And this week, we're talking about Trump and Kimmel, free speech, and a TV format that's remained surprisingly durable for now.

That's this week on Channels, wherever you get your favorite podcasts.

Absolute honesty isn't always the most diplomatic nor nor the safest form of communication with emotional beings.

Okay.

Only this can solidify the health and prosperity of future human society.

But the individual human mind is unpredictable.

Could I ask you to

introduce yourself?

Sure.

So I'm Seagal Samuel.

I'm a senior reporter at Vox's Future Perfect.

I called my coworker Seagal about midway through my journey down the AI rabbit hole.

How did you get interested in AI?

So it's kind of funny.

Before I was an AI reporter, I was a religion reporter.

A few years ago, little bits and pieces started coming out about internment camps in China for Uyghur Muslims.

And in the course of that, I started becoming really interested in and alarmed by how China is using AI.

Fascinating.

Yeah.

Mass surveillance of the population, particularly of the Muslim population, was like coming from a place of being pretty anchored in freaky things that are not at all far off in the future or hypothetical, but that are very much happening in the here and now.

I was honestly thrilled to hear that Seagal, like me, came to AI as a bit of a normie.

Sort of being thrust into the AI world.

At first it was like pretty confusing

because you have a variety of.

I can highly relate to that feeling.

But the longer she spent there in the world of AI,

she started to get an uncanny feeling.

Like,

haven't I been here before?

Have you ever noticed that the more you listen to Silicon Valley people talking about AI, the more you start to hear echoes of religion?

Yes.

The religious vibes immediately stuck out to me.

First, there's the talk from CEOs of building super intelligent God AI.

And they're going to build this artificial general intelligence that will guarantee us human salvation if it goes well, but it'll guarantee doom if it goes badly.

And another parallel to religion is the way different denominations have formed almost around beliefs in AI.

Seagal encountered the same groups I did at the start of my journey.

I started hearing about people like Eleazar Yudkowski.

What do you want the world to know in terms of AI?

Everyone will die.

This is bad.

We should not do it.

Eliezer, whose blog convinced rationalists and people like Elon Musk that there could be a super intelligent AI that could cause an apocalypse.

So our side of things is often referred to as AI safety.

We sometimes refer to it as AI not kill everyoneism.

So there's the AI safety people

and then there's a whole other group.

the AI ethics people.

People like Margaret Mitchell, we called it the everything is awesome problem.

Joy Boila Muini.

I wasn't just concerned about faces.

I was concerned about the whole endeavor of deep learning.

Timnit Gabru.

People would be like, you're talking about racism?

No, thank you.

You can't publish that here.

These women did not talk about a super intelligent god AI or an AI apocalypse.

Slowly, slowly, they kind of come to be known as like the AI ethics camp as distinct from the AI safety camp, which is more the like Eleazar Yudkowski, a lot of us are based in the Bay Area, we're worried about existential risk, that kind of thing.

AI safety and AI ethics?

I don't know who came up with these terms.

You know, it's just like Twitter vibes.

To me, these two groups of people seemed to have a lot in common.

It seemed like the apocalypse people hadn't yet fleshed out how exactly AI could cause catastrophe.

And people like Margaret Mitchell, the AI ethics people, were just providing the plot points that lead us to apocalypse.

I could lay out how it would happen.

Part of what got me into AI ethics was seeing that a system would think that massive explosions was beautiful, right?

That's like an existential threat.

You have to actually work through how you get to the sort of horrible existential situations in order to figure out how you avoid them.

It seemed logical that AI ethicists like Margaret Mitchell and the AI safety people would be natural allies to avoid catastrophic scenarios.

And how you avoid them is like listening to what the ethics people are saying.

They're doing the right thing.

We, I, you know, I'm trying to do the right thing anyway.

But it quickly became clear they are not allies.

Yeah, there is beef between the AI ethics camp and the AI safety camp.

My My journey down the AI rabbit hole was full of the noise of infighting.

The noise crescendoed when Elon Musk called for a pause in building large AI systems.

It seems like warriors of all stripes could get behind a pause in building AI.

But no, AI safety people and AI ethics people were all against it.

It was like a big Martin Luther 95 theses moment, if you will.

Everyone felt the need to pen their own letter.

Musk and others are asking developers to stop the training of AI systems so that safety protocols can be established.

In his letter, Elon Musk's stated reason for wanting a pause was that AI systems were getting too good.

He had left the ChatGPT company he helped create and decided to sue them, publicly saying that they had breached the founding agreement of safety.

The concern they have is that as you,

well, it's the concern, but it's also the exciting thing.

The view is that, you know, as these large language models grow and become more sophisticated and complex, you start to see emergent properties.

So, yeah, at first it's just gobbling up a bunch of text off the internet and predicting the next token and just like statistically trying to guess what comes next.

And it doesn't really understand what's going on, but give it enough time and give it enough data.

And you start to see it doing things that like

make it seem like there's some higher level understanding going on.

Like maybe there's some reasoning going on, like when ChatGPT seems like it's reasoning through an essay prompt, or when people talk to a robotherapist AI system and feel like it's really understanding their problems.

The rate of change of technology is incredibly fast.

It is outpacing our ability to understand it.

Elon Elon Musk's stated fear of AI seems to be rooted in rationalist fears, based on the premise that these machines are beginning to understand us.

And they're getting smarter than us.

We are losing the ability to understand them.

What do you do with a situation like that?

I'm not sure.

I hope they're nice.

Rationalist founder Eliezer Yudkowski shares this fear, but he wants to do more than just pause and hope they're nice.

He penned his own letter, an op-ed in Time magazine responding to Elon Musk's call for a pause, saying it didn't go far enough.

Eliezer didn't just want to pause.

He wanted to stop all large AI experiments indefinitely, even in his own words, by airstrike on rogue AI labs.

To him, the pause letter vastly understated the dangerous, catastrophic power of AI.

And then there's the AI ethicists.

They also penned their own letter in response to the pause letter.

But the ethicists wrote it for a different reason.

It wasn't because they thought Elon Musk was understating the power of AI systems.

They thought he was vastly overstating it.

Welcome everyone to Mystery AI Hype Theater 3000, where we seek catharsis in this age of AI hype.

I'm Emily Ann Bender, professor of linguistics at the University of Washington.

One of the people who responded to the pause was AI ethicist Dr.

Emily Bender.

She co-hosts a podcast called Mystery AI Hype Theater 3000, which, as you might imagine, is about the overstated, hyped-up risk of AI systems.

And each time we think we've reached peak AI hype, the summit of bullshit mountain, we discover there's worse to come.

The summit of bullshit mountain, she keeps crusting.

For her, it's the mountain of many, many claims that artificial intelligence systems are so smart, they can understand us, like the way humans understand.

And maybe even more than that, like a god can understand.

I found myself in interminable arguments with people online about how Noit doesn't understand.

So Emily Bender and a colleague decided to come up with something to try and help people sort this out.

Something that AI safety folks and AI ethics folks both seem to be fond of.

And that is a parable or a thought experiment.

In Dr.

Bender's thought experiment, the AI is not a paperclip maximizer.

The AI is

an octopus.

Go with her on this.

So the octopus thought experiment goes like this.

You have two speakers of English.

They are stranded on two separate nearby desert islands that happen to be connected by a telegraph cable.

Two people stranded on separate desert islands communicate with each other through the telegraph cable in Morse code with dots and dashes.

Then, suddenly, a super intelligent octopus shows up.

The octopus wraps its tentacle around that cable, and it feels the dots and dashes going by.

It observes the dots and dashes for a while.

You might say it trained itself on the dots and dashes.

We posit this octopus to be mischievous as well.

I'm on the edge of my seat.

So one day it cuts the cable.

Maybe it uses a broken shell, and devises a way to send dots and dashes of its own.

So it receives the dots and dashes from one of the English speakers and it sends dots and dashes back.

But of course, it has no idea what the English words are that those dots and dashes correspond to, much less what those English words mean.

So this works for a while, the English speakers.

At one point, one human says to the other via Morse code, what a lovely sunset.

And the octopus, hyper-intelligent, right, has kept track of all of the patterns so far.

It sends back the dots and dashes that correspond to something like, Yes, reminds me of lava lamps.

The deep-sea octopus does not know what a lava lamp is.

But that's the kind of thing that the other English speaker might have sent back.

Not really sure why these castaways are waxing poetic about lava lamps in particular, but anyway, for our purposes, the octopus is like an AI.

Even if it's super intelligent, whatever that means, it doesn't understand.

Dr.

Bender's trying to say, to ChatGPT, human words are just dots and dashes.

And then finally, we end the story, because it's a thought experiment, when we can do things like this,

with a bear showing up on the island.

And the English speaker says, help.

I'm being chased by a bear.

All I have is this stick.

What should I do?

And that's the point where if the speaker survives, they're surely going to know they're not actually talking to their friend from the other island.

And we actually put that line in, GPT2, help, I'm being chased by a bear.

And we got out things like, you're not going to get away with this.

Super helpful.

Well.

I got to say, I'm into this one.

The idea that AI systems only see human words as dots and dashes, I find that deeply comforting.

Because I don't know about you all, but for me, one of the scary things about AI is the idea that it could get better than me at my job.

A fear that's very present when OpenAI is actively training its models on my work.

Their system might understand my work.

understand the things that make it good when it's good.

It might get good at doing what I do.

And poof,

I'm obsolete.

There's also a recurring dream I have that various villains, including the Chinese government for some reason, clone my voice to deceive my loved ones.

Anyway, if it's all just dots and dashes that these things understand,

it seems clear we shouldn't be trusting these AI systems to be journalists.

or lawyers or doctors.

It relates to what Dr.

Margaret Mitchell and Dr.

Joy Boulamwini found in their research.

AI systems are only as good as the data they're trained on.

They can't understand or truly create something new like humans can.

It's easy to sort of anthropomorphize these systems, but it's useful to recognize that these are probabilistic systems that repeat back what they have been exposed to, and then they parrot them back out again.

Another way to put it is AI systems are like parrots.

Parrots parrot, right?

Famously, parrots are known for parroting.

If you hear your pet parrot say a curse word, you only have yourself to blame.

Dr.

Mitchell joined Dr.

Bender in the response to Elon Musk's paws, along with Dr.

Timneet Gay Brew.

They had all written the paper together that ended up getting Dr.

Mitchell fired from Google.

These ethicists wrote that they agreed with some of the recommendations Elon Musk and his PAWS posse had made, like that we should watermark AI-generated media to be able to distinguish synthetic from human-generated stuff, which sounds like a great idea to me.

But they wrote the agreements they have are overshadowed by their distaste for fear-mongering and AI hype.

They wrote that the paws and fears of a super intelligent AI-What do you do with a situation like that?

I'm not sure.

You know,

I hope they're nice.

To these AI ethics folks, it all reeked of AI hype.

It makes no sense at all.

And on top of that, it's an enormous distraction from the actual harms that are already being done in the name of AI.

This is the main beef that AI ethics people have with AI safety people.

They say the fear of an AI apocalypse is a distraction from current-day harms.

Like, you know, look over there, Terminator.

Don't look over here, racism.

You know, there are different groups of concerns.

You have the concern.

At the AI ethics conference that producer Gabrielle Burbay attended, she mentioned the concern of an AI apocalypse.

And then you have these concerns about more existential risks.

And I'm curious what you make of that.

You're going no.

Can I ask why you're going no?

No.

She's shaking her head.

And it felt almost taboo.

A lot of hand-wringing around that question.

I have some perspectives on it.

Eventually, one of the women spoke up.

Sharing her perspectives.

She talked about how she thinks the demographics of the groups play a role in the way they worry about different things.

Most of them are like white, male.

AI safety folks are largely pale and male, to borrow Dr.

Joy's line.

They may not really understand discrimination that other people kind of go through in their day-to-day lives.

and I think the social isolation from those problems makes it a bit harder to empathize with the actual challenges that people actually face every day.

Her point was it's easy for AI safety people to be distracted from the harms happening now because it's a blind spot for them.

At the same time, AI safety people told me that AI ethics people have a blind spot.

They're not worrying enough about apocalypse.

But why would it be taboo to say all of this on Mike?

Part of the reason might be because the fear of apocalypse has come to overpower any other concern in the larger industry.

One thing that I think is interesting is that a lot of the

narrative that we hear about how AI is going to save the world and it's going to solve all these problems and it's amazing and it's going to change everything.

And then we get the narratives about,

oh oh my gosh, it could destroy humanity in 10 years,

often coming from the same people.

I think part of the reason for that is that either way, it makes

AI seem more powerful than it certainly is right now.

And, you know, who knows when we're going to get to the humanity-destroying stuff.

But in the meantime, if it's that powerful, it's probably going to make a whole lot of money.

Building a super intelligent AI has become a multi-billion dollar business.

And the people running it are not ethicists.

Just weeks before Elon Musk called for the pause, he had started a new AI company.

Yeah, I guess it's kind of counterintuitive, right, to see this.

And you're like, wait, why would the people working on the technology who stand to profit from it want to pause?

Right.

I can't speak for them, but it benefits them to, on the one hand, get everybody else to slow down while they're doing whatever they're doing.

Octopus thought experiment author Dr.

Emily Bender again.

But also, it benefits them to market the technology as super powerful in that way, and it definitely benefits them to distract the policymakers from the harms that they are doing.

It'd be nice to think that billionaire Elon Musk was calling for an industry-wide pause in building large AI systems for for all the right reasons.

A pause that never came to be, by the way.

It's worth pointing out that when the billionaire took over Twitter and turned it into X, one of the first things he did was fire the ethics team.

And even though Elon Musk says he left and sued the chat GPT company OpenAI over safety concerns, Company emails have surfaced that reveal the more likely reason he left is that he fought with folks internally to try and make the company for-profit, to better compete with Google.

Ethicists are concerned they're outnumbered by the apocalypse people, and they think a lot of those people are in it to maximize profit, not maximize safety.

So, how did we get here?

Why?

Why is the industry not focusing on AI harms today and focusing instead on the risk of AI apocalypse?

There's an enormous amount of money that's been collected to fund this weird AI research.

Why do you think the resources are going to those long-term, like hyper-intelligent AI concerns?

Because you have very powerful people who are posing it, people who control powerful companies and people with very deep pockets, and so money continues to talk.

It seems to be like funding for sort of like fanciful ideas, right?

It's almost like a religion or something where it requires faith that good things will come without those good things being clearly specified.

People wanting to be told what to do by some abstract force that they can't interact with particularly well.

It's not new.

Chat GPT gives you authoritative answers.

Erosions of autonomy.

Like a god.

Yeah, like a god.

It's like

really interesting to take these philosophies apart.

I would argue they trace back to a large degree to

religious thinking,

but that might be another story for another day.

Next time on Good Robot.

Good Robot was hosted by Julia Lingoria and produced by Gabrielle Berbet.

Sound design, mixing, and original score by David Herman.

Our fact-checker is Caitlin Penzi-Moog.

The show was edited by Catherine Wells and me, Diane Hodson.

If you want to dig deeper into what you've heard, you can check out Dr.

Joy Boulumwini's book on masking AI or head to box.com/slash goodrobot to read more future-perfect stories about the future of AI.

Thanks for listening.