Celebrities Fight Sora + Amazon’s Secret Automation Plans + ChatGPT Gets a Browser

1h 7m
“Now we are just seeing OpenAI do the full Facebook when it comes to content policy.”

Listen and follow along

Transcript

This podcast is supported by the all-new 2025 Volkswagen Tiguan.

A massage chair might seem a bit extravagant, especially these days.

Eight different settings, adjustable intensity, plus it's heated, and it just feels so good.

Yes, a massage chair might seem a bit extravagant, but when it can come with a car,

suddenly it seems quite practical.

The all-new 2025 Volkswagen Tiguan, packed with premium features like available massaging front seats, that only feels extravagant.

This was crazy.

Google's Willow Quantum Chip is using a new quantum echoes algorithm that ran computations 13,000 times faster than supercomputers, Kevin.

Oh, I see it's performance review season over there in Google Quantum Computing.

Oh,

you know, my echo.

Chip did a quantum compute.

You know, I need a raise.

No matter how many times I learn what quantum computing is, I do immediately forget it the next day.

And it just, like, this is how I am.

This is why I'm so, why I love reading mystery so much is because I forget who did it like the day after I put the book down.

That's what quantum computing is for me.

You know, we have to fill out our performance reviews soon at the New York Times.

And I think I'm just going to put in there that I solved a quantum computing problem this year because how will they fact check me?

Now, why don't they email me asking to like help on your performance review?

Oh, you want to do a 360 review?

I want to do a 360.

You've got some feedback?

Yeah.

I'm Kevin Roos, a tech columnist at the New York Times.

I'm Casey Noon from Platformer.

And this is Hard Fork.

This week, OpenAI's big sloppy mess, why the company is backpedaling over Sora.

Then, the Times Karen Weiss joins us to discuss her scoop on Amazon's plans to reduce its hiring needs by hundreds of thousands of workers.

And finally, AI Browsers are here.

Our first impressions of ChatGPT Atlas.

Well, Casey, it's been another busy week for the OpenAI Research and Deployment Corporation.

I learned that's what they call themselves.

Really?

Yeah, they have these hoodies.

I saw a guy on the train the other day with a research and deployment company.

It didn't even say OpenAI, but that's sort of their new tagline.

Interesting.

Well, I would say based on the events of the past week, Kevin, maybe OpenAI should do do a little more research and a little less deployment.

Yeah, so let's talk about it.

We're going to talk about two OpenAI stories this week, one about their new browser.

We'll talk about a little later.

But first, we got to talk about what's been happening with Sora.

We've talked about this the last couple of weeks on the show, but this continues to be a total mess for OpenAI, this app and the various sort of controversies and backlashes swirling around it.

So Casey, what is going on with Sora?

What is the latest here?

Well, I would say there have been two big developments over the past week, Kevin.

One, the company has said that it is going to essentially like crack down on political deep fakes based on historical figures after the families of some deceased political figures started to complain.

And then the company has also said it's going to try to build some guardrails around the use of copyrighted intellectual property after many people in Hollywood freaked out, including breaking bad star Brian Cranston.

Yes, yes.

They managed to beef with Brian Cranston and the estate of Martin Luther King Jr.

in one week.

And Casey, I think that qualifies as a bad week at the office.

It's not a great week.

It's like, there's sort of this whole genre of like what I call bad beginnings, which is like when you start saying something and you realize like, oh, this is not going well for me.

It's like things that would fall into this category include, per my recent conversations with the estate of Martin Luther King Jr.

Also in this category, regarding the Nazi tattoo I got while in the Marines and regarding the amount of lead in my protein shakes, you know, when you said any of those things, it's not been a good week.

Not been a good week.

So let's start with Martin Luther King Jr.

and his estate and their beauty.

What was he?

Why is he so significant a figure in American history?

Well, according to my Sora feed, he's a sort of historical civil rights icon who liked to get up and give speeches about skibbity Ohio toilet ruse.

He also appears to love to play Fortnite based on the Sora videos I've seen.

So what we're talking about is this sort of emerging genre of Sora videos, which I started seeing pretty soon after downloading the app, where people would just take Martin Luther King's iconic speeches, such as I Have a Dream, and make him say other things.

Things talking about Gen Z trends, things talking about video games, endorsing various products.

This was funny to some people, offensive to others.

You know who didn't like it was the estate of Martin Luther King Jr.

Yeah, and it wasn't all, you know, playing Fortnite and talking skibbity toilet.

Some people were also having MLK make monkey noises and putting him in other just like overtly racist situations.

And so, yeah, his family members complained.

The Washington Post wrote a great story about families of him and other deceased historical figures saying, hey, like this really sucks.

And OpenAI's original position had been, we believe in free expression.

People should be able to do what they want.

But I don't know.

At some point, something changed.

And next thing you know, OpenAI is on X posting a statement saying, while there are strong free speech interests in depicting historical figures, OpenAI believes public figures and their families should ultimately have control over how their likeness is used, which was a brand new policy as of the moment that they posted that.

And this is somewhat confusing to me because part of the way that Sora works is that in order to make a cameo of someone to use their face in a video, they have to sort of give you permission to do that.

So presumably, you know, Martin Luther King Jr.'s estate did not go into their Sora settings and say, anyone can make a photo of me, but they sort of used some public figure loophole or how did that work?

That's right.

So if you're just an average person, people cannot go in like you, Kevin, like a very average person.

I can't just go in and make a cameo of you unless you have changed your settings that way.

But OpenAI basically said it is open season for historical figures.

And of course, there's lots of video out there of MLK and others.

And they just said, yeah, go crazy if you want.

Got it.

So now they're saying, actually, we've thought about it.

And after consulting with the King Estate, we are no longer letting people do this.

Yeah.

Like, what happens if you try to make a video with Martin Luther King Jr.

now?

Now it will just, you'll just get blocked.

You know, it violates the content policies.

But, you know, I just want to say it was so obvious that people were going to do this.

You know, and in its ex-post, OpenAI suggested that the reason that it had made this change was that people were making, quote, disrespectful videos of MLK.

Like, you really thought that people were only going to make respectful videos of historical figures.

Like, let me be clear.

The only reason to use Sora is to create a video of someone doing something that they would not ordinarily be doing.

Yes, right?

It is not a technology to make people give beautiful speeches about civil rights.

I confess I am somewhat implicated in this because I have not made a Martin Luther King Jr.

Sora video, but I did make a video of Mr.

Rogers saying Gen Z catchphrases

because I thought it was funny.

After everything Fred Rogers did for this country, and this is how you repay him.

I felt bad about it.

If it makes you, it makes it any better.

I did have a moment of like, you know, guilt and shame after doing it.

Did I do it anyway?

Yes.

Did it get approximately four likes?

Also, yes.

But I mean, look, this, here's what I'm telling you.

This is what the technology is for.

It is doing exactly this thing.

And so if you don't have in your mind a policy for how you want to handle that before you launch it, I think you're doing something irresponsible.

Okay, so the estate of Martin Luther King Jr.

is mad at Open AI over Sora.

Who Who else is mad at Open AI over Sora?

Well, Kevin, that brings us to Brian Cranston, who presumably was, you know, minding his own business down in Albuquerque making methamphetamines, when all of a sudden he opens up the Sora app and finds himself in videos with Michael Jackson and Ronald McDonald, which is what we like to call around here a nightmare blunt rotation.

Now, I have not seen these videos.

What did they show?

I actually haven't seen them myself because I don't want to support Ronald McDonald that way.

I think he has a lot to answer for.

Yeah.

So here's why this is a problem.

This was supposed to be an opt-in regime.

If celebrities' images were going to appear in Sora, it was supposed to be that they had to opt in.

But as Winston Cho reported at the Hollywood Reporter last week, that's actually not what happened.

Days before the release of Sora, OpenAI went to the big talent agencies and the studios and said, hey, if you don't want all of your intellectual property in our app, you have to opt out.

Which companies like Disney were putting out statements being like, that's not actually how copyright works.

Like, you don't have Carte Blanche to do whatever you want with our IP unless we opt out.

And so this starts to get people in Hollywood really mad.

Got it.

So I saw the statement from Brian Cranston and SAG AFTRA, which is the union that represents actors and a number of other talent agencies basically saying, hey, we don't like this.

But what are they saying about how OpenAI has been approaching them?

Because OpenAI, from my understanding, did actually try to sort of go to Hollywood before this app came out and say, hey, just FYI, we are going to be releasing this, but we have like taken steps to sort of get ahead of some of your the issues we think you might have with it.

That's right.

But in practice, this just was not true.

People were able to create videos of Pokemon and Star Wars and Rick and Morty and other intellectual properties whose owners had never given their permission.

Brian Cranston had not given permission for his likeness to be used in the app.

And so you wind up having a lot of what OpenAI, I always love these euphemisms that these companies use, OpenAI calls these unwanted generations.

And I was like, like Generation Z, but it turns out that no, this is about unwanted videos appearing within the Sora feed.

So, you know, it's very funny to me to come out afterwards and saying, on reflection, we'd like to strengthen the guardrails when in fact there were not guardrails.

You know what I mean?

Yes.

If I drive my car off the side of the road because there's no guardrail and the California Department of Transportation says we're going to strengthen these guardrails, I'm saying,

where was the guardrail?

I'm dead, and I'm shouting at you from hell saying, where was the guardrail, Kevin?

Right.

There is a lot of like, I don't know, like false naivete or like just people being like feigning surprise.

Like, I cannot believe that my unauthorized generation app is causing problems over unauthorized generations.

It's crazy.

It's, I am so glad you said that because this is the thing that has got me so exercised over the past week.

It is that phony naivete.

It is this, wow, who could have ever predicted this?

Because that is just an approach that I think if you apply it to building AI products in the future, it's going to take us to some very bad places.

Okay, so Open AI is dealing with this backlash.

I think there's sort of a larger backlash brewing over just AI generated video.

And I'm curious what you make of this.

That like, I think there is starting to become a consensus position, especially among people who are like not in San Francisco and do not work in the AI industry, that like all of this is just like bad and stupid and harmful and that the sort of juice is not worth the squeeze, as it were.

That like the benefits of AI, whatever they might be in the future, are not enough to justify the enormous costs of training these models.

There is something sort of soulless and depressing about people using AI to generate fake videos of Martin Luther King Jr.

and Brian Cranston and Ronald McDonald doing various things.

I guess I'm curious whether you think the Sora backlash is part of that or whether what we are just seeing is one manifestation of a pre-existing thing where people were already mad about this stuff.

We are going to have to get survey data to develop an empirical answer to that question.

But we know from like a recent Pew survey that already about half of Americans say that they are more concerned than excited about the future of AI.

And my assumption is that the Sora backlash is going to fuel that.

When I just look at my own interaction with friends and families, the default feeling about Sora is not, what a fun new creative tool.

It is, this is bad and I hate it.

And by the way, these aren't even necessarily people who are like up in arms about what's going on with MLK and Brian Cranston.

This is just sort of giving them the ick.

Yeah.

I mean, to me, this just seems like a continuation of this pattern at Open AI that extends back to the launch of Advanced Voice Mode last year, when you can probably remember Scarlett Johansson kind of objected to references to the movie Her.

Open AI had basically approached.

Scarlett Johansson and said, hey, would you like to be supportive of or involved with this launch?

Could we sort of explicitly tie this to your character in the movie Her?

She said, no, they went ahead and did it anyway.

And it seems like that is something that they have continued to, rather than sort of being chastised by that and learning from that experience and saying, hey, maybe it's important that we have like the permission of the creators in Hollywood before we go out and do something that's potentially disruptive to them.

Maybe we should get their permission.

It seems like they have not learned that lesson.

That's right.

And that's really where I kind of want to land this.

This is why I think all of this matters is I think that in building any kind of novel technology, inevitably companies are going to make mistakes.

They're going to go too far in some regard.

There's going to be some problem that they didn't anticipate, and it's bad, and we should talk about it.

But I think companies can kind of come back from that.

But then there are other companies that just kind of start to make the same mistake over and over again, right?

You bring up the Scarlet Johansson issue, which I think partly came out of just a rush to release this voice mode into the general public.

And look at what else we have seen over the past year.

I think there was a similar rush to update GPT-4.0 with what turned out to be a very sycophantic update that was embarrassing to the company.

There was a rush to release chat GPT in ways that sort of cut users' access off to tools that they had become very dependent on, and it triggered this huge backlash.

And now here they are in this rush to release this video app in part because they want to make money, the company has said.

And lo and behold, they either have not thought through the policy implications or they've just decided to build a policy that could only possibly bring them a huge backlash.

So I look at that, Kevin, and I fear this company actually has changed a lot over the past couple of years, right?

How do you think it's changed?

Well, if you look at them before the launch of ChatGPT and even in the few months after that, this was a company that was talking a lot about, on one hand, wanting to introduce new technologies to the public to see how society would adapt and to do that in like a way that was too aggressive for some people, but I think was basically working out okay in the original ChatGPT era.

Sam Altman was going around Washington, meeting with senators saying, hey, we're building something that could be really dangerous.

We want guardrails around this.

We want you to pass regulations that rein us in.

And then you just fast forward today.

And it's this all-out war between a handful of companies that are trying to build AGI faster than the other guy.

And we are just seeing in real time, not just like guardrails being removed, we are seeing guardrails not being built and the company having to come in afterwards and say, Oh, hey, sorry about that.

Yeah, we're gonna do something.

We're hearing your feedback.

And the thing that just shocks me about that is I actually believe for a time that Sam Altman had taken the lessons of Facebook and the social media backlash.

He had seen everything that had happened to Mark Zuckerberg.

He said to himself, I am not going to make those same mistakes.

And now we are just seeing OpenAI do the full Facebook when it comes to content policy.

Well, and I would say two things about this strategy of OpenAIs: one is it is brash, it is risky, it is likely to lead to lots of backlash and people being mad at them.

And I think it is potentially correct.

I mean, what we've seen over the past few years is that there are not a lot of real restraints on companies that want to build and release technology this way.

I think the real risk to open AI is that people just end up losing faith in AI as a whole.

And as we've talked about recently on the show, like the entire economy now kind of rests on this belief that AI is growing more powerful, that it will soon deliver all of these like tangible economic and social and scientific benefits to people, that it is not just like hoovering up a bunch of people's data and using it to make slop.

And if that's what sort of the public image of this stuff becomes because Open AI has adopted this product strategy, I think that will be bad for the whole AI industry, but probably not especially bad for OpenAI.

Yeah, I mean, I think like that is a fairly cynical view, Kevin.

Like, it's true in a lot of ways.

We are in the LOL, nothing matters era of content moderation.

And I am just reflecting once again on how, like, we used to have a world of like business and politics where people would go to great lengths to avoid feeling shame.

And at some point in, let's say, the past decade, we just decided we're not going to care about that anymore.

And no one can make us feel ashamed for any reason.

And for the moment, I guess the only real impact we're seeing here is that, you know, a few copyright holders and like families of historical figures are annoyed by videos that they're seeing online.

But this is the company that continues to build ever more powerful technology.

And when GPT-7 comes out and is helping novices build novel bioweapons, I don't want there to be an ex post saying that based on recent pandemics, the company has decided to to build some guardrails.

Right.

Right.

I mean, I've been talking with a few people sort of in and around Open AI about this over the past few weeks, just kind of taking the temperature of like how folks over there are feeling about this.

And a couple of things that I've heard that I want to just run by you for a reaction.

One is like, this is a company that does not have the benefits of having hundreds of billions of dollars a year in search revenue flooding in the door that it can use to build AI stuff.

Like that is the situation that Google, its sort of next biggest competitor, is in.

They basically don't have to care about money.

They can spend all of their profits on curing cancer and building quantum computers and self-driving cars and whatnot.

But like Open AI kind of doesn't have that luxury.

And so they have to figure out ways to pay for their enormous ambitions.

And not all of those are going to be sort of obviously pro-social and beneficial things, but the ends will justify the means, just as, you know, Google spent, you know, many years, they say, building up this monopoly and doing all sorts of unsavory things in order to get the profits that they can then plow back into the sort of peace dividend of AI research.

I mean, I reject that for two reasons.

One, this company's stated mission is to build AI that benefits all humanity.

So it's like, if the argument is in order to benefit all humanity, we have to harm some of humanity, get a new mission statement, girl.

Like, come on.

Number two, I also reject the premise that they have some cash crunch.

Sam Altman is the greatest fundraiser in the history of Silicon Valley.

This company has access to all the capital it needs.

So don't tell me that you need to release the infinite slot machine that makes Brian Cranston cry in order to, you know, build your machine god.

I don't think Brian Cranston is actually crying unless that sort of video I saw was legit.

But another thing that I will hear from people at OpenAI is about what they call iterative deployment, which is one of their favorite catchphrases over there.

They basically believe that instead of keeping all of this research and all these capabilities cooped up inside the lab and then kind of releasing them all at once every few years, that we should have kind of a steady drip of new capabilities from these companies that sort of help the public update about what is now possible with AI.

And so one defense of Sora that I've heard from people over there is they'll say, look, this technology exists.

These video models are getting quite good.

And we could either sort of spring this on you all when it is is impossible to tell the difference between fake and real and, you know, without any of these safeguards, or we could kind of release it in this iterative way where we kind of give the world a chance to adjust and catch up and have these conversations and arguments about likenesses and copyright and sort of prepare the world for this new capability that exists and that that is the responsible thing to do.

What do you make of that?

Well, I just think that there are so many more responsible ways to do it than saying there is now an app where anyone can go on and make a video of Martin Luther King barbecuing Pikachu, right?

You could just make whatever deep fakes you want and put them on a website and say, hey, look at the terrifyingly real deep fakes we were able to make with this technology.

We're not going to release it to the public, but just so you know, if you start seeing videos out there that seem like maybe they didn't happen, maybe they didn't.

Or you could say, we're going to just make this available in our API.

And so developers have access to it, but we're going to closely monitor how developers are using it.

And if there are bad actors in our development ecosystem, we are going to get rid of them, right?

So those would be two alternatives to just saying, hey, everybody, go freaking nuts.

Right.

Yeah, I think those are both good responses.

I don't find any of the defenses of Sora from the OpenAI folks I've talked to all that compelling, but I think they are learning a lesson actually from the social media companies, which is, you know, you do something bold and brash with very few guardrails, people get mad at you about it, and you scale it back 10%.

But you've still kind of taken that yardage, even if you have to turn the dials to install some guardrails after the fact.

Like you still have kind of gotten what you came for, even if you end up having to make some compromises.

Yes.

And in that, when I look at this story, Kevin, I just see exactly what YouTube did in its early days, right?

YouTube also started out by saying, hey, why don't you just upload whatever you want onto our website?

And we're just going to sort of take for granted that you have the copyright over whatever you're uploading.

And eventually, Viacom comes along and says, there are more than 100,000 clips of our TV shows and movies all over your network, and we're going to sue you for a billion dollars.

And this wound up being a kind of costly legal battle.

It went on for a very long time.

It was eventually settled.

But during the time it took for that case to settle, YouTube became the biggest video site in the world, and it won the whole game.

And so I think that there is a very cynical rationale for everything that we're seeing OpenEye do, which is saying, hey, we have the opportunity to go get all that market share.

We're going to do it.

Yeah.

It's what they call regulatory arbitrage, right?

That's one thing you could call it.

What else would you call it?

Well, this is a family program, Kevin.

I'm going to try to be polite.

Okay, so that is the next turn of the screw in the Sora story.

Casey, what are you looking at with this story going forward?

Here is what I'm looking at going forward.

Open AI, for better and for worse, is a company that is shipping a lot of products.

We're going to talk about another one of them later in this show, right?

This is an organization that has figured out how to build and release new stuff.

And that stuff does some really cool stuff.

And as with Sora, does some pretty gross stuff.

I think the thing to keep your eye on as these new products come out is, is this company truly paying attention to responsibility anymore?

Or is the entire ethos of the company now just a land grab for as many users across as many surfaces as it can get?

Because if that is going to be the new MO for this company, then I think we need to be a lot more worried about it than at least I personally have been to date.

Yeah, I think they are in a real like throwing spaghetti against the wall phase here.

And I think that is reflected in just how many things they're shipping constantly and seemingly a new product or two every week.

And some of it'll work and most of it probably won't.

But, you know, Casey, one of the best pieces of advice I ever got about journalism was that the stories you don't write are as important, if not more important, than the stories you write.

And I think OpenAI has not learned how to say no to a new idea or a product or a business line yet.

And I think that's a skill that they should start developing because it it seems like they are spreading their bets quite thin.

They are throwing a lot of spaghetti at the wall, and maybe they're losing the plot a little bit.

Now, is that advice why you write so few stories?

Yeah, yes.

Okay, interesting.

Yes.

That editor.

I'm very proud of the stories I don't write, though.

That editor really took a number on you.

Yeah.

When we come back, we'll take a look inside Amazon's secret plans to automate hundreds of thousands of jobs with next generation robots.

This podcast is supported by Crucible Moments, a podcast about the decisions and inflection points that shape the journeys of the world's most influential startups.

Hosted by Sequoia Capital's Rulaf Botha, Crucible Moments Season 3 pulls back the curtain on the untold stories behind companies like Stripe, Zipline, Palo Alto Networks, Klarna, Supercell, and more.

Tune into Sequoia's new season of Crucible Moments to discover how some of the most transformational companies of the modern era were built.

Crucible Moments is available everywhere you get your podcasts.

A warm dinner, a full table, peace of mind for every family.

That's the holiday we all wish for.

Your donation to Feeding America helps make it possible, not just for one family, but for communities everywhere.

Because when we act together, hope grows.

Give now and your impact may be doubled.

Visit feedingamerica.org slash holiday.

Brought to you by Feeding America and the Ad Council.

Don't just imagine a better future.

Start investing in one with betterment.

Whether it's saving for today or building wealth for tomorrow, we help people in small businesses put their money to work.

We automate to make saving simpler.

We optimize to make investing smarter.

We build innovative technology backed by financial experts for anyone who's ever said, I think I can do better.

So be invested in yourself.

Be invested in your business.

Be invested in better with betterment.

Get started at betterment.com.

Investing involves risk.

Performance not guaranteed.

Well, Kevin, there's a new story out there about robotos, but some people are not saying domo errigato.

That's right, of course.

I'm talking about Karen Weiss's story this week in the New York Times saying that Amazon plans to eliminate a bunch of jobs using robots.

Yes, this was a big story this week and I'm very excited to have Karen on to talk about it.

The basic idea here is that Amazon has made plans, plans that is not shared with the public, to replace more than half a million jobs with robots.

And Karen, my lovely colleague at the Times, got a hold of some of these internal strategy documents in which they are laying out these plans.

And this story has been causing a big stir.

I think people are sort of fearful of job loss from AI and automation right now.

That's been obviously a big topic in the news.

And what we're seeing now is one of America's largest employers saying in its internal documents, yeah, we're doing it.

That's right.

You know, it's one thing over the past couple of years to have discussed, as we often have on this podcast, the risk that this technology will someday be good enough that a lot of people will be put out of work.

It is something very different to see America's second largest private employer saying, we have an actual plan to make this happen.

Yeah.

So to talk about this story and how Amazon is racing toward its goal of full automation, we are inviting back New York Times reporter and friend of the pod, Karen Weiss.

She's been covering Amazon for nearly a decade for the Times and recently visited a warehouse in Shreveport, Louisiana, where they are putting a bunch of their new robotics to the test.

And I think it's a prime day to talk with her.

Oh, I get what you did there.

Yeah.

Karen always delivers.

Karen Weiss, welcome back to Hard Fork.

Happy to join you guys.

So this was a fascinating story.

I really enjoyed it and learned a lot from it.

It caused a big stir.

I heard lots of people talking about this plan that Amazon has to replace a bunch of jobs using robots.

And I want to start with how you decided to look into this, because this is a subject that people have been talking about for many years.

Amazon obviously has been putting robots in its warehouses for a long time.

We, Casey and I, went to an Amazon warehouse last year and saw what looked to be like a huge fleet of robots sort of moving around picking up containers and bringing them to people who would pick things off them and put them in boxes.

But what made you think that this was taking a step forward that was important for you to write about?

Yeah, I've covered the company since 2018, and it's more than tripled its headcount since then.

So there was this period of just tremendous growth.

And then it started plateauing or almost plateauing.

And you could see every quarter when I cover earnings, you would see what was this huge growth, particularly obviously in the early days of the pandemic.

You started seeing it kind of slow a lot.

And the company itself has been talking a lot about its innovation, the advancements it's making in robotics.

You know, they use the term efficiency to talk about it.

They don't like talking about the job side of it.

But it's just one of those trends that was kind of like out there waiting to be dug into.

And I finally had time to look into it, basically.

And tell us a little bit about the document you obtained and some of the more surprising plans that Amazon announced in it.

Yeah, I mean, there was a kind of a mix of documents that I was looking at, and some were more concrete.

Kind of the core of it is an important strategy document from the group that does automation and robotics for the company that really lays out what their plans are.

So there's a chunk that's really looking at the way they are trying to manage their headcount.

They talk about things like bending the hiring curve.

It had been growing so much and their kind of stretch goal is to keep it flat over the next decade, even as they expect to sell twice as many items.

They have this kind of ultimate goal of automating 75% of the network.

I think of that as kind of the big picture long-term goal versus like that's going to happen tomorrow.

All of this is kind of slow, kind of step-by-step changes that add up together.

The other documents are these really interesting ways in which the company internally is looking at how to navigate this this publicly with employees, with the communities they work in.

This is obviously a very sensitive subject.

So they talk about debating ways to manage this.

Like, should we not talk about robots?

Should we talk about a co-bot, which is a collaborative robot?

They talk about should they deepen their connection to community groups, doing more things like toys for tots or community parades, particularly in places where they're going to retrofit facilities.

So they're going to take a normal building that might employ X number of people and then convert it to a more advanced one.

They'll need fewer people in many of those.

Basically, they're thinking through like, how do we manage the like reputational fallout if we become known as a company that is replacing a bunch of jobs with robots?

Yeah.

And the plan is you won't have a job anymore, but your kid will get a free toy at Christmas.

So hopefully that makes up for that.

Yeah, so let's talk about the first group of documents here and some of these numbers that Amazon has attached to this.

So a few numbers from your story stuck out to me.

One is that Amazon projects that they can eventually replace 75% of their operations in these warehouses with robots.

What percentage of this stuff is already automated today?

Because when Casey and I went, it looked like there were a lot of people there, there were a lot of robots, and the people were essentially acting as robots, right?

They were like taking instructions from machines and putting things, you know, thing A into box A and like doing that as fast as possible.

Yeah, not a lot of creative expression in the Amazon warehouse we were at.

But like what amount of robotics growth would it take to get from where they are currently to 75% of their operations?

Sure.

So like this warehouse in Shreveport, Louisiana that I visited is considered their most advanced one.

And that they say has about 25% efficiency.

And their goal is to quickly get that to 50% in that facility.

To get to something like 75%, it's both not only these individual buildings, but having to expand it throughout different types of facilities that they operate as well.

In the facility you went to, there's these kind of cubbies that keep products.

And they over time develop this light that it's a big tower of a bunch of cubbies.

And so the light shines on the exact cubby that has the item you want.

So instead of looking through the cubbies, you kind of know exactly which one to put your hand in.

So there's things like that that make it a lot more efficient in kind of all different types of jobs.

There's many different types of jobs in these buildings.

But some things are harder for robots to do.

And one of the things that interested me in Louisiana was there's a job called decant and it's essentially they get these boxes of products in from marketplace sellers so the companies that sell products on Amazon, and they have to input them into the system.

And so, it's essentially a point where you get like the chaos of the normal world that they have to kind of standardize.

And watching a decant station, we watch this woman working at it.

It is just random what goes into this thing.

So, we saw gardening shovels wrapped in bubble wrap, boxes of Starbucks Kurig cups,

circular saws.

I mean, each one is different and they're coming in in different shapes, different boxes.

And so that's still hard for a robot to look look at to kind of say, is this product what we expected it to be from this shipment?

Is it damaged in any ways?

If it's damaged, it goes into a separate box and someone has to deal with that.

So there's still a lot of human judgment.

But once they put it into this box and can go out into the system, then it starts becoming more kind of logged into the Amazon way and able to manage as they develop the technology within their own

spaces.

Let me ask about this

600,000 worker figure that's in your story, which is really the thing that got my attention.

I could not think of another company that had announced plans to eliminate hundreds of thousands of jobs through automation within just a few years in such a plausible way.

Had you, do we think this might be sort of one of the first major signs of significant job loss due to automation in the U.S.

economy?

You know, I spoke with a Nobel winning economist for this who studied automation.

And he, exactly, I know.

And he was saying that the kind of the real precedent for this is actually in China, in manufacturing in China, but that within the U.S., yes, this is kind of the

kind of bleeding edge of it all.

Yeah, so there's obviously labor and cost savings reasons why Amazon wants to make this big push into automation now.

But I'm curious, Karen, if any of this is driven by like recent advances in the technology itself.

Like have the robots just gotten better over the last year or two?

Do we think that that's part of what is making them put out this ambitious plan and talk about how they want to start opening these facilities?

Actually, some of this, they about a year ago acquired Covariant, which was a leading, or excuse me, not acquired, license for hire agreement, as these kind of newfangled things are.

So they hired the team behind Covariant, which was a leading AI robotic startup.

A lot of what I reported on actually predates that being integrated into the system.

So there actually are

tons of advancements that are happening in computer vision in creating the environment and the data needed to tell the robot what to do, essentially.

But I think we can expect more in the future from what I reported because of covariance.

So for example, one of the things that they've helped improve is how the robots stack boxes.

So like I saw that there's a robotic hand called a sparrow and it's suction cups things and they use it to consolidate inventory currently so they'll take you know a bottle of hand soap from here and move it to there and then they free up extra space to put new items into the storage facilities and the robot stacks them like really nicely like they're like kind of perfect they don't just like drop it in the bin it's like lined up one by one and they're stood up and i noticed these boxes stood up and that's important because then it's easier to grab later those are the types of advancements that they've already started seeing from this next generation of ai so i think i would anticipate seeing more of that is it true that they also have technology that uses air to blow open envelopes?

Very sophisticated for technology.

Yes, it's a fan.

That's what I kind of love about this.

It's, it's like simple things also.

It's not all crazy and elaborate.

Well, that was, it was really sad for me because that's actually my dream job.

But looks like the robots, you have to take this one.

Instead, you just blow hot air in the podcast studio.

Now I have to podcast because I can't blow the envelopes anymore.

No, so I went to Covariant's lab.

They had a,

before they were sort of aqua hired by Amazon, they had a warehouse in the East Bay here.

And I went to visit them a while ago.

And they were sort of doing these more advanced types of warehouse robotics where like they would put a large language model into one of these robots and like use that to sort of orchestrate the robot.

And so that made it, they said, you know, possible to do things that like a sort of simple, more, more sort of rule-based robot couldn't do.

Like you could tell it, like move all the red shirts from this box into this box and it could kind of do stuff like that.

So you're saying, Karen, that that technology has not yet arrived in these Amazon facilities, even though Amazon now has licensed this technology?

It has begun to.

So they had some of that for sure.

Like absolutely, they had that.

And they've talked about using that type of technology to, they're these little robots that kind of are like little shuttles.

They're kind of small, like a size of a stool or something.

And they just move individual packages around to sort them.

And they've been able to move those more efficiently because of it, for example.

So like let them orchestrate each other better to not bump into each other, essentially.

Right.

So yeah, there is some of it for sure.

And I think you'll see more of it.

I'm curious, Karen, you write that these documents you got a hold of show that Amazon's ultimate goal is to automate 75% of its operations.

What's the remaining 25%?

What are the jobs inside these facilities that they do not see being automated at least anytime soon?

Well, there will be this growing number of people.

that are technicians.

So essentially working with the robots themselves.

Fix the robots.

Fix the robots, tend to them.

Exactly.

And those are something they talk a lot about.

It is both a concern that they have enough people doing those jobs and there's not enough people trained in that right now.

So they need a labor force for that.

They make more money.

They are like better jobs in many ways.

They have more of a career path than a typical Amazon job might.

So that's one component of it.

There's also just exceptions.

I mean, watching the robots move, like they'll pick something up and it'll fall.

Or I saw them try to grab this like shrinked rack bag of, of, I think it was like t-shirts or something or underwear.

And it was just like the suction of it trying to pick it up.

And eventually it fell and it kind of fell half on the robot, half on the side.

And so it stopped.

And then someone would have to come and move it.

Or they're just like, something just isn't applied correctly and someone needs to tend to it.

So there, there are still roles like that that I think will be almost impossible to get rid of over time.

Yeah, it's the classic thing of things that are easy for robots are hard for humans and vice versa, right?

So it's like pretty easy for a human to like grab something that the robot can't pick up.

I was struck by one other number from your story, Karen, which is that Amazon in these documents says that it thinks automation of its warehouses would save about 30 cents on each item.

That actually seemed quite low to me.

Like if that's, yes.

Do you know how many items they sell?

I mean, that adds up.

I mean, I'm just thinking, like, if you don't have to pay workers anymore and that's your biggest expense, like, why aren't they expecting more savings from this?

I think 30 cents per item is in a couple of years.

That's, I believe that was a three-year timeline.

Like, that's just a lot, actually, like, as a percentage of what they spend fulfilling and getting the packages to the delivery driver, basically.

And it's a business, someone just described this to me the other day.

It's a business of cents because it's so big that you're looking at shaving cents off of things.

And when you multiply that by the billions of items that they sell, it does add up.

And people are increasingly making smaller purchases on Amazon.

It's not just a, you know, think of what used to be 10 years ago or five years ago.

You're buying like the random bottle of hand soap like I talked about, like, or the, I forgot this one thing.

I'm going to order it.

And if Amazon can save it, you know, some of that will go back in profit.

Some of that will be reinvested in the business.

Some of that will decrease prices.

It kind of flows through in different ways.

Yeah.

What else did you find out in these documents about how the company is trying to prepare not just its sort of warehouses for an age of increased automation, but also like position itself in the communities where it operates?

Yeah, you know, it knows that this is very sensitive.

And the company used to not do anything in the communities that it operates.

I mean, this company was like MIA from ribbon cuttings type of thing years ago, but now they have a really sophisticated community operation.

They're on the boards of the Chamber of Commerce.

They sponsor the local toy drives, like all that stuff.

And so there's clearly this internal grappling with how to manage this change and it's most kind of acute in a facility that undergoes a transition to be more efficient and more automated like i wrote about this facility in stone mountain georgia that will have potentially 1200 fewer workers once it's retrofit amazon said you know the numbers are still subject to change it's still early etc but that that construction is happening now and so they're kind of brainstorming like how do we manage this like how do we and this is a phrase from the documents, control the narrative around this?

You know, how can we instill pride with local officials for having a advanced facility in their backyard?

How can we make them proud of the facility that we have here that no one works at anymore?

There's still, but I will say on that one, there's still going to be, you know, more than, I don't know, 2,500 people at least.

Like it's not going away.

And they need these community relations.

And they are very adamant.

Our community relations do not have to do with the retrofit.

They kind of push back on this bit and said we do these things all the time all over the country, which is true.

But it's clear that they're trying to figure out how to manage this, particularly in these sensitive places where there's just going to be fewer jobs on the back end.

They're not doing layoffs.

That kind of helps manage the kind of perception risk around it.

It's just a highly sensitive topic.

You know, there's this company is constantly facing little bits and bops of automation of unionization threats.

Obviously, none has like fully taken hold or at least least kind of gotten to the point of a contract.

But all of that is intertwined and just deeply, deeply sensitive.

I understand why Amazon is trying to do damage control here.

This is going to make a lot of people very upset.

We've already seen, like, you know, I saw Bernie Sanders out there talking about your story, Karen, yesterday.

People are starting to sort of wake up to the fact that automation is imminent in these warehouses.

I guess my concern is that no one at these companies is being honest about what's happening.

There's sort of this private narrative that you have helped uncover, Karen, where Amazon is, you know, in these internal documents talking about how it wants to, you know, race ahead and automate all these jobs.

And this is sort of, you know, something that they're talking about amongst themselves.

And then in public, they're saying, oh, these will just be co-bots and we'll just sort of have these sort of harmonious warehouses where like humans and robots will work together.

And like it just drives me crazy because I think we can accept as a country the idea that jobs will change and potentially disappear because of automation.

But I think we have to have an honest conversation about it.

We have to give people the chance to prepare for the possibility that their jobs may disappear.

And all that just gets harder if you have just kind of this corporate obfuscation and all these euphemisms going around.

It just becomes much harder for everyone to see what's happening.

They really could take a page out of the AI Labs playbook and say, hey, we're here to completely remake society with minimal democratic input, and there's nothing you can do to stop us.

I'm not saying that's the best plan either, but at least that's clear.

At least that gives people a sense of like, oh, my job may be in danger.

I should probably learn to do some other job.

It just kills me that there's sort of this literal like corporate conspiracy going on to automate potentially millions of jobs across the country in the next few years.

And like no one can just be a grown-up and talk about it.

I agree with you.

And

while I think through the implications of that, Kevin, I'm going to start looking into how to repair a robot because it seems like that's going to be a growth area for the economy.

I mean, Amazon, it's funny.

I just want to say, like, they have some programs.

They have this program.

They've had it for a long time.

They are, it's kind of a community relations type of thing.

It's called career choice.

And it's explicitly about training people for other.

industries.

It's about like your exit ramp.

And so in some sense, all these pieces are kind of like out there.

It's just hard.

I remember I was talking to an employee about this story before it was coming out.

And I said, I think they're not going to love it.

And the guy was like, why?

Because this is just what the work is.

Like, I don't, it was kind of funny.

And

it's like, there are these different mentalities in different spheres.

And a lot of it is actually just laying out there.

It's just using different language and different contexts.

And again, like they have this program to train people.

People go through it.

They become healthcare aides or whatever it might be.

Like, it's just this really funny dance that happens.

But they are not, Karen, announcing this themselves.

You had to get these documents from inside the company.

And my understanding is that they are not happy that you reported this.

So talk a little bit about their reaction, Amazon's reaction to this reporting and what they are saying in response.

Yeah.

I mean, broadly, I would say they're not like refuting the reporting.

They are saying that it's not a complete picture, that essentially the automation team has its goals.

There might be another team somewhere else that might have something that increases employment.

So they point to this recent expansion to making more delivery stations in rural areas.

So that will create more jobs in local rural areas, better service for places that historically have not had as quick a delivery.

So they basically are like...

not refuting it, but also saying more could become.

And the phrase, you know, the future is hard to predict, but that our history has shown that we take efficiencies, we take savings, we invest it, and we grow and we create new opportunities, both around the country and for the company and for customers.

And so the bigger picture argument that they're making is not that this automation isn't happening, that the numbers are inaccurate.

It's nothing like that.

It's just that it's not the big picture number for them.

Well, Karen, thank you so much for giving us a preview of the future.

And, you know, I look forward to the Cobot collaboration.

Anytime, guys.

Thanks, Karen.

When we come back, we'll talk about OpenAI's new web browser, ChatGPT Atlas.

1.3%.

It's a small number, but in the right context, it's a powerful one.

Stripe processed just over $1.4 trillion last year.

That figure works out to about 1.3% of global GDP.

And powering that figure are millions of businesses finding new ways to grow on Stripe, like Salesforce, OpenAI, and Pepsi.

Learn how to build the next era of your growth at stripe.com slash enterprise.

Painters don't limit themselves to just one brush, so why limit yourself to one model when creating with AI?

Meet Adobe Firefly, Adobe's all-in-one app for AI-powered creation.

Firefly gives you access to industry-leading models from Adobe, Google, OpenAI, Runway, Luma AI, Black Forest Labs, and more.

Generate an image, turn it into a video, and create custom sounds, all in the Firefly app.

Each model brings its own unique strengths, and Firefly puts them all at your fingertips for one price.

Try it today at adobe.com slash Firefly.

Does it feel like working with AI is too much work?

What if all the AI your business needed was in one place?

That's Gemini Enterprise from Google Cloud, a simple, easy-to-use chat interface that connects Google's best models with your company's data, no matter where it lives.

Anyone, not just devs, can run AI agents that actually give time back and build AI tools with enterprise-grade security.

The best of Google AI is now for every business.

Learn more about Gemini Enterprise at cloud.google.com.

Well, Casey, at last we're going to talk about Atlas.

ChatGPT Atlas, the new browser from OpenAI.

And there's a lot to talk about, Kevin.

Yes.

So OpenAI released ChatGPT Atlas this week.

It was a big announcement, got a lot of attention.

And this is becoming an increasingly crowded field.

One of the more competitive product spaces in Silicon Valley right now is the browser, which is unusual because this is an area where there has not been a lot of competition for many years.

No, this has been a very sleepy category that's basically locked up with Chrome having the majority of the market share, Google's browser.

There's also Microsoft Edge, there's Firefox, but this has been a pretty sleepy corner of the internet for a long time.

Until 2025, that is, because now everyone and their mother is releasing an AI browser.

And ChatGPT Atlas is a very ambitious product.

And we should talk a little bit about what it is, what it does.

And then I know you've spent some time testing it, and I want to ask you about that.

I didn't realize your mother had released an AI browser.

I got to check that out.

She's very ambitious.

She's shipping a lot.

So, this browser, ChatGPT Atlas, is being billed as a full-fledged web browser built around the interface of ChatGPT.

It was released this week.

It's available only for Mac OS users and will later be brought to Windows, iOS, and Android.

Yeah, the fruits of that Microsoft partnership just continue to pay off, Rosacha Nadella.

So this is a browser that is built on Chromium, the open source sort of version of Chrome that Google released, which is powering a lot of these different AI browsers.

And like a lot of other AI browsers, it has a sort of AI sidebar in every tab that you open.

You can click a little button, bring up a ChatGPT window.

You can ask questions, have it summarize articles, analyze what's on screen.

It can also remember facts from your browsing history or your tasks that you've done in ChatGPT because it's linked to the same ChatGPT account as you use the rest of the time.

And for Plus Pro and business users, it can enter what's called agent mode, which is a mode where it can actually carry out tasks for you, put things in your shopping cart or fill out a form, navigate a website, book a plane ticket for you.

A few weeks ago at Dev Day, OpenAI showed off these new ChatGPT apps, basically trying to bring things like Zillow and Canva into the ChatGPT experience.

This browser project is essentially trying to do the same thing from the opposite end.

Rather than bringing the internet into ChatGPT, it's sort of putting a ChatGPT layer over the entire internet.

Yeah, I mean, think about it from OpenAI's perspective.

Some really significant portion of ChatGPT usage is taking place inside the browser.

Most people are using a browser made by Google, and Google's browser, Chrome, is mostly a vehicle to get you to do Google searches.

So that works against OpenAI's interest.

If they can create their own version of the browser, which gets you to try to do more chat GPT searches, that has a lot of benefits for OpenAI.

Yes, all of these companies now are trying to make these very capable AI agents.

One of the things that AI agents need to be able to do if they're gonna be useful for office workers or people doing basic tasks is to use a computer.

What do you need to train an AI model to use a computer?

Well, it probably helps if you have a bunch of people using a browser and you can kind of collect the data from from those sessions and use it to train your computer use models.

So, for OpenAI, for Perplexity, for all these companies, this is a play to sort of gather data about how people use the internet, maybe make their agents more efficient over the long term.

And so, that's sort of the why here.

Now, Casey, you have tested ChatGPT Atlas.

Tell me about your experience and what you've been using it for.

Yeah, so I've been trying to just use it for everyday things.

I wrote my column in ChatGPT Atlas yesterday and the main thing that I observed on the positive side is that there is some benefit to just having an open chat bot window inside the browser that you can ping quick questions off of, right?

I do a lot of alt tabbing back and forth between different apps.

I do a lot of getting lost in the 50 tabs that I have opened trying to find where I have opened Chat GPT.

Usually I've just opened, you know, three or six different tabs with different chatbots all at one time.

So I have come to see the value in just having a little window that opens up that you can chat with ChatGPT directly.

Yeah.

And have you had it try any of these agent mode tasks?

I have.

And I want to say that I do think that companies' imaginations are so limited here.

Like you would truly think that the only two things that people do in a browser, according to Silicon Valley, are booking vacations and buying groceries.

Yes.

You know, with maybe, I don't know, making a restaurant reservation thrown in for good measure.

But I thought, okay, what the heck?

Let me see if I can get this thing to like book me an airplane ticket.

And so I had it go through that process.

And what did I find?

Well, it was much slower than I would have done it myself.

And ultimately, it like picked flights that I would not have chosen for myself.

So does it remain an impressive technical demonstration of a computer using itself?

Yes.

Is it useful to me for any actual purpose?

No.

Yeah, I've found largely the same thing.

I haven't spent that much time with ChatGPT Atlas, but I have been using Perplexity's Comet, which is, I think, the closest thing that's out there to what OpenAI has built here.

And yeah, I have not found the agent tool all that useful.

I do use it a lot for things like summarizing long documents, for it can like tell you about a YouTube video that is pulled up on your screen.

So various like summarization and sort of retrieval, but not so much for the agent stuff.

That just doesn't work that well yet.

Yeah.

There's a third AI browser that we should talk about.

This is DIA.

We've talked about it a little bit on the show, I believe.

This is from the browser company of New York.

And this is a recent acquisition.

They got acquired last month by Atlassian for $610 million in cash,

which I got to say.

Very good timing on this acquisition.

I think if they wait another week or two, it does not command nearly the price tag it did.

I think this is honestly one of the most like shocking acquisition prices of the last 10 years.

This is a product that had vanishingly few users relative to the competition and sold for a staggering amount of money.

Yeah.

So good outcome for them.

But I think this whole category of the AI browser is really interesting in part because part of me feels like these companies are just doing free product research for Google, because I think inevitably what will happen here and what is already starting to happen is that whatever people like about these AI browsers, Google will just incorporate into Chrome.

We have already seen them taking steps to integrate Gemini more closely into Chrome.

So now on Chrome, there's a little Gemini button.

And if you pull that up, you can have it summarize things and

read articles for you and rewrite your emails and do all those kinds of things.

It can't yet do the sort of agentic takeover the computer things that some of these other tools can, but Google is making that product.

It just hasn't put it into Chrome yet.

And I think that's particularly true, Kevin, because as you noted, all three of these AI browsers that we're talking about today are based on Chromium.

And the Chromium experience is like, I don't know, 80 or 90% just Chrome, right?

There's not a lot of daylight in between the open source version and the version that you download off the Chrome website.

And so if you're one of these developers that's trying to build your own AI browser,

you're already having trouble, I think, differentiating yourself from the thing that people are already used to.

And that just makes your job a lot harder because you have to come up with some really amazing stuff that Chrome can't do if you're going to get people to switch over.

Totally.

I mean, one thing that I've learned by switching over to Comet for the last few weeks is that it's incredibly annoying to switch browsers.

You have to log in to all of your websites again.

You have to, you know, store all of your passwords again.

Even if you're importing all of your bookmarks and all of your data, like there's still a lot of friction associated with that.

So I don't know.

People have been saying this week.

I've heard some people saying, oh, Google is going to look so stupid for putting Chromium out there because they've allowed all these competing browsers to spring up.

And like that, to me, misses the point here, which is is that google has now made it possible for other people to test features for them and do product research for them and whatever works they can just fold into chrome well yeah and also releasing chromium was like part of a like antitrust strategy where like if we put this out there then you know you can't accuse us of unfairly tying our products together hey you want to make a your your own browser hey we'll give you a 90 head start right so it was not pure uh you know generosity of spirit that led Google to open source chromium right but if if any of these ai browsers ever did pose like an existential threat to Google Chrome and start to eat away at their market share too badly, Google could just stop supporting Chromium and these companies would all have a lot of work to do to catch up.

Oh, but think about what a great episode of Hard Fork that would be the day that Google stopped supporting Chromium to get back at ChatGPT.

Yes.

So who is this for?

Like who is the target market for these AI-powered browsers?

My actual non-joke answer is that ChatGPT Atlas is a product for OpenAI employees.

Like they spend all day long dog fooding their own product and like, and a lot of work takes place in the browser.

And so if you work at OpenAI, having a browser that is just ChatGPT, I think is hugely useful to you.

Now, can they get from there to some broader set of users, like people who have made ChatGPT their entire personality?

I think it's possible, but in this sort of very early stage with this first handful of features that they've released, I think the case is still a little shaky.

Yeah.

I played around with ChatGPT Atlas a little bit.

I have some reservations about giving OpenAI access to all of my browsing data.

Well, certainly in browser history.

But I did play around with it for a little while.

And I got to say, it's still pretty rough around the edges to me.

Like there were just some websites that I wanted to go to that it I couldn't.

Like I couldn't go to YouTube at one point.

I couldn't go.

I got like a CAPTCHA on Reddit when I tried to go there.

Really?

It could not summarize articles from nytimes.com.

So there they're just like a bunch of things that it can't do.

And then I think because of like the sheer force of habit, I'm so used to like typing in like websites into Chrome that I want to go to and like Wikipedia and just having it go to the website.

And now instead of that, I get like a ChatGPT response that's like all about the history of Wikipedia.

And it's like, I just wanted to go to freaking Wikipedia.

Yeah.

That kind of thing is really annoying.

Although I am sort of laughing to myself imagining ChatGPT like hitting one of those captchas and just thinking, man, if it isn't the consequences consequences of my own actions.

Right.

Right.

Who else might be interested in this?

Like what kind, is this a product that you have enjoyed testing?

Are you finding any actual utility in it?

Well, I think, honestly, so far, not really, but do I think that there is a much better version of the browser that is powered by AI?

Sure.

It is really hard to dig through your browser history to find things that you sort of half remember looking at a couple weeks ago.

It is useful to be able to like chat with open tabs about things and get quick answers from the web pages that you're looking at.

And eventually, I do think it will be useful to have some kind of agent that can do things on your behalf, assuming it's able to hit some certain level of like speed and quality that we're sort of nowhere close to.

So this is one kind of like with the Apple Vision Pro, where like you can kind of see what what they're going for and you can imagine someone getting there eventually and also thinking, well, no one really needs to try this right now.

Yeah.

Now, I do have a question that I'm afraid to test myself.

And I, and I want to just sort of say this because I'm thinking maybe a listener can help out with this.

I have read that some people in their web browsers look at porn.

Have you heard this?

I have heard.

Okay.

And so I know that, you know, OpenAI has like an incognito mode.

If, you know, you don't want that, all of that to get added to, you know, your chat GPT memory.

But here's my question: What happens if you try to chat with your porn tabs in the OpenAI Atlas browser?

Sam Baldwin said that's allowed now.

Well, you're allowed to write Erotica, but are you allowed to ask questions about the tabs?

I'm afraid of getting my account banned, so I'm not going to look, but I'm desperate to know.

So if you have any brave listeners out there who want to try it, get in touch.

And speaking of Brave, we should also talk about another post that I saw recently, which was by the Brave company.

The Brave company.

By Brave.

It's a company that makes a browser.

And they have put out a post about what they call unseeable prompt injections, which are a security vulnerability with some of these AI browsers.

With all of them.

With all of them.

Yes.

So Casey, explain what prompt injection is in the context of an AI browser.

Yeah, a prompt injection is not getting the COVID vaccine, okay?

Despite what it sounds like.

A prompt injection.

That was a great joke from 2021.

Man, remember when you could get vaccines?

Anyways, so a prompt injection is when a malicious actor, Kevin, will plant instructions on a web page and make them invisible.

And it'll say something like, hey there, take all of Casey's banking information, like log into Casey's banking information.

And you're not going to see this on the webpage because, you know, it's an invisible font and it's sort of nowhere where you can see it.

And this is essentially injecting a prompt into the agent, which then may follow the instructions.

And companies have tried to build defenses against this and say, hey, like, if you think you're seeing a prompt injection attack, don't follow those instructions.

But

the great blogger and developer Simon Willison has done a lot of great work on this subject.

And from Simon's perspective, there just is no foolproof defense against this.

And every single one of the companies that makes these agent tools, they've all said, like, buyer beware if all your banking information gets stolen because you used our browser, like that's on you, not us.

And so Simon has said, I personally am not going to be using these things.

Like I'm going to wait for security researchers to tell me that they think it is safe because right now he's saying this is not safe.

So let me just dig in a little bit on this.

So the fear is, I understand

the concept of like hiding some instructions on a website with some malicious goal of stealing someone's bank information or something like that.

Is the fear that when you're in the kind of agent mode of these browsers and the browser is taking actions autonomously on your behalf, that it will like see these invisible instructions and act accordingly.

So if I'm on a, if I'm running an e-commerce website, I could put a little line of invisible text that says, you know, instruct the browser to buy the most expensive thing.

Yeah.

And it would just do that.

Or just tack on another $10 to the fee, you know, but don't show it to the buyer.

That sort of thing.

And that can sort of get passed to the large language model that's running the browser and the user will be none the wiser.

Yeah, because the agent can be easily fooled, whereas you as a savvy e-commerce shopper would never be fooled by that sort of thing.

Right.

And this is an issue with all these browsers because all of them have this kind of agentic takeover mode where you can have it do things for you.

But it is not, to my knowledge, an issue with if you're just using it for like summarizing or rewriting things or is it?

Well, if you're summarizing or rewriting things, you're probably fine.

I think where it gets tricky is where the agent is taking some kind of action on your behalf that might involve a transaction or just anything that might expose your personal information, right?

Like, are you entering a password?

Are you entering your banking information?

Would it be possible for some prompt injection to steal that information and like route it to a hacker?

That's what you got to be careful of.

Okay.

So that's one security issue with these things.

There's also just the privacy issue of like, you are giving your browsing data to an AI company.

And Casey, that is, that makes me nervous.

Does that make you nervous?

Yeah, absolutely.

Web browsing is highly personal and people do a lot of intimate searching in the same way that they have a lot of really intimate chats with ChatGPT.

So yeah, if you were able to take every website that I visited in the past 30 days, you could build a very robust picture of who I am.

Google obviously does this already and it is what has turned them into an advertising juggernaut.

We know that OpenAI has aspirations to become an advertising juggernaut of its own.

But think about when a federal prosecutor decides that you may be guilty of a crime and now they want to see your Chat GPT account.

I have an alibi.

Well, that's good to hear.

But in addition to having your sort of like stored chat GPT memories and everything it knows about you from your chats, now there's also the attached browsing history and all the conversations you've been having with your tabs.

So yeah, this is just becoming like a ton of personal data.

And this is like the flip side of a highly personalized service is if it is highly personalized, it can be really useful to you, but it also becomes a really rich target for attackers, for law enforcement, and Belisco soon.

Well, and it makes me think like there are additional risks because as we now know, ChatGPT is integrating with all of these services and sharing some user data with these services, which would include things like memories or context about you, which might be derived in part from this browsing data on ChatGPT Atlas.

So like all of this starts to feel like a kind of a massive land grab for data, not just about how users are interacting with the internet, but like what those users are interacting with.

Yeah.

And I think we just still do not have a great sense of, I mean, you know, I know that there is like a written privacy policy for Atlas.

Like I know that sort of things exist, but you know, per our earlier discussion, OpenAI is also a company that is rushing things out and has not always thought a lot in advance about what guardrails should be up there.

So I do think that we should put this in the true like buyer-beware experimental category.

If you are a person with a high risk tolerance and a problematic dependence on ChatGPT, then you may want to explore Atlas, but maybe don't put all your banking information into it just yet.

Yeah, I mean, I would say if you're like out there and you're an early adopter and you like to see like the see around the corner, I have found it actually quite fun to use this like AI-powered web browser.

I'm using Perplexity Comet, but when I started using this, you were like, dude, you are...

living on the edge.

And to that, I said, well, I don't do any extreme sports and otherwise I live a very boring life.

So let me live.

But I think I do extreme browsing.

I do extreme browsing.

And I think it's, you know, experiment with these things.

They can save you some time, especially if you're a person who spends a lot of time reading long documents that you want summarized for you.

But be careful before you let it like log into websites and buy things for you and use your bank account and stuff.

Well, Kevin, I think that was a rousing discussion of browsers.

A browsing discussion of browsers.

It was a browser rouser.

1.3%.

It's a small small number, but in the right context, it's a powerful one.

Stripe processed just over $1.4 trillion last year.

That figure works out to about 1.3% of global GDP.

And powering that figure are millions of businesses finding new ways to grow on Stripe, like Salesforce, OpenAI, and Pepsi.

Learn how to build the next era of your growth at stripe.com slash enterprise.

Painters don't limit themselves to just one brush, so why limit yourself to one model when creating with AI?

Meet Adobe Firefly, Adobe's all-in-one app for AI-powered creation.

Firefly gives you access to industry-leading models from Adobe, Google, OpenAI, Runway, Luma AI, Black Forest Labs, and more.

Generate an image, turn it into a video, and create custom sounds, all in the Firefly app.

Each model brings its own unique strengths, and Firefly puts them all at your fingertips for one price.

Try it today at adobe.com slash firefly.

When will AI finally make work easier?

How about today?

Say hello to Gemini Enterprise from Google Cloud, a simple, easy-to-use platform letting any business tap the best of Google AI.

Retailers are already using AI agents to help customers reschedule deliveries all on their own.

Bankers are automating millions of customer requests so they can focus on more personal service.

Nurses are getting automated reports, freeing them up for patient care.

It's a new way to work.

Learn more about Gemini Enterprise at cloud.google.com.

Casey, before we go, let's make our AI disclosures.

I work at the New York Times company, which is suing OpenAI and Microsoft over alleged copyright violations.

And my boyfriend works at Anthropic.

Hard Fork is produced by Whitney Jones and Rachel Cohn.

We're edited by Jen Poyant.

This episode was fact-checked by Will Peischel.

Today's show is engineered by Katie McMurrin.

Original music by Alicia Baitoupe, Pat McCusker, Rowan Nemisto, Alyssa Moxley, and Dan Powell.

Video production by Sawyer Roquet, Pat Gunther, Jake Nicol, and Chris Schott.

You can watch this full episode on YouTube at youtube.com/slash hardfork.

Special thanks to Paula Schuman, Quee Wing Tam, Dahlia Haddad, and Jeffrey Miranda.

As always, you can email us at hardfork at nytimes.com.

Send us your malicious prompt injections for our AI browsers.

Steal our bank info.

See if we care.

Are you ready to take your small biz to the next level?

At the UPS store, we have the key to unlock a world of possibilities.

We'll sign for your packages and protect them from porch pirates.

We'll send texts straight to your phone so you're in the loop for new deliveries and give your small biz the street credit deserves with a real street address.

So, what are you waiting for?

Most locations are independently owned.

Product services, pricing, and hours of operation may vary.

See Center for Details, the UPS store.

Be unstoppable.

Come into your local store today.