Google: The AI Company

4h 6m

Google faces the greatest innovator's dilemma in history. They invented the Transformer — the breakthrough technology powering every modern AI system from ChatGPT to Claude (and, of course, Gemini). They employed nearly all the top AI talent: Ilya Sutskever, Geoff Hinton, Demis Hassabis, Dario Amodei — more or less everyone who leads modern AI worked at Google circa 2014. They built the best dedicated AI infrastructure (TPUs!) and deployed AI at massive scale years before anyone else. And yet... the launch of ChatGPT in November 2022 caught them completely flat-footed. How on earth did the greatest business in history wind up playing catch-up to a nonprofit-turned-startup?

Today we tell the complete story of Google's 20+ year AI journey: from their first tiny language model in 2001 through the creation Google Brain, the birth of the transformer, the talent exodus to OpenAI (sparked by Elon Musk's fury over Google’s DeepMind acquisition), and their current all-hands-on-deck response with Gemini. And oh yeah — a little business called Waymo that went from crazy moonshot idea to doing more rides than Lyft in San Francisco, potentially building another Google-sized business within Google. This is the story of how the world's greatest business faces its greatest test: can they disrupt themselves without losing their $140B annual profit-generating machine in Search?

Sponsors:

Many thanks to our fantastic Fall ‘25 Season partners:

Acquired’s 10th Anniversary Celebration!

Links:

Carve Outs:

More Acquired:

‍Note: Acquired hosts and guests may hold assets discussed in this episode. This podcast is not investment advice, and is intended for informational and entertainment purposes only. You should do your own research and make your own independent decisions when considering any financial transactions.

Listen and follow along

Transcript

I went and looked at a studio.

Well, a little office that I was going to turn into a studio nearby, but it was not good at all.

It had dropped ceilings, so I could hear the guy in the office next to me.

You would be able to hear him talking on episodes.

Third co-host.

Third co-host.

Is it Howard?

No, it was like a lawyer.

They seemed to be like talking through some horrible problem that I didn't want to listen to, but I could hear every word.

Does he want millions of people listening to his conversations?

Right.

All right.

All right.

Let's do a podcast.

Let's do a podcast.

Welcome to the fall 2025 season of Acquired, the podcast about great companies and the stories and playbooks behind them.

I'm Ben Gilbert.

I'm David Rosenthal.

And we are your hosts.

Here's a dilemma.

Imagine you have a profitable business.

You make giant margins on every single unit you sell.

And the market you compete in is also giant, one of the largest in the world, you might say.

But then on top of that, lucky for you, you also are a monopoly in that giant market with 90% share and a lot of lock-in.

And when you say monopoly, monopoly as defined by the US government.

That is correct.

But then, imagine this.

In your research lab, your brilliant scientists come up with an invention.

This particular invention, when combined with a whole bunch of your old inventions by all your other brilliant scientists, turns out to create the product that is much better for most purposes than your current product.

So you launch the new product based on this new invention, right?

Right.

I mean, especially because out of pure benevolence, your scientists had published research papers about how awesome the new invention is and and lots of the inventions before also.

So now there's new startup competitors quickly commercializing that invention.

So of course, David, you change your whole product to be based on a new thing, right?

This sounds like a movie.

Yes, but here is the problem.

You haven't figured out how to make this new incredible product anywhere near as profitable as your old giant cash printing business.

So maybe you shouldn't launch that new product.

David, this sounds like quite the dilemma to me.

Of course, listeners, this is Google today and in perhaps the most classic textbook case of the innovator's dilemma ever.

The entire AI revolution that we are in right now is predicated by the invention of the transformer out of the Google brain team in 2017.

So think OpenAI and ChatGPT, Anthropic, Nvidia, hitting all-time highs.

All the craziness right now depends on that one research paper published by Google in 2017.

And consider this.

Not only did Google have the densest concentration of AI talent in the world 10 years ago that led to this breakthrough, but today they have just about the best collection of assets that you could possibly ask for.

They've got a top-tier AI model with Gemini.

They don't rely on some public cloud to host their model.

They have their own in Google Cloud that now does $50 billion in revenue.

That is real scale.

They're a chip company with their tensor processing units or tpus which is the only real scale deployment of ai chips in the world besides nvidia gpus

maybe amd maybe but these are definitely the top two somebody put it to me in research that if you don't have a foundational frontier model or you don't have an ai chip

you might just be a commodity in the ai market and google is the only company that has both google still has a crazy bench of talent And despite ChatGPT becoming kind of the Kleenex of the era, Google does still own the text box, the single one that is the front door to the internet for the vast majority of people, anytime anyone has intent to do anything online.

But the question remains, what should Google do strategically?

Should they risk it all and lean into their birthright to win in artificial intelligence?

Or will protecting their gobs of profits from search hamstring them as the AI wave passes passes them by.

But perhaps first, we must answer the question, how did Google get here, David Rosenthal?

So listeners, today we tell the story of Google, the AI company.

Woo.

You like that, David?

I love it.

I love it.

Did you hire like a Hollywood scriptwriting consultant without telling me?

I wrote that 100% myself with no AI.

Thank you very much.

No AI.

Listeners, if you want to know every time an episode drops, vote on future episode topics or get access to corrections from past episodes.

Check out our email list.

That's acquired.fm slash email.

Come talk about this episode with the entire acquired community in Slack after you listen.

That's acquired.fm slash Slack.

Speaking of the acquired community, we have an anniversary celebration coming up.

We do.

10 years of the show.

We're going to do an open Zoom call with everyone to celebrate.

kind of like how we used to do our LP calls back in the day with LPs.

And we are going to do that on October 20th, 2025 at 4 p.m.

Pacific time.

Check out the show notes for more details.

If you want more acquired, check out our interview show, ACQ2.

Our last interview was super fun.

We sat down with Toby Lutke, the founder and CEO of Shopify, about how AI has changed his life and where he thinks it will go from here.

So search ACQ2 and any podcast player.

And before we dive in, we want to briefly thank our presenting partner, JPMorgan Payments.

Yes, just like how we say every company has a story, every company's story is powered by payments, and JP Morgan Payments is a part of so many of their journeys from seed to IPO and beyond.

So, with that, this show is not investment advice.

David and I may have investments in the companies we discuss, and this show is for informational and entertainment purposes only.

David, Google, the AI company.

So, Ben, as you were alluding to in that fantastic intro, really, you're really up in the game again.

If we rewind 10 years ago from today,

before the Transformer paper comes out,

all of the following people,

as we've talked about before, were Google employees.

Ilya Sitskeever, founding chief scientist of OpenAI, who along with Jeff Hinton and Alex Koszewski had done the seminal AI work on AlexNet and just published that a few years before.

All three of them were Google employees, as was Dario Amoudei, the founder of Anthropic,

Andrei Karpathy, chief scientist at Tesla until recently, Andrew Ng, Sebastian Thrun, Noam Shazir, all the DeepMind folks, Demis Asabis, Shane Legge, Mustafa Suleiman.

Mustafa now, in addition to in the past, having been a founder of DeepMind, runs AI at Microsoft.

Basically,

every single person of note in AI worked at Google, with the one exception of Jan Lacun, who worked at Facebook.

Yeah.

It's pretty difficult to trace a big AI lab now back and not find Google in its origin story.

Yeah.

I mean, the analogy here is it's almost as if at the dawn of the computer era itself, a single company like say IBM had hired every single person who knows how to code.

So it'd be like, you know, if anybody else wants to write a computer program, oh, sorry, you can't do that.

Anybody who knows how to program works at IBM.

This is how it was with AI and Google in the mid-2010s.

But learning how to program a computer wasn't so hard that people out there couldn't learn how to do it.

Learning how to be an AI researcher, significantly more difficult.

Right.

It was the stuff of very specific PhD programs with a very limited set of advisors and a lot of infighting in the field of where the direction of the field was going, what was legitimate versus what was crazy, heretical, religious stuff.

Yep.

So then, yes, the question is, how do we get to this point?

Well, it goes back to...

the start of the company.

I mean, Larry Page always thought of Google as an artificial intelligence company.

And in fact, Larry Page's dad was a computer science professor and had done his PhD at the University of Michigan in machine learning and artificial intelligence, which was not a popular field in computer science back then.

Yeah.

In fact, a lot of people thought specializing in AI was a waste of time because so many of the big theories from 30 years prior to that had been kind of disproven at that point, or at least people thought they were disproven.

And so it was frankly contrarian for Larry's dad to spend his life and career and research work in AI.

And that rubbed off on Larry.

I mean, if you squint, PageRank, the PageRank algorithm that Google was founded upon, is a statistical method.

You could classify it as part of AI within computer science.

And Larry, of course, was always dreaming much, much bigger.

I mean, there's the quote that we've said before on this show in the year 2000, two years after Google's founding, when Larry says artificial intelligence would be the ultimate version of Google.

If we had the ultimate search engine, it would understand everything on the web.

It would understand exactly what you wanted, and it would give you the right thing.

That's obviously artificial intelligence.

We're nowhere near doing that now.

However, we can get incrementally closer, and that is basically what we work on here.

It's always been an AI company.

Yep.

And that was in 2000.

Well, one day, in either late 2000 or early 2001, the timelines are a bit hazy here, a Google engineer named Georges Herrick is talking over lunch with Ben Gomes, famous Google engineer, who I think would go on to lead search, and a relatively new engineering hire named Noam Shazir.

Now, Georges was one of Google's first 10 employees, incredible engineer, and just like Larry Page's dad, he had a PhD in machine learning from the University of Michigan.

And even when Georges went there, it was still a relatively rare contrarian subfield within computer science.

So the three of of them are having lunch, and George says offhandedly to the group that he has a theory from his time as a PhD student that compressing data is actually

technically equivalent to understanding it.

And the thought process is if you can take a given piece of information and make it smaller, store it away, and then later reinstantiate it in its original form, the only way that you could possibly do that is if if whatever force is acting on the data actually understands what it means because you're losing information, going down to something smaller, and then recreating the original thing.

It's like you're a kid in school.

You learn something in school, you read a long textbook, you store the information in your memory, then you take a test to see if you really understood the material.

And if you can recreate the concepts, then you really understand it.

Which kind of foreshadows big LLMs today are like compressing the entire world's knowledge into some number of terabytes.

That's just like the smashdown little vector set, little, at least compared to all the information in the world.

But it's kind of that idea, right?

You can store all the world's information in an AI model in something that is like kind of incomprehensible and hard to understand.

But then if you uncompress it, you can kind of bring knowledge back to its original form.

Yep.

And these models demonstrate understanding, right?

Do they?

That's the question.

And that's the question.

They certainly mimic understanding.

So this conversation is happening.

You know, this is 25 years ago.

And Noam, the new hire, the young buck, he sort of stops in his tracks and he's like, wow, if that's true, that's really profound.

Is this in one of Google's micro kitchens?

This is in one of Google's micro kitchens.

They're having lunch.

Where did you find this, by the way, a 25-year-old?

This is in In the Plex.

This is like a small little passage in Stephen Levy's great book that's been a source for all of our Google episodes, In the Plex.

This is a small little throwaway passage in here about this because this book came out before ChatGPT and AI and all that.

So Gnome kind of latches on to Georges and keeps vibing over this idea.

And over the next couple of months, the two of them decide in the most googly fashion possible that they are just going to stop working on everything else and they're going to go work on this idea on language models and compressing data and can they generate machine understanding with data.

And if they can do that, that that would be good for Google.

I think this coincides with that period in 2001 when Larry Page fired all the managers in the engineering organization.

And so everybody was just doing whatever they wanted to do.

Funny.

So there's this great quote from Georges in the book.

A large number of people thought it was a really bad thing for Noam and I to spend our talents on, but Sanjay Gemawat, Sanjay, of course, being Jeff Dean's famous, prolific coding partner, thought it was cool.

So Georges would posit the following argument to any doubters that they came across.

Sanjay thinks it's a good idea, and no one in the world is as smart as Sanjay.

So, why should Noam and I accept your view that it's a bad idea?

It's like if you beat the best team in football, are you the new best team in football no matter what?

Yeah.

So, all of this ends up taking Noam and George deep down the rabbit hole of probabilistic models for natural language.

Meaning, for any given sequence of words that appears on the internet, what is the probability for another specific sequence of words to follow?

This should sound pretty familiar for anybody who knows about LLM's work today.

Oh, kind of like a next word predictor.

Yeah, or next token predictor if you generalized it.

Yep.

So the first thing that they do with this work is they create the did-you-mean spelling correction in Google search.

Oh, that came out of this?

That came out of this.

GNOME created this.

So this is huge for Google because obviously it's it's a bad user experience when you mistype a query and then need to type another one, but it's attacked to Google's infrastructure because every time these mistyped queries are going, well, Google's infrastructure goes and serves the results to that query that are useless and immediately overwritten with the new one.

Right.

And it's a really tightly scoped problem where you can see like, oh, wow, 80% of the time that someone types in God groomer, oh, they actually mean dog groomer and they retype it.

And if it's really high confidence, then you actually just correct it without even asking them and then ask them if they want to opt out instead of opting in.

It's a great feature and it's sort of a great first use case for this in a very narrowly scoped domain.

Totally.

So they get this win, they keep working on it, Noman George, and they end up creating a fairly large, I'm using large in quotes here, you know, for the time, language model.

that they call affectionately Phil, the probabilistic hierarchical inferential learner.

These AI researchers love creating their backronyms.

They love their word buttons.

Yeah.

Yep.

So fast forward to 2003 and Susan Wojiski and Jeff Dean are getting ready to launch AdSense.

They need a way to understand the content of these third-party web pages, the publishers, in order to run.

the Google Ad corpus against them.

Well, Phil is the tool that they use to do it.

Huh.

I had no idea that language models were involved in this.

Yeah.

So Jeff Dean borrows Phil and famously uses it to code up his implementation of AdSense in a week because he's Jeff Dean.

And boom, AdSense.

I mean, this is billions of dollars of new revenue to Google overnight because it's the same corpus of ads that are ad words that are search ads that they're now serving on third-party pages.

They just massively expanded the inventory for the the ads that they already have in the system.

Thanks to Phil.

Thanks to Phil.

All right.

This is a moment where we got to stop and just give some Jeff Dean facts.

Jeff Dean is going to be the through line of this episode of, wait, how did Google pull that off?

How did Jeff Dean just go home and over the weekend rewrite some entire giant distributed system and figure out all of Google's problems?

Back when Chuck Norris facts were big, Jeff Dean facts became a thing internally at Google.

I just want to give you some of my favorites.

The speed of light in a vacuum used to be about 35 miles per hour.

Then Jeff Dean spent a weekend optimizing physics.

So good.

Jeff Dean's pin is the last four digits of pie.

Only Googlers would come up with these.

Yes.

To Jeff Dean, NP means no problemo.

Oh yeah.

I've seen that one before.

I think that one's my favorite.

Yes.

Oh man.

So, so good.

Also, a wonderful human being who we spoke to in research and was very, very helpful.

Thank you, Jeff.

Yes.

So, language models definitely work, definitely going to drive a lot of value for Google.

And they also fit pretty beautifully into Google's mission to organize the world's information and make it universally accessible and useful.

If you can understand the world's information and compress it and then recreate it, yeah, that fits the mission, I think.

I think that checks the box.

Absolutely.

So, Phil gets so big that apparently by the mid-2000s, Phil is using 15% of Google's entire data center infrastructure.

And I assume a lot of that is AdSense Ad serving, but also did you mean and all the other stuff that they start using it for within Google.

So early natural language systems, computationally expensive.

Yes.

So, okay, now mid-2000s, fast forward to 2007, which is a very, very big year.

for the purposes of our story.

Google had just recently launched the Google Translate product.

This is the era of all the great, great products coming out of Google that we've talked about.

Maps and Gmail and Docs and all the wonderful things that Chrome and Android are going to come later.

They had like a 10-year run where they basically launched everything you know of at Google except for search.

Truly in a 10-year run.

And then there were about 10 years after that from 2013 on where they basically didn't launch any new products that you've heard about until we get to Gemini, which is this fascinating thing.

But this 03 to 2013 era was just so rich with hit after hit after hit.

Magical.

And so one of those products was Google Translate, you know, not the same level of user base or perhaps impact on the world as Gmail or maps or whatnot, but still a magical, magical product.

And the chief architect for Google Translate was another incredible machine learning PhD named Franz Ock.

So Franz had a background in natural language processing and machine learning, and that was his PhD.

He was German.

He got his PhD in Germany.

At the time, DARPA, the Defense Advanced Research Projects Agency, division of the government, had one of their famous challenges going for machine translation.

So Google and Franz, of course, enters this, and Franz builds an even larger language model that blows away the competition in this year's version of the DARPA challenge.

This is either 2006 or 2007.

Gets a astronomically high blue score for the time.

It's called the bilingual evaluation understudy is the sort of algorithmic benchmark for judging the quality of translations.

At the time, higher than anything else possible.

So Jeff Dean hears about this and the work that Franz and the Translate team have done and it's like, this is great.

This is amazing.

When are you guys going to ship this in production?

Oh, I heard this story.

So Jeff and Noam talk about this on the DoorCash podcast.

Yes.

That episode is so, so good.

And Franz is like, no, no, no, no, Jeff, you don't understand.

This is research.

This isn't for the product.

We can't ship this model that we built.

This is a n-gram language model.

Grams are like a number of words in a cluster.

And we've trained it on a corpus of two trillion words from the Google search index.

This thing is so large, it takes it 12 hours to translate a sentence.

So the way the the DARPA challenge worked in this case was you got a set of sentences on Monday and then you had to submit your machine translation of those set of sentences by Friday.

Plenty of time for the servers to run.

Yeah, they were like, okay, so we have whatever number of hours it is from Monday to Friday.

Let's use as much compute as we can to translate these couple sentences.

Hey, learn the rules of the game and use them to your advantage.

Exactly.

So Jeff Dean being the engineering equivalent of Chuck Norris, he's like, hmm, let me see your code.

So Jeff goes and parachutes in and works with the translate team for a few months.

And he re-architects the algorithm to run on the words and the sentences in parallel instead of sequentially.

Because when you're translating a set of sentences or a set of words in a sentence, you don't necessarily need to do it in order.

You can break up the problem into different pieces, work on it independently.

You can parallelize it.

And you won't get a perfect translation, but imagine you just translate every single word.

You can at least go translate those all at the same time in parallel, reassemble the sentence, and mostly understand what the initial meaning was.

Yep.

And as Jeff knows very well, because he and Sanjay basically built it with Urz Holza, Google's infrastructure is extremely parallelizable, distributed.

You can break up workloads into little chunks, send them all over the various data centers that Google has, reassemble the projects, return that to the user.

They are the single best company in the world at parallelizing workloads across CPUs across multiple data centers.

CPUs.

We're still talking CPUs here.

Yep.

And Jeff's work with the team gets that average sentence translation time down from 12 hours to 100 milliseconds.

And so then they ship it in Google Translate, and it's amazing.

This sounds like a Jeff Dean fact.

Well, you know, it used to take 12 hours and then Jeff Dean took a few months with it.

Now it's 100 milliseconds.

Right, right, right, right, right.

So this is the first large, I'm using large in quotes here, language model used in production in a product at Google.

They see how well this works.

Like, hmm, maybe we could use this for other things, like predicting search queries as you type.

That might be interesting.

And of course, the crown jewel of Google's business that also might be interesting application for this, the ad ad quality score for AdWords is literally the predicted click-through rate on a given set of ad copy.

You can see how an LLM that is really good at ingesting information, understanding it, and predicting things based on that might be really useful for calculating ad quality for Google.

Yep, which is the direct translation to Google's bottom line.

Indeed.

Okay, so obviously all of that is great on the language model front.

I said 2007 was a big year.

Also in 2007 begins the sort of momentous intersection of several computer science professors

on the Google campus.

So in April of 2007, Larry Page hires Sebastian Thrun from Stanford to come to Google and work first part-time and then full-time on

machine learning applications.

Sebastian was the head of SAIL at Stanford, the Stanford Artificial Intelligence Laboratory, legendary AI laboratory that was big in the sort of first wave of AI back in the 60s, 70s when Larry's dad was active in the field, then actually shut down for a while and then had been restarted and re-energized here in the early 2000s.

And Sebastian was the leader, the head of SAIL.

Funny story about Sebastian, the way that he actually comes to Google.

Sebastian was kind enough to speak with us to prep for this episode.

I didn't realize it was basically an Aqua hire.

He and some, I think it was grad students were in the process of starting a company, had term sheets from Benchmark and Sequoia.

Yes.

And Larry came over and said, what if we just acquire your company before it's even started in the form of signing bonuses?

Yes.

Probably a very good decision on their part.

So SAIL, this group within the CS department at Stanford, not only had some of the most incredible, most accomplished professors and PhD AI researchers in the world, they also had this stream of Stanford undergrads that would come through and work there as researchers while they were working on their CS degrees or symbolic system degrees or, you know, whatever it was that they were doing as Stanford undergrads.

One of those people.

was Chris Cox, who's the chief product officer at Meta.

Yeah, that was kind of how he got his start in all of this and AI.

And obviously Facebook and Meta are going to come back into the story here in a little bit.

Wow.

You really can't make this up.

Another undergrad who passed through SAIL while Sebastian was there was a young freshman and sophomore who would later drop out of Stanford to start a company that went through Y Combinator's very first batch.

in summer 2005.

I'm on the edge of my seat.

Who is this?

Any guesses?

Dropbox, Reddit.

I'm trying to think who else was in the first batch.

Oh, no, no.

But way more on the nose for this episode.

The company was a failed local mobile social network.

Oh,

Sam Altman looped.

Sam Altman.

That's amazing.

He was at sale at the same time.

He was at sale.

Yep.

As an undergrad researcher.

Wow.

Wild, right?

We told you that it's a very small set of people that are all doing all of this.

Man, I miss those days.

Sam presenting at the WWEC with Steve Jobs on stage with the double pop collar.

Right.

Different time in tech.

The double popped collar.

That was amazing.

That was a vibe.

That was a moment.

Oh, man.

All right.

So April 2007, Sebastian comes over from sale into Google, Sebastian Thrun.

One of the first things he does over the next set of months is a project called Ground Truth for Google Maps, which is essentially Google Maps.

It is essentially Google Maps.

So before Ground Truth, Google Maps existed as a product, but they had to get all the mapping data from a company called Tele Atlas.

And I think there were two.

They were sort of a duopoly.

Navtech was the other one.

Yeah, Navtech and Tele Atlas.

But it was this like kind of crappy source of truth map data that everyone used, and you really couldn't do any better than anyone else because you all just used the same data.

Yep.

It was not that good, and it cost a lot of money.

Tele Atlas and Navtech were multi-billion dollar companies.

I think maybe one or both of them were public at some point, then got acquired, but a lot of money, a lot of revenue.

Yep.

And Sebastian's first thing was Street View, right?

So he already had the experience of orchestrating this fleet of all these cars to drive around and take pictures.

Yes.

So then coming into Google, Ground Truth is this sort of moonshot type project to recreate all the teleatlas data.

Mostly from their own photographs of streets from Street View.

And they incorporated some other data.

There was like census data they used.

I think it was 40 something data sources to bring it all together.

But Ground Truth was this very ambitious effort to create new maps from whole cloth.

Yep.

And just like all of the AI and AI enabled projects within Google that we're talking about here worked very, very well, very quickly.

Huge win.

Well, especially when you hire a thousand people in India to help you sift through all the discrepancies in the data and actually hand draw all the maps.

Yes, we are not yet in an era of a whole lot of AI automation.

So, on the back of this win with ground truth, Sebastian starts lobbying to Larry and Sergei.

Hey, we should do this a lot.

We should bring in AI professors, academics.

I know all these people into Google part-time.

They don't have to be full-time employees.

Let them keep their posts in academia, but come here and work with us on projects for our products.

They'll love it.

They get to see their work used by millions and millions of people.

We'll pay them.

They'll make a lot of money.

They'll get Google stock and they get to stay professors at their academic institutions.

Win, win, win.

Win-win-win.

So as you would expect, Larry and Sergei are like, yeah, yeah, yeah, that's a good idea.

Let's do that.

More of that.

So in December of 2007, Sebastian brings in a relatively little-known machine learning professor from the University of Toronto named Jeff Hinton.

to the Google campus to come and give a tech talk.

Not yet hiring him, but come give a tech talk to, you know, all the folks at Google and talk about some of the new work, Jeff, that you and your PhD and postdoc students there at the University of Toronto are doing on blazing new paths with neural networks.

And Jeff Hinton, for anybody who doesn't know the name, now very much known as the godfather of neural networks and really the godfather of kind of the whole direction that AI went in.

Modern AI.

He was kind of a fringe academic

at this point in history.

I mean, neural networks were not a respected subtree of AI.

No, totally not.

And part of the reason is there had been a lot of hype 30, 40 years before around neural networks that just didn't pan out.

So it was effectively, everyone thought, disproven and certainly backwater.

Yep.

Ben, do you remember from our NVIDIA episodes my favorite piece of trivia about Jeff Hinton?

Oh, yes, that his grandfather, great-grandfather was George Boole.

Yep.

He is the great-great-grandson of George and Mary Boole, who invented Boolean algebra and Boolean logic.

Which is hilarious now that I know more about this, because that's the basic building block of symbolic logic, of defined deterministic computer science logic.

And the hilarious thing about neural nets is it's not.

It's not symbolic AI.

It's not I feed you the specific instructions and you follow a big if-then tree.

it is non-deterministic.

It is the opposite of that field.

Which actually just underscores, again, how sort of heretical this branch of machine learning and computer science was.

Right.

So, Ben, as you were saying earlier, neural networks, not a new idea and had all of this great promise in theory, but in practice, just took too much computation to do multiple layers.

You could really only have a single or maybe small single-digit number of layers in in a computer neural network up until this time.

But

Jeff and his former postdoc, a guy named Jan Lacun,

started vandalizing within the community, hey, if we can find a way to have multi-layered, deep-layered neural networks, something we call deep learning, we could actually realize the promise here.

It's not that the idea is bad.

It's that the implementation, which would take a ton of compute to actually do all the math, to do all the multiplication required to propagate through layer after layer after layer of neural networks to sort of detect and understand and store patterns.

If we could actually do that, a big multi-layered neural network would be very valuable and possibly could work.

Yes.

Here we are now in 2007, mid-2000s.

Moore's Law has increased enough that you could actually start to try to test some of these theories.

Yep.

So Jeff comes and he gives this talk at Google.

It's on YouTube.

You can go watch it.

We'll link to it in the show notes.

This is incredible.

This is an artifact of history sitting there on YouTube.

And

people at Google, Sebastian, Jeff Dean, all the other folks we're talking about, they get very, very, very excited because they've already been doing stuff like this with Translate and the language models that they're working with.

That's not using...

deep neural networks that Jeff's working on.

So here's this whole new architectural approach that if they could get it to work, would enable these models that they're building to work way better, recognize more sophisticated patterns, understand the data better.

Very, very promising.

Again, kind of all in theory at this point.

Yep.

So Sebastian Thrun brings Jeff Hinton into the Google fold after this tech talk.

I think first as a consultant over the next couple of years, and then this is amazing.

Later, Jeff Hinton technically becomes an intern at Google.

Like that's how they get around the part-time, full-time policies here.

Yep.

He was a summer intern in somewhere around 2011, 2012.

And mind you, at this point, he's like 60 years old.

Yes.

So in the next couple of years after 2007 here, Sebastian's concept of bringing these computer science machine learning academics into Google as contractors or part-time or interns, basically letting them keep their academic posts and work on big projects for Google's products internally go so well that by late 2009, Sebastian and Larry and Sergei decide, hey, we should just start a whole new division within Google.

And it becomes Google X, the Moonshot Factory.

The first project within Google X, Sebastian leads himself.

Ooh, David, don't say it.

Don't say it.

I won't say the name of it.

We will come back to it later.

But for our purposes for now, the second project would be critically important, not only for our story, but to the whole world.

Everything in AI, changing the entire world.

And that second project is called Google Brain.

But before we tell the Google Brain story, now is a great time to thank our friends at JPMorgan Payments.

Yes.

So today we are going to talk about one of the core components of JP Morgan Payments, their treasury solutions.

Now, treasury is something that most listeners probably do not spend a lot of time thinking about, but it's fundamental to every company.

Yep.

Treasury used to be just a back office function, but now great companies are using it as a strategic lever.

With JP Morgan Payments Treasury Solutions, you can view and manage all your cash positions in real time and all of your financial activities across 120 currencies in 200 countries.

And the other thing that they acknowledge really in their whole strategy is that every business has its own quirks.

So it's not a cookie-cutter approach.

They work with you to figure out what matters most for you and your business and then help you gain clarity, control, and confidence.

So whether you need advanced automation or just want to cut down on manual processes and approvals, their real-time treasury solutions are designed to keep things running smoothly, whether your treasury is in the millions or billions, or perhaps like the company we're talking about this episode, in the hundreds of billions of dollars.

And they have some great strategic offerings like Pay By Bank, which lets customers pay you directly from their bank account.

It's simple, secure, tokenized, and you get faster access to funds and enhanced data to optimize revenue and reduce fees.

This lets you send and receive real-time payments instantly just with a single API connection to JP Morgan.

And because JP Morgan's platform is global, that one integration lets you access 45 countries and counting and lets you scale basically infinitely as you expand.

As we've said before, JP Morgan Payments moves $10 trillion a day.

So scale is not an issue for your business.

Not at all.

If you're wondering how to actually manage all that global cash, JPMorgan again has you covered with their liquidity and account solutions that make sure you have the right amount of cash and the right currencies in the right places for what you need.

So whether you're expanding into new markets or just want more control over your funds, JPMorgan Payments is the partner you want to optimize liquidity, streamline operations, and transform your treasury.

To learn more about how JPMorgan can help you and your company, just go to jpmorgan.com slash acquired and tell them that Ben and David sent you.

All right, David.

So Google Brain.

So when Sebastian left Stanford full-time and joined Google full-time, of course, somebody else had to take over sale.

And the person who did is another computer science professor, brilliant guy named Andrew Ng.

This is like all the hits.

All the hits.

This is all the AI hits on this episode.

So what does Sebastian do?

He recruits Andrew to come part-time, start spending a day a week on the Google campus.

And this coincides right with the start of X and Sebastian formalizing this division.

So one day in 2010, 2011 timeframe, Andrew's spending his day a week on the Google campus and he bumps into who else?

Jeff Dean.

And Jeff Dean is telling Andrew about what he and Franz have done with language models and what Jeff Hinton is doing in deep learning.

Of course, Andrew knows all this.

And Andrew's talking about what he and Sale are doing at Stanford.

And they decide, you know,

the time might finally be right to try and take a real big swing on this within Google and build a massive,

really large, deep learning model in the vein of what Jeff Hinton has been talking about on highly parallelizable Google infrastructure.

And when you say the time might be right, Google had tried twice before and neither project really worked.

They tried this thing called Brains on Borg.

Borg is sort of an internal system that they use to run all of their infrastructure.

They tried the Cortex project and neither of these really worked.

So there's a little bit of scar tissue in the sort of research group at Google of our large-scale neural networks actually going to work for us on Google infrastructure.

So the two of them, Andrew Ng and Jeff Dean, pull in Greg Corrado, who is a neuroscience PhD and amazing researcher who was already working at Google.

And in 2011, the three of them launch the second official project within X, appropriately enough, called Google Brain.

And the three of them get to work building a really, really big,

deep neural network model.

And if they're going to do this, they need a system to run it on.

You know, Google is all about taking this sort of frontier research and then doing the architectural and engineering system to make it actually run.

Yes.

So Jeff Dean is working on this system on the infrastructure, and he decides to name the infrastructure DIST Belief, which of course is a pun, both on the distributed nature of the system and also on, of course, the word disbelief, because no one thought it was going to work.

Most people in the field thought this was not going to work and most people in Google thought this was not going to work.

And here's a little bit on why.

And it's a little technical, but follow me for a second.

All the research from that period of time pointed to the idea that you needed to be synchronous.

So all the compute needed to be sort of really dense, happening on a single machine with really high parallelism, kind of like what GPUs do, that you really would want it all sort of happening in one place.

So it's really easy to kind of go look up and see, hey, what are the computed values for everything else in the system before I take my next move?

What Jeff Dean wrote with Disc Belief was the opposite.

It was distributed across a whole bunch of CPU cores and potentially all over a data center or maybe even in different data centers.

So in theory, this is really bad because it means you would need to be constantly waiting around on any given given machine for the other machines to sync their updated parameters before you could proceed.

But instead, the system actually worked asynchronously without bothering to go and get the latest parameters from other cores.

So you were sort of updating parameters on stale data.

You would think that wouldn't work.

The crazy thing is it did.

Yes.

Okay, so you've got disk belief.

What do they do with it now?

They want to do some research.

So they try out, can we do cool neural network stuff?

And what they do in a paper that they submitted in 2011, right at the end of the year, is

I'll give you the name of the paper first: Building High-Level Features Using Large-Scale Unsupervised Learning.

But everyone just calls it the cat paper.

The cat paper.

You talk to anyone at Google, you talk to anyone at AI, they're like, oh, yeah, the cat paper.

What they did was they trained a large nine-layer neural network to recognize cats from unlabeled frames of YouTube videos using 16,000 CPU cores on a thousand different machines.

And listeners, just to like underscore how seminal this is, we actually talked with Sundar in prep for the episode, and he cited seeing the cat paper come across his desk as one of the key moments that sticks in his brain in Google's story.

Yeah.

A little later on, they would do a TGIF where they would present the results of the cat paper and you talk to people at Google.

They're like, that TGIF.

Oh my God, that's when it all changed.

Yeah.

It proved that large neural networks could actually learn meaningful patterns without supervision and without labeled data.

And not only that, it could run on a distributed system that Google built to actually make it work on their infrastructure.

And that is a huge unlock of the whole thing.

Google's got this big infrastructure asset.

Can we take this theoretical computer science idea that the researchers have come up with and use dist belief to actually run it on our system yep that is the amazing technical achievement here that is almost secondary to the business impact of the cat paper i think it's not that much of a leap to say that the cat paper led to probably hundreds of billions of dollars of revenue generated by Google and Facebook and by dance over the next decade.

Definitely.

Pattern recognizers in data.

So YouTube had a big problem at this time, which was that people would upload these videos.

There's tons of videos being uploaded to YouTube, but people are really bad at describing what is in the videos that they're uploading.

And YouTube is trying to become more of a destination site, trying to get people to watch more videos, trying to build a feed, increase dwell time, et cetera, et cetera.

And the problem is the recommender is trying to figure out what what to feed and it's only just working off titles and descriptions that people were writing about their own videos.

Right.

And whether you're searching for a video or they're trying to figure out what video to recommend next, they need to know what the video is about.

Yep.

So the cat paper proves that you can use this technology, a deep neural network running on disk belief

to go inside.

of the videos in the YouTube library and understand what they were about and use that data to then figure out what videos to serve to people.

If you can answer the question, cat or not a cat, you can answer a whole lot more questions too.

Here's a quote from Jeff Dean about this.

We built a system that enabled us to train pretty large neural nets through both model and data parallelism.

We had a system for unsupervised learning on 10 million randomly selected YouTube frames, as you were saying, Ben.

It would build up unsupervised representations based on trying to reconstruct the frame from the high-level representations.

We got that working and training on on 2,000 computers using 16,000 cores.

After a little while, that model was actually able to build a representation at the highest neural net level where one neuron would get excited by images of cats.

It had never been told what a cat was, but it had seen enough examples of them in the training data of head-on facial views of cats that that neuron would then turn on for cats and not much else.

It's so crazy.

I mean, this is the craziest thing about unlabeled data, unsupervised learning, that a system can learn what a cat is without ever being explicitly told what a cat is.

And that there's a cat neuron.

Yeah.

And so then there's a iPhone neuron and a San Francisco Giants neuron and all the things that YouTube recommends.

Not to mention porn filtering, explicit content filtering.

Not to mention copyright identification and enabling revenue share with copyright holders.

Yeah, this leads to everything in YouTube.

Basically puts YouTube on the path to today becoming the single biggest property on the internet and the single biggest media company in the planet.

This kicks off a 10-year period from 2012 when this happens until ChatGPT on November 30th, 2022, when AI is already shaping the human existence for all of us and driving hundreds of billions of dollars of revenue.

It's just in the YouTube feed and then Facebook borrows it and they hire Jan Lacun and they start Facebook AI research and then they bring it into Instagram and then TikTok and ByteDance take it and then it goes back to Facebook and YouTube with reels and shorts.

This is the primary way that humans on the planet spend their leisure time for the next 10 years.

This is my favorite David Rosenthalism.

Everyone talks about 2022 onward as the AI era.

And I love this point from you that actually, for anyone that could make good use of a recommender system and a classifier system, basically a company with a social feed, the AI era started in 2012.

Yes, the AI era started in 2012.

and part of it was the cat paper.

The other part of it was what Jensen and Nvidia always calls the big bang moment for AI,

which was AlexNet.

Yes.

So we talked about Jeff Hinton.

Back at the University of Toronto, he's got two grad students who he's working with in this era.

Alex Kruszewski and Ilya Sutskever.

Of course.

Future co-founder and chief scientist of OpenAI.

And the three of them are working with Jeff's deep neural network ideas and algorithms to create an entry

for the famous ImageNet competition in computer science.

This is Feifei Lee's thing from Stanford.

It is an annual machine vision algorithm.

competition.

And what it was was Feifei had assembled a database database of 14 million images that were hand-labeled.

Famously, she used Mechanical Turk on Amazon, I think, to get them all hand-labeled.

Yes, I think that's right.

And so then the competition was what team can write the algorithm that without looking at the labels, so just seeing the images, could correctly identify the largest percentage.

The best algorithms that would win the competitions year over year, we're still getting more than a quarter of the images wrong.

So like 75% success rate, great.

Way worse than a human.

Can't use it for much in a production setting when a quarter of the time you're wrong.

So then the 2012 competition.

Along comes AlexNet.

Its error rate was 15%.

Still high, but a 10% leap from the previous best being a 25% error rate all the way down to 15 in one year.

A leap like that had never happened before.

It's 40% better than the next best.

Yes.

On a relative basis.

Yes.

And why is it so much better, David?

What did they figure out that would create a $4 trillion company in the future?

So, what Jeff and Alex and Ilya did

is they knew, like we've been talking about all episode, that deep neural networks had all this potential, and Moore's Law had advanced enough that you could use CPUs to create a few layers.

They had the aha moment of what if we re-architected this stuff, not to run on CPUs, but to run on a whole different class of computer chips that were by their very nature highly, highly, highly parallelizable video game graphics cards made by the leading company in the space at the time.

NVIDIA.

Not obvious at the time, and especially not obvious that this highly advanced cutting-edge academic computer science research that was being done on supercomputers, usually that was being done on supercomputers with incredible CPUs would use these toy video game cards that retail for $1,000.

Yeah, less at that point in time, a couple hundred bucks.

So the team in Toronto, they go out to like the local Best Buy or something.

They buy two NVIDIA GeForce GTX 580s, which were NVIDIA's top top-of-the-line of the gaming cards at the time.

The Toronto team rewrites their neural network algorithms in CUDA, NVIDIA's programming language.

They train it on these two off-the-shelf GTX 580s, and

this is how they achieve their deep neural network and do 40% better than any other entry in the ImageNet competition.

So when Jetson says that this was the big bang moment of artificial intelligence, A, he's right.

This shows everybody that, holy crap, if you can do this with two off-the-shelf GTX 580s, imagine what you could do with more of them or with specialized chips.

And B,

this event is what sets NVIDIA on the path from a somewhat struggling PC gaming accessory maker to the leader of the AI wave and the most valuable company in the world today.

And this is how AI research tends to work is there's some breakthrough that gets you this big step change function.

And then there's actually a multi-year process of optimizing from there, where you get these kind of diminishing returns curves on breakthroughs where the first half of the advancement happens all at once.

And then the second half takes many years after that to figure out.

But it's rare and amazing and must be so cool when you have an idea, you do it, and then you realize, oh my God, I just found the next giant leap in the field.

It's like I unlocked the next level to use the video game analogy.

Yes.

I leveled up.

So after alexnet the whole computer science world is a buzz people are starting to stop doubting neural networks at this point yes so after alexnet the three of them from toronto jeff hinton alex kuszewski and ilya sitskeever do the natural thing they start a company called dnn research deep neural network research this company does not have any products this company has ai researchers who just want a big competition and predictably as you might imagine, it gets acquired by Google almost immediately.

Oh, are you intentionally shortening this?

That's what I thought the story was.

Oh, it is not immediately.

Oh, okay.

There's a whole crazy thing that happens where

the first bid is actually from Baidu.

Oh,

I did not know that.

So Baidu offers $12 million.

Jeff Hinton doesn't really know how to value the company and doesn't know if that's fair.

And so he does what any academic would do to best determine the market value of the company.

He says, thank you so much.

I'm going to run an auction now and I'm going to run it in a highly structured manner where every time anybody wants to bid, the clock resets and there's another hour where anybody else can submit another bid.

No way.

So I didn't know this.

This is crazy.

He gets in touch with everyone that he knows from the research community who is now working at a big company who he thinks, hey, this would be a good place for us to do our research.

That includes Baidu.

That includes Google.

That includes Microsoft.

And there's one other.

Facebook, of course.

It's a two-year-old startup.

Oh, wait, so it does not include Facebook?

It does not include Facebook.

Think about the year.

This is 2012.

So

Facebook's not really in the AI game yet.

They're still trying to build their own AI lab.

Yeah, yeah, because Jan Lacoon and Fair would start in in 2013.

Is it Instagram?

Nope.

It is the most important part of the end of this episode.

Wait, well, it can't be Tesla, because Tesla was older than that.

Nope.

Well, OpenAI wouldn't get founded for years.

Wow.

Okay, you really got me here.

What company slightly predated OpenAI?

Doing effectively the same mission?

Oh,

of course, of course.

Hiding in plain sight.

DeepMind.

Wow.

DeepMind, baby.

They are the fourth bidder in a four-way auction for DNN research.

Now, of course, right after the bidding starts, DeepMind has to drop out.

They're a startup.

They don't actually have the cash to be able to buy.

Yeah.

Didn't even cross my mind because my first question was like, where the hell would they get the money?

Because they had no money.

But Jeff Hinton already knows and respects Demis.

Even though he's just doing this at the time startup called DeepMind.

That's amazing.

Wait, how is DeepMind in the auction, but Facebook is not?

Isn't that wild?

That's wild.

So the timing of this is

concurrent with the, it was then called NIPS.

Now it's called Nurips Conference.

So Jeff Hinton actually runs the auction from his hotel room at the Hera's Casino in Lake Tahoe.

Oh my God.

Amazing.

So the bids all come in, and we got to thank Cade Metz, the author of Genius Makers, great book on the whole history of AI that we're actually going to reference a lot in this episode.

The bidding goes up and up and up.

At some point, Microsoft drops out, they come back in, told you DeepMind drops out.

So it's Baidu and Google really going at the end.

And finally, at some point, the researchers look at each other and they say, Where do we actually want to land?

We want to land at Google.

And so they stop the bidding at $44 million and just say, Google, this is more than enough money.

We're going with you.

Wow.

I knew it was about $40 million.

I did not know that old story.

It's almost like Google itself and, you know, the Dutch auction IPO process.

Right?

How fitting.

That's kind of a perfect DNA.

Yes.

Wow.

And the three of them were supposed to split at 33 each.

And Alex and Ilya go to Jeff and say, I really think you should have a bigger percent.

I think you should have 40% and we should each have 30%.

And that's how it ends up breaking down.

Ah, wow.

What a team.

Well,

that leads to the three of them joining Google Brain directly and turbocharging everything going on there.

Spoiler alert, a couple of years later, Astro Teller, who would take over running Google X after Sebastian Thrun left, he would get quoted in the New York Times in a profile of Google X that the gains to Google's core businesses and search and ads and YouTube from Google Brain have way more than funded all of the other bets that they have made within Google X and throughout the company over the years.

It's one of these things that if you make something a few percent better that happens to do tens of billions of dollars or hundreds of billions of dollars in revenue, you find quite a bit of loose change in those couch cushions.

Yes, quite quite a bit of loose change.

But that's not where the AI history ends within Google.

There is another

very important piece of the Google AI story that is an acquisition from outside of Google, the AI equivalent of Google's acquisition of YouTube.

It's what we talked about in a minute ago, Deep Mind.

But before we tell the Deep Mind story, now is a great time to thank a new partner of ours, Sentry.

Yes.

Listeners, that is S-E-N-T-R-Y, like someone standing guard.

Yes.

Sentry helps developers debug everything from errors to latency and performance issues, pretty much any software problem, and fix them before users get mad.

As their homepage puts it, they are considered quote-unquote not bad by over 4 million software developers.

And today we're talking about the way that Sentry works with another company in the acquired universe, Anthropic.

Anthropic used to have some older monitoring systems in place, but as they scaled and became more complex, they adopted Sentry to find and fix issues faster.

So when you're building AI models, like we're talking about all episode here, small issues can ripple out into big ones fast.

Let's say you're running a huge compute job like training a model.

If one node fails, it can have massive downstream impacts, costing huge amounts of time and money.

Sentry helped Anthropic detect bad hardware early so they could reject it before causing a cascading problem and taking debugging down to hours instead of days for them.

And one other fun update from Sentry, they now have an AI debugging agent called Sear.

Sear uses all the context that Sentry has about your app usage to run root cause analysis as issues are detected.

It uses errors, span data, logs, and tracing, and your code to understand the root cause, fix it, and get you back to shipping.

It even creates pull requests to merge code fixes in.

And on top of that, they also recently launched agent and MCP server monitoring.

AI tooling tends to offer limited visibility into what's going on under the hood, shall we say.

Sentry's new tools make it easy to understand exactly what's going on.

This is everything from actual AI tool calls to performance across different models and interactions between AI and the downstream services.

We're pumped to be working with Sentry.

We're big fans of the company and of all the great folks we're working with there.

They have an incredible customer list, including not only Anthropic, but Cursor, Vercel, Linear, and more.

And actually, if you're in San Francisco or the Bay Area, Century is hosting a small invite-only event with Dave and I in San Francisco for product builders on October 23rd.

You can register your interest at century.io slash acquired.

That's century, S-E-N-T-R-Y dot I-O slash acquired, and just tell them that Ben and David sent you.

All right, David, deep mind.

I kind of like your framing, the YouTube of AI.

The YouTube of AI for Google.

They bought this thing for, we'll talk about the purchase price, but it's worth, what, $500 billion today?

I mean, this is as good as Instagram or YouTube in terms of greatest acquisitions of all time.

100%.

So I remember when this deal happened, just like I remember when the Instagram deal happened.

Because the number was big at the time.

It was big, but I remember it for a different reason.

It was like when Facebook bought Instagram, like, oh my God, this is, wow, what a tectonic shift in the landscape of tech.

In January 2014, I remember reading on TechCrunch this random news.

Right.

You're like, deep what?

That Google is spending a lot of money to buy something in London that I've never heard of.

That's working on artificial intelligence question mark.

Right.

This really illustrates how outside of mainstream tech AI was at the time.

Yeah.

And then you dig in a little further and you're like, this company doesn't seem to have any products.

And it also doesn't even really say anything on its website about what DeepMind is.

It says it is a quote-unquote cutting-edge artificial intelligence company.

Wait, did you look this up on the Wayback Machine?

I did.

I did.

Oh, nice.

To build general-purpose learning algorithms for simulations, e-commerce, and games.

This is 2014.

This does not compute, does not register.

Simulations, e-commerce, and games.

It's kind of a random spattering of.

Exactly.

It turns out, though, not only was that description of what DeepMind was fairly accurate, this

company and this purchase of it by Google was the butterfly flapping its wings equivalent moment that directly leads to OpenAI, ChatGPT, Anthropic, and basically everything.

Certainly Gemini that we know.

Yeah.

Gemini directly in the world of AI today.

And probably XAI, given Elon's involvement.

Yeah, of course XAI.

In a weird way, it sort of leads to Tesla self-driving too, with Karpathy.

Yeah, definitely.

Okay, so what is the story here?

DeepMind was founded in 2010 by a neuroscience PhD named Demis Hassabas.

Who previously started a video game company?

Oh, yeah.

And a postdoc named Shane Legg at University College London.

And a third co-founder who was one of Demis' friends from growing up, Mustafa Suleiman.

This was unlikely, to say the least.

This would go on to produce a knight and Nobel Prize winner.

Yes.

So Demis,

the CEO, was a childhood chess prodigy turned video game developer who, when he was age 17 in 1994

he had gotten accepted to the university of cambridge but he was too young and the university told him hey take a you know gap year come back he decided that he was going to go work at a video game developer at a video game studio called bullfrog productions for the year and while he's there he created the game theme park if you remember that it was like a theme park version of sim city this was a big big game.

This was very commercially successful.

Rollercoaster Tycoon would be sort of a clone of this that would have many, many sequels over the years.

Oh, I played a ton of that.

Yeah.

It sells 15 million copies in the mid-90s.

Wow.

Wild.

Then after this, he goes to Cambridge, studies computer science there.

After Cambridge, he gets back into gaming.

founds another game studio called Elixir.

That would ultimately fail.

And then he decides, you know what, I'm going to go get my PhD in neuroscience.

And that is how Demis ends up at University College, London.

There he meets Shane Legg, who's there as a postdoc.

Shane is a self-described, at the time, member of the lunatic fringe in the AI community,

in that

he believes,

this is 2008, 9, 10, he believes that AI is going to get more and more and more powerful every year, and that it will become so powerful that it will become more intelligent than humans.

And Shane is one of the people who actually popularizes the term artificial general intelligence, AGI.

Oh, interesting.

Which, of course, lots of people talk about now, and approximately zero people were afraid of that.

I mean, you had like the Nick Bostrom type folks, but very few people were thinking about super intelligence or the singularity or anything like that.

For what it's worth, not Elon Musk.

He's not included in that list because Demis would be the one who tells Elon about this.

Yes, we'll get to it.

So Demis and Shane hit it off.

They pull in Mustafa, Demis' childhood friend, who is himself extremely intelligent.

He had gone to the University of Oxford and then dropped out, I think, at age 19 to do other startup-y type stuff.

So the three of them decided to start a company, DeepMind, the name, of course, being a reference to deep learning, Jeff Hinton's work and everything coming out of the University of Toronto, and the goal that the three of these guys have of actually creating an intelligent mind with deep learning.

Like Jeff and Ilya and Alex aren't really thinking about this yet.

As we said, this is lunatic fringe type stuff.

Yes.

AlexNet, the cat paper, that whole world is about better classifying data.

Can we better sort into patterns?

It's a giant leap from there to say, oh, we're going to create intelligence.

Yes.

I think probably some people, probably almost certainly at Google, were thinking, oh, we can create narrow intelligence that'll be better than humans at certain tasks.

I mean, a calculator is better than humans at certain tasks.

Right.

But I don't think too many people were thinking, oh, this is going to be general intelligence smarter than humans.

Right.

So they decide on the tagline for the company.

is going to be solve intelligence and use it to solve everything else.

Ooh, I like it.

I like it.

Yeah.

Yeah.

They're good marketers too, these guys.

So there's just one problem.

To do what they want to do.

Money.

Just say it.

Money is the problem.

Right, right, right.

Money is the problem for lots of reasons, but even more so than any other given startup in the 2010 era, it's not like they can just go spin up an AWS instance and like build an app and deploy it to the App Store.

They want to build really, really, really, really, really big deep learning neural networks that requires Google-sized levels of compute.

Well, it's interesting.

It actually, they don't require that much funding yet.

The AI of the time was go grab a few GPUs.

We're not training giant LLMs.

That's the ambition eventually.

But right now, what they just need to do is raise a few million bucks.

But who's going to give you a few million bucks when there's no business plan?

When you're just trying to solve intelligence, you need to find some lunatics.

It's a tough sell to VCs.

Except for the exact same thing.

As you say, they need to find some lunatics.

Oh, I chose my words carefully.

Yeah.

We use the term lunatic in

most endearing possible way here, given that they were all basically right.

So in June 2010, Demis and Shane managed to get invited to the Singularity Summit.

in San Francisco, California.

Because they're not raising money for this in London.

Yeah, definitely not.

i think they tried for a couple months and learned that that was not going to be a viable path yes the summit the singularity summit organized by ray kurzweil a future google employee i think chief futurist and noted futurist eliezer yudkowski

and

peter teal

yes

so Demis and Shane are excited about getting this invite.

They're like, this is probably our one chance to get funded.

But we probably shouldn't just walk in guns blazing and say, Peter, can we pitch you?

Yeah.

So they finagle their way into Demis

getting to give a talk on stage at the summit.

Always the hack.

They're like, this is great.

This is going to be the hack.

The talk is going to be our pitch to Peter and Founders Fund.

Peter has just started Founders Fund at this point, obviously.

member of the PayPal Mafia, very wealthy.

I think he had a big Roth IRA at this point is the right way to frame it.

Big Big Roth IRA that he had invested in Facebook, first investor in Facebook.

He is the perfect target.

They architect the presentation at the summit to be a pitch directly to Peter, essentially, a thinly veiled pitch.

Shane has a quote in Parmi Olson's great book, Supremacy, that we used as a source for a lot of this deep mind story.

And Shane says, we needed someone crazy enough to fund an AGI company, somebody who had the resources not to sweat a few million and liked super ambitious stuff.

They also had to be massively contrarian because every professor that he would go talk to would certainly tell him, absolutely do not even think about funding this.

That Venn diagram sure sounds a lot like Peter Thiel.

So they show up at the conference.

Demis is going to give the talk.

Goes out on stage.

He looks out into the audience.

Peter is not there.

Turns out Peter wasn't actually that involved in the conference.

No, he's a busy guy.

He's a co-founder, co-organizer, but is a busy guy.

Yes.

Guy's like, shoot.

Oh, we missed our chance.

What are we going to do?

And then fortune turns in their favor.

They find out that Peter is hosting an after-party that night at his house in San Francisco.

They get into the party.

Demis seeks out Peter and he's like, Demis is very, very, very smart, as anybody who's ever listened to him talk would immediately know.

He's like, rather than just pitching Peter head on, I'm going to come about this obliquely.

He starts talking to Peter about chess because he knows, as everybody does, that Peter Thiel loves chess.

And Demis had been the second highest ranked player in the world as a teenager in the under 14 category.

Good strategy.

Great strategy.

The man knows his chess moves.

So Peter's like, hmm, I like you.

You seem smart.

What do you do?

And Demis explains he's got this AGI startup.

They were actually here.

He gave a talk on stage as as part of the conference.

People are excited about this.

And Peter said, oh, okay.

All right.

Come back to Founders Fund tomorrow and give me the pitch.

So they do.

They make the pitch.

It goes well.

Founders Fund leads DeepMind's seed round of about $2 million.

My, how times have changed for AI company seed rounds these days.

Oh, yes.

Imagine leading DeepMinds seed round with.

less than $2 million check.

And through Peter and Founders Fund, they get introduced.

Hey, Elon, you should meet this guy.

To another member of the PayPal Mafia, Elon Musk.

Yes.

So it's teed up in a pretty low-key way.

Hey, Elon, you should meet this guy.

He's smart.

He's thinking about artificial intelligence.

So Elon says, great, come over to SpaceX.

I'll give you the tour of the place.

So Demis comes over for lunch and a tour of the factory.

Of course, Demis thinks it's very cool, but really he's trying to reorient the conversation over to artificial intelligence.

And I'll read this great excerpt from an article in The Guardian.

Musk told Hasabis his priority was getting to Mars as a backup planet in case something went wrong here.

I don't think he'd thought much about AI at this point.

Hasabis pointed out a flaw in his plan.

I said, what if AI was the thing that went wrong here?

Then being on Mars wouldn't help you because if we got there, then it would obviously be easy for an AI to get there through our communication systems or whatever it was.

He hadn't thought about that.

So he sat there for a minute without saying anything, just sort of thinking, hmm, that's probably true.

Shortly after, Musk, too, became an investor in DeepMind.

Yes.

Yes, yes.

I think it's crazy that Demis is sort of the one that woke Elon up to this idea of we might not be safe from the AI on Mars either.

Right, right.

I hadn't considered that.

So this is the first time the bit flips for Elon of we really need to figure out a safe, secure AI for the good of the people, that sort of seed being planted in his head.

Yep.

Which of course is what DeepMind's ambition is.

We are here doing research for the good of humanity like scientists in a peer-reviewed way.

Yep.

I think all that is true.

Also

in the intervening months to year

after

this meeting between Demis and Elon and Elon investing in DeepMind, Elon also starts to get really, really excited and convinced about the capabilities of AI in the near term.

And specifically, the capabilities of AI for Tesla.

Yes.

Like with everything else in Elon's world, once the bit flips and he becomes interested, he completely changes the way he views the world, completely sheds all the old ways and actions that he was taking.

And it's all about what do I most do to embrace this new worldview that I have.

And other people have been working on for a while already by this point, AI driving cars.

Yep.

That sounds like it would be a pretty good idea for Tesla.

Does.

So Elon

starts trying to recruit as many AI researchers as he possibly can and machine vision and machine learning experts into Tesla.

And then AlexNet happens.

And man, AlexNet's really, really, really good at identifying and classifying images and cat videos on YouTube and the YouTube recommender feed.

Well, is that really that different from a live feed of video from a car that's being driven and understanding what's going on there?

Can we process it in real time and look at differences between frames?

Perhaps controlling the car?

Not all that different.

So Elon's excitement, channeled initially through DeepMind and Demis about AI and AI for Tesla, starts ratcheting up big time.

Yep.

Meanwhile, back in London, DeepMind is getting to work.

They're hiring researchers.

They're getting to work on models.

They're making some vague noises about products to their investors.

Maybe we could do something in shopping, maybe something in gaming, like the description on the website at the time of acquisition said.

But mostly what they really, really want to do is just build these models and work on intelligence.

And then one day in late 2013, they get a call from Mark Zuckerberg.

He wants to buy the company.

Mark has woken up to everything that's going on at Google after AlexNet and what AI is doing for social media feed recommendations at YouTube, the possibility of what it can do at Facebook and for Instagram.

He's gone out and recruited Jan Lacun, Jeff Hinton's old postdoc, who's together with Jeff, one of the sort of godfathers of AI and deep learning.

And really popularized the idea of convolutional neural networks, the next hot thing in the field of AI at this point in time.

And so with Jan, they have created FAIR, Facebook AI Research, which is a Google brain rival within Facebook.

And remember who the first investor in Facebook was, who's still on the board, Peter Thiel.

And is also the lead investor in DeepMind.

Where do you think Mark learned about DeepMind?

Peter Thiel.

Was Was it, do you know for sure that it was from Peter?

No, I don't know for sure, but like, how else could Mark have learned about this startup in London?

I've got a great story of how Larry Page found out about it.

Oh, okay.

Well, we'll get to that in one sec.

So Mark calls and offers to buy the company.

And there are various rumors of how much Mark offered, but according to Parmi Olson in her book, Supremacy, the reports are that it was up to $800 million.

Company with no products and a long way from AGI.

That squares with what Cade Metz has in his book, that the founders would have made about twice as much money from taking Facebook's offer versus taking Google's offer.

Yep.

So Demis, of course, takes this news to the investor group.

Which, by the way, is kind of against everything the company was founded on.

The whole aim of the company and what he's promised the team is that DeepMind is going to stay independent, do research, publish in the scientific community.

We're not going to be sort of captured and told what to do by the whims of a capitalist institution.

Yep.

So definitely some deal point negotiating that has to happen with Mark and Facebook if this offer is going to come through.

But Mark is so desperate at this point.

He is open to these very large deal point negotiations, such as Jan Lacun gets to stay in New York.

Jan Lacun gets to stay operating his lab at NYU.

Jan Lacun is a professor.

He's flexible on some things.

Turns out, Mark is not flexible on letting Demis keep control of DeepMind if he buys it.

Demis sort of argued for we need to stay separate and carved out and we need this independent oversight board with his ability to intervene if the mission of deep mind is no longer being followed.

And Mark's like, no,

you'll be a part of Facebook.

Yeah.

And you'll make a lot of money.

So as this negotiation is going on, of course, the investors in DeepMind get wind of this.

Elon

finds out about what's going on.

He immediately calls up Demis and says, I will buy the company right now with Tesla stock.

This is late 2013, like early 2014.

Tesla's market cap is about $20 billion.

So Tesla stock from then to today is about a 70x run up.

Demis and Shane and Mustafa are like, wow, okay, there's a lot going on right now.

But to your point, They have the same issues with Elon and Tesla that they had with Mark.

Elon wants them to come in and work on autonomous driving for Tesla.

They don't want to work on autonomous driving.

Right.

Or at least exclusively.

At least exclusively.

Yep.

So

then Demis gets a third call from Larry Page.

Do you want my story of how Larry knows about the company?

I absolutely want your story of how Larry knows about the company.

All right.

So this is still early in Deep Mind's life.

We haven't progressed all the way to this acquisition point yet.

Apparently, Elon Musk Musk is on a private jet with Luke Nosick, who's another member of the PayPal Mafia and an angel investor in DeepMind.

And they're reading an email from Demis with an update about a breakthrough that they had where DeepMind AI figured out a clever way to win at the Atari game breakout.

Yes.

And the strategy it figured out with no human training was that you could bounce the ball up around the edges of the bricks and then without needing to intervene, it could bounce around along the top and win the game faster without you needing to have a whole bunch of interactions with the paddle down at the bottom.

They're watching this video of how clever it is, and flying with them on the same private plane is Larry Page.

Of course, because Elon and Larry used to be very good friends, yes.

And Larry is like,

Wait, what are you watching?

What company is this?

And that's how he finds out.

Wow, yes, Elon must have been so

angry about all this.

And the crazy thing is, this kinship between Larry and Demis is, I think, the reason why the deal gets done at Google.

Once the two of them get together, they are like peas in a pod.

Larry has always viewed Google as an AI company.

Yep.

Demis, of course, views DeepMind so much as an AI company that he doesn't even want to make any products until they can get.

to AGI.

And Demis, in fact, we should share with listeners, Demis told us this when we were talking to him to prep for this episode.

Just felt like Larry got it.

Larry was completely on board with the mission of everything that DeepMind was doing.

And there's something else very convenient about Google.

They already have brain.

So Larry doesn't need Demis and Shane and Mustafa and DeepMind to come work on products within Google.

Right.

Brain is already working on products within Google.

Demis can really believe Larry when Larry says, nah, stay in London.

Keep working on intelligence.

Do what you're doing.

I don't need you to come work on products within Google.

Brain is like actively going and engaging with the product groups, trying to figure out, hey, how can we deploy neural nets into your product to make it better?

That's like their reason for being.

So they're happy to agree to this.

And it's working.

Brain and neural nets are getting integrated into search, into ads, into Gmail, into everything.

It is the perfect home.

for DeepMind.

Home away from home, shall we say.

Yes.

And there's a third reason why Google is the perfect fit for deep mind.

Infrastructure.

Google has all the compute infrastructure you could ever want right there on tap.

Yes, at least with CPUs so far.

Yes.

So how's the deal actually happen?

Well, after buying DNN research, Alan Eustis, who David you spoke with, right?

Yep.

Was Google's head of engineering at the time.

He makes up his mind that he wanted to hire.

all the best deep learning research talent that he possibly could and he had a clear path to do so a few months earlier earlier, Larry Page held a strategy meeting on an island in the South Pacific.

In Cade Metz's book, It's an Undisclosed Island.

Of course he did.

Larry thought that deep learning was going to completely change the whole industry.

And so he tells his team, this is a quote, let's really go big.

which effectively gave Alan a blank check to go secure all the best researchers that he possibly could.

So in 2013, he decides I'm going to get on a plane in December before the holidays and go meet DeepMind.

Crazy story about this.

Jeff Hinton, who's at Google at the time, had a thing with his back where he couldn't sit down he either has to stand or lay and so a long flight across the ocean is not doable but he needs to be there as a part of the diligence process you have jeff hinton you need to use him to figure out if you're going to buy a deep learning company and so alan eustis decides he's going to charter a private jet and he's going to build this crazy custom harness rig so that

jeff hinton won't be sliding around oh my god when he's laying on the floor during takeoff and landing.

Wow.

I was thinking the first part of this, I'm pretty sure Google has planes.

They could just get into Google Play.

For whatever reason, this was a separate charter.

But it's not solvable just with a private plane.

You need also a harness.

Right.

And Alan is the guy who set the record for jumping out of the world's highest, was it a balloon?

I actually don't know.

The highest free fall jump that anyone has ever done, even higher than that Red Bull stunt a few years before.

So he's like very used to designing these custom rigs for airplanes.

He's like, oh, no problem.

You just need a bed and some straps.

I jumped out of the atmosphere in a scuba suit.

I think we'll be fine.

That is amazing.

So they fly to London, they do the diligence, they make the deal.

Demis has true kinship with Larry, and it's done.

550 million US dollars.

There's an independent oversight board that is set up to make sure that the mission and goals of DeepMind are actually being followed.

And this is an asset that Google owns today that, again, I think is worth half a trillion dollars if it's independent.

Do you know what other member of the PayPal Mafia gets put on the ethics board after the acquisition?

Reed Hoffman?

Reed Hoffman.

Has to be, given the OpenAI tie later.

We are going to come back to Reed in just a little bit here.

Yes.

So after the acquisition, it goes very well, very quickly.

Famously, the data center cooling thing happens where DeepMind carved off some some part of the team to go and be an emissary to Google and look for ways to use DeepMind.

And one of them is around data center cooling.

Very quickly, July of 2016, Google announces a 40% reduction in the energy required to cool data centers.

I mean, Google's got a lot of data centers, a 40% energy reduction.

I actually talked with Jim Gao, who's a friend of the show, and actually led a big part of this project.

And I mean, it was just the most obvious application of neural networks inside of Google right away.

Pays for itself.

Yeah, imagine that paid for the acquisition pretty quickly there.

Yes.

David, should we talk about AlphaGo on this episode?

Yeah, yeah, yeah.

I watched the whole documentary that Google produced about it.

It's awesome.

This is actually something that you would enjoy watching, even if you're not researching a podcast episode and you're just looking to pull something up and spend an hour or two.

I highly recommend it.

It's on YouTube.

It's the story of how DeepMind post-acquisition from Google trained a model to beat the world Go champion at Go.

And I mean, everyone, the whole Go community coming in thought, there's no chance this guy, Lee Seedal, is so good that there's no way that an AI could possibly beat him.

It's a five-game thing.

It just won the first three games straight.

I mean, completely cleaned up and with inventive, new, creative moves that no human has played before.

That's sort of the big, crazy takeaway.

There's a moment in one of the games, right, where it makes a move of people like, is that a mistake?

I think that's just been an error.

Yeah, move 37.

Yeah, yeah.

yeah and then a hundred moves later it plays out and that it was like completely genius and humans are now learning from deep minds strategy of playing the game and discovering new strategies a fun thing for acquired listeners who are like why is it go

go is so complicated compared to chess chess has 20 moves that you can make at the beginning of the game in any given turn and then mid-game there's like 30 to 40 moves that you could make go on any given turn has about 200 and so if you think combinatorially, the number of possible configurations of the board is more than the number of atoms in the universe.

That's a great demis quote, by the way.

And so he says, even if you took all the computers in the world and ran them for a million years as of 2017, that wouldn't be enough compute power to calculate all the possible variations.

So it's cool because it's a problem that you can't brute force.

You have to do something like neural networks, and there is this white space to be creative and explore.

And so it served as this amazing breeding ground for watching a neural network be creative against a human.

Yep.

And of course, it's totally in with Demis' background and the DNA of the company of playing games.

Demis was chess champion.

And then after Go, then they play StarCraft, right?

Oh, really?

I actually didn't know that.

Yeah, that was the next game that they tackle was StarCraft, a real-time strategy game against an opponent.

And that'll

come back up in a sec with another opponent here in OpenAI.

Yes, David, but before we talk about the creation of the other opponent, should we thank another one of our friends here at Acquired?

Yes, we should.

All right, listeners, we are here today to tell you about a new friend of the show we are very excited about, WorkOS.

Yes, if you're building software that is used in enterprises, you've probably felt the pain of integrating things like SSO, SCIM, or SCIM, permissions, audit logs, and all the other features that are required by big customers.

And if you haven't felt this pain yet, just wait until you get your first big enterprise customer.

And trust us, you will.

Yes.

WorkOS turns these potential deal blockers into simple drop-in APIs.

And while WorkOS had great product market fit a few years ago with developers who just want to save on some headache, they really have become essential in the AI era.

Yeah.

I was...

shocked when they sent over their latest customer list.

Almost all the big AI companies use Work OS today as the way way that they've been able to rapidly scale revenue so fast.

Companies like OpenAI, Anthropic, Cursor, Perplexity, Sierra, Replit, Vercel, hundreds of other AI startups all rely on Work OS as their auth solution.

So I called the founder to ask why, and he said it's basically two things.

One, in the AI era, these companies scale so much faster that they need things like authorization, authentication, and SSO quickly to become enterprise ready and keep up with customer demand even early in life, unlike older SaaS companies of yesteryear.

And two, unlike that world where you could bring your own little SaaS product just for you and your little team, these AI products reach deep into your company's systems and data to become the most effective.

So IT departments are scrutinizing heavier than ever to make sure that new products are compliant before they can adopt them.

Yeah, it's this kind of like second order effect of the AI era that the days of, oh, just swipe a credit card, bring your own SaaS solution for your product team.

You actually need to be enterprise ready a lot sooner than you did before.

Yeah.

It's not just about picking up that big potential customer for the revenue itself either.

It's about doing it so your competitors don't.

Enterprise readiness has become so table stakes for companies no matter their stage.

And WorkOS is basically the weapon of choice for the best software companies to shortcut this process and get back to focusing on what makes their beer taste better, building the product itself.

Amen, amen.

So if you're ready to get started with just a few lines of code for SAML, Skim, RBOC, SSO, authorization, authentication, and everything else to please IT admins and their checklists, check out WorkOS.

It's the modern software platform to make all this happen.

That's workos.com.

And just tell them that Ben and David sent you.

All right, David.

So what are the second order effects of Google buying DMIN?

Well, there's one person who is really, really, really upset about this.

And maybe two people, if you include Mark Zuckerberg, but Mark tends to play his cards a little closer to the vest.

Of course, Elon Musk is very upset about this acquisition.

When Google buys DeepMind out from under him, Elon goes ballistic.

As we said, Elon and Larry had always been very close.

And now here's Google, who Elon has already started to sour on a little bit as he's now trying to hire AI researchers.

And you've got Alan Eustis flying around the world, sucking up all of the AI researchers into Google.

And Elon's invested in DeepMind, wanted to bring DeepMind into his own AI team at Tesla and gone out from under him.

So this leads to one of the most fateful dinners in Silicon Valley's history.

Organized in the summer of 2015.

at the Rosewood Hotel on Sandhill Road.

Of course, where else would you do a dinner in Silicon Valley, but the Rosewood by two of the most leading figures in the valley at the time, Elon Musk and Sam Altman.

Sam, of course, being president of Y Combinator at the time.

So what is the purpose of this dinner?

They are there

to make a pitch to all of the AI researchers.

that Google and to a certain extent Facebook have sucked up and basically created this duopoly status on.

Again, Google's business model and Facebook's business model, these feed recommenders or these classifiers turn out to be unbelievably valuable.

So they can, it's funny in hindsight saying this, pay tons of money to these people.

Tons of money, like millions of dollars.

Take them out of academia and put them into their dirty capitalist research labs inside the companies.

Selling advertising.

Yes.

How dirty could you be?

And the question and the pitch that Elon and Sam have for these researchers gathered at this dinner is,

what

would it take to get you out of Google for you to leave?

And the answer that goes around the table from almost everybody is

nothing.

You can't.

Why would we leave?

We're getting paid way more money than we ever imagined.

Many of us get to keep our academic positions and affiliations.

and we get to hang out here at Google.

With each other.

With each other.

Iron sharpens iron.

These are some of the best minds in the world getting to do cutting edge research with enormous amount of resources and hardware at their disposal.

It's amazing.

It's the best infrastructure in the world.

We've got Jeff Dean here.

There is nothing you could tell us

that would cause us to leave Google.

Except there's one person

who is intrigued.

And to quote from an amazing Wired article at the time by Cade Metz, who would later write Genius Makers, right?

Yep, exactly.

Quote is, the trouble was so many of the people most qualified to solve these problems were already working for Google.

And no one at the dinner was quite sure that these thinkers could be lured into a new startup, even if Musk and Maltman were behind it.

But one key player was at least open to the idea of jumping ship.

And then there's a quote from that key player.

I felt like there were risks involved, but I also felt like it would be a very interesting thing to try.

It's the most Ilya quote of all time.

The most illia quote of all time, because that person was Ilya Sutskeever,

of course, of AlexNet and DNN Research and Google and about to become founding chief scientist of Open AI.

So the pitch that Elon and Sam are making to these researchers is, let's start a new non-profit AI research lab where we can do all this work out in the open.

You can publish.

Free of the forces of Facebook and Google and independent of their control.

Yes, you don't have to work on products.

You can only work on research.

You can publish your work.

It will be open.

It will be for the good of humanity.

All of these incredible advances, this intelligence that we believe is to come will be for the good of everyone, not just for Google and Facebook.

And for money researchers, it seemed too good to be true.

So they basically weren't doing it because they didn't think anyone else would do it.

It's sort of an activation energy problem where once Ilya said, okay, I'm in.

And once he said, I'm in, by the way, Google came back with a big counter, something like double the offer.

And I think it was delivered from Jeff Dean personally.

And Ilya said, nope, I'm doing this.

That was massive for getting the rest of the top researchers to go with him.

And it was nowhere near all of the top researchers who left Google to do this, but it was enough.

It was a group of seven or so researchers who left Google and joined Elon and Sam and Greg Brackman from Stripe who came over to create open

AI.

Because that was the pitch.

We're all going to do this in the open.

And that's totally what it was.

It totally is what it was.

And the stated mission of Open AI was to, quote, advance digital intelligence in the way that is most likely to benefit humanity as a whole, unconstrained by a need to generate financial return.

Which is fine, as long as the thing that you need to fulfill your mission doesn't take tens of billions of dollars.

Yes.

So here's how they would fund it originally.

There was a billion dollars pledged.

Yes.

And that came from famously Elon Musk, Sam Altman, Reid Hoffman, Jessica Livingston, who I think most people don't realize was part of that initial tranche, and Peter Thiel.

Yep.

Founders Fund, of course, would go on to put massive amounts of money into OpenAI itself later as well.

The funny thing is, it was later reported that a billion dollars was not actually collected.

Only about 130 million of it was actually collected to fund this nonprofit.

And for the first few years, that was plenty for the type of research they were doing, the type of compute they needed.

Most of that money was going to paying salaries to the researchers.

Not as much as they could make at Google and Facebook, but still a million or two million dollars for these folks.

Right.

And yeah, so that really worked until it really didn't.

Yeah.

So David, what were they doing in the early days?

Well, in the first days, it was all hands on deck recruiting and hiring researchers.

And there was the initial crew that came over.

And then pretty quickly after that in early 2016, they get a big, big win when Dario Amadei leaves Google, comes over.

joins Ilya and crew at OpenAI.

Dream team, you know, assembling here.

And was he on Google Brain before this?

He was on Google Brain.

Yep.

And he, along with Ilya, would run large parts of OpenAI for the next couple of years before, of course, leaving to start Anthropic.

But we're still a couple years away from Anthropic, Claude, ChatGPT, Gemini, everything today.

For at least the first year or two,

basically the plan at OpenAI is let's look at what's happening at DeepMind and show the research community that we can do, as a new lab, do the same incredible things that they're doing and maybe even do them better.

Is that why it looks so game-like and game-focused?

Yes.

Yes.

So they start building models to play games.

Famously, the big one that they do is Dota 2, Defense of the Agents 2, the massively online battle arena video game.

They're like, all right, well.

DeepMind, you're playing StarCraft.

Well, we'll go play Dota 2.

That's even more complex, more real-time.

And similar to the emergent properties of Go, the game would devise unique strategies that you wouldn't see humans trying.

So it clearly wasn't humans coded their favorite strategies and rules in.

It was emergent.

Yep.

They did other things.

They had a product called Universe, which was around training computers to play thousands of games from Atari games to open world games like Grand Theft Auto.

They had something where they were teaching a model how to do a Rubik's Cube.

And so it was a diverse set of projects that didn't seem to coalesce around one of these is going to be the big thing.

Yep.

It was research stuff.

It was what DeepMind was doing.

Yeah.

It was like a university research.

It was like DeepMind.

And if you think back to Elon being an investor in DeepMind, being really upset about Google acquiring it out from under him, makes sense.

And I think Elon deserves a lot of credit for having his name and his time attached to OpenAI at the beginning.

A lot of the big heavy hitter recruiting was Elon throwing his weight behind this.

I'm willing to take a chance.

Absolutely.

Okay, so that's what's going on over at OpenAI, doing a lot of DeepMind-like stuff, bunch of projects, not one single obvious big thing they're coalescing around.

It's not chat GPT time.

Let's put it that way.

Let's go back to Google because last we sort of checked in on them.

Yeah, they bought DeepMind, but they had their talent rated.

And I don't want you to get the wrong impression about where Google is sitting just because some people left to go to OpenAI.

So back in 2013, when Alex Kruszewski arrives at Google with Jeff Hinton and Ilya Sitskeever, he was shocked to discover that all their existing machine learning models were running on CPUs.

People had asked in the past for GPUs since machine learning workloads were well-suited to run in parallel, but Google's infrastructure team had pushed back and said, the added complexity and expanding and diversifying the fleet, let's keep things simple.

That doesn't seem important for us.

We're a CPU shop here.

Yes.

And so to quote from Genius Makers, in his first days at the company, he went out and bought a GPU machine, this is Alex, from a local electronic store, stuck it in the closet down the hall from his desk, plugged it into the network, and started training his neural networks on this lone piece of hardware.

Just like he did in academia, except this time, Google's paying for the electricity.

Obviously, one GPU was not sufficient, especially as more Googlers wanted to start using it too.

And Jeff Dean and Alan Eustis had also also come to the conclusion that diss belief, while amazing, had to be re-architected to run on GPUs and not CPUs.

So spring of 2014 rolls around.

Jeff Dean and John Geandra, who we haven't talked about this episode.

Yeah, JG.

Yes, you might be wondering, wait, isn't that the Apple guy?

Yes, he went on to be Apple's head of AI, who at this point in time was at Google and oversaw Google Brain 2014.

They sit down to make a plan for how to actually formally put GPUs into the fleet of Google's data centers, which is a big deal.

It's a big change.

But they're seeing enough reactions to neural networks that they know to do this.

Yeah, after Alex Nett is just a matter of time.

Yeah.

So they settle on a plan to order 40,000 GPUs

from NVIDIA.

Yeah, of course.

Who else are you going to order them from?

For a cost of $130 million.

That's a big enough price tag that the request gets elevated to Larry Page, who personally approves it, even though finance wanted to kill it, because he goes, look, the feature of Google is deep learning.

As an aside, let's look at NVIDIA at the time.

This is a giant, giant order.

Their total revenue was $4 billion.

This is one order for 130 million.

I mean, NVIDIA is primarily a consumer graphics card company at this point.

Yes, and their market cap is $10 billion.

It's almost like Google gave NVIDIA a secret that, hey, not only does this this work in research like the ImageNet competition, but neural networks are valuable enough to us as a business to make a $100 plus million dollar investment in right now, no questions asked.

We got to ask Jensen about this at some point.

This had to be a tell.

This had to really give NVIDIA the confidence, oh, we should way forward invest on this being a giant thing in the future.

So all of Google wakes up to this idea.

They start really putting it into their products.

Google Google Photos happened.

Gmail starts offering typing suggestions.

David, as you pointed out earlier, Google's giant AdWords business started finding more ways to make more money with deep learning.

In particular, when they integrated it, they could start predicting what ads people would click in the future.

And so Google started spending hundreds of millions more on GPUs on top of that 130 million, but very quickly paying it back from their ad system.

So it became more and more of a no-brainer to just buy as many GPUs as they possibly could.

But once neural nets started to work, anyone using them, especially at Google Scale, kind of had this problem.

Well, now we need to do giant amounts of matrix multiplications.

Anytime anybody wants to use one, the matrix multiplications are effectively how you do that propagation through the layers of the neural network.

So you sort of have this problem.

Yes, totally.

There's the inefficiency of it, but then there's also the business problem of, wait a minute, it looks like we're just going to be shipping hundreds of millions, millions, soon to be billions of dollars over to NVIDIA every year for the foreseeable future.

Right.

So there's this amazing moment right after Google rolls out speech recognition, their latest use case for neural nets, just on Nexus phones.

Because again, they don't have the infrastructure to support it on all Android phones.

It becomes a super popular feature.

And Jeff Dean does the math and figures out if people use this for, I don't know, call it three minutes a day, and we roll it out to all billion Android phones, We're going to need twice the number of data centers that we currently have across all of Google just to handle it.

Just for this feature.

Yeah.

There's a great quote where Jeff goes to Erz Holtzl and goes, We need another Google.

Or David, as you were hinting at, the other option is we build a new type of chip customized for just

our particular use case.

Yep.

Matrix multiplication, multiplication, tensor multiplication, a tensor processing unit, you might say.

Ah, yes.

Wouldn't that be nice?

So conveniently, Jonathan Ross, who's an engineer at Google, has been spending his 20% time at this point in history working on an effort involving FPGAs.

These are essentially expensive but programmable chips that yield really fantastic results.

So they decide to create a formal project to take that work, combine it with some other existing work, and build a custom ASIC or an application-specific integrated circuit.

So enter, David, as you said, the tensor processing unit made just for neural networks that is far more efficient from GPUs at the time, with the trade-off that you can't really use it for anything else.

It's not good for graphics processing.

It's not good for lots of other GPU workloads, just matrix multiplication and just neural networks.

But it would enable Google to scale their data centers without having to double their entire footprint.

So the big idea behind the TPU, if you're trying to figure out like what was the core insight, they use reduced computational precision.

So it would take numbers like 4,586.8272 and round it just to 4,586.8, or maybe even just 4,586 with nothing after the decimal point.

And this sounds kind of counterintuitive at first.

Why would you want less precise rounded numbers for this complicated math?

The answer is efficiency.

If you can do the heavy lifting in your software architecture, or what's called quantization to account for it, you can store information as less precise numbers.

Then you can use the same amount of power and the same amount of memory and the same amount of transistors on a chip to do far more calculations per second.

So you can either spit out answers faster or use bigger models.

The whole thing is quite clever behind the TPU.

The other thing that has to happen with the TPU is it needs to happen now.

Because it's very clear speech to text is a thing.

It's very clear some of these other use cases at Google.

Yeah.

Demand for all of this stuff that's coming out of Google brain is through the roof immediately.

Right.

And we're not even two LLMs yet.

It's just like everyone sort of expects some of this, whether it's computer vision and photos or speech recognition, like it's just becoming a thing that we expect.

And it's going to flip Google's economics upside down if they don't have it.

So the TPU was designed, verified, built, and deployed into data centers in 15 months.

Wow.

It was not like a research project that could just happen over several years.

This was like a hair on fire problem that they launched immediately.

One very clever thing that they did was A, they used the FPGAs as a stopgap.

So even though they were like too expensive on a unit basis, they could get them out as a test fleet and just make sure all the math worked before they actually had the ASICs printed at, I don't know if it was a TSMC, but, you know, fabbed and ready.

The other thing they did is they fit the TPU into the form factor of a hard drive.

So it could actually slot into the existing server racks.

You just pop out a hard drive and you pop in a TPU without needing to do any physical re-architecture.

Wow.

That's amazing.

That's the most googly infrastructure story since the corkboards.

Exactly.

Also, all of this didn't happen in Mountain View.

It was at a Google satellite office in Madison, Wisconsin.

Whoa.

Yes.

Why Madison, Wisconsin?

There was a particular professor out of the university, and there was a lot of students that they could recruit from.

Wow.

Yeah.

I mean, it was probably them or Epic.

Where are you going to go work?

yeah

wow they also then just kept this a secret right why would you tell anybody about this because it's not like they're offering these in google cloud at least at first and why would you want to tell the rest of the world what you're doing so the whole thing was a complete secret for at least a year before they announced it at google io so really crazy the other thing to know about the tpus is they were done in time for the alpha go match.

So that match ran on a single machine with four TPUs in Google Cloud.

And once that worked, obviously that gave Google a little bit of extra confidence to go really, really reap production.

So that's the TPU.

V1, by all accounts, was not great.

They're on V7 or V8 now.

It's gotten much better.

TPUs and GPUs look a lot more similar than they used to, than they've sort of adopted features from each other.

But today, Google, it's estimated, has two to three million TPUs.

For reference, NVIDIA shipped, people don't know for sure, somewhere around 4 million GPUs last year.

So people talk about AI chips like it's this just, oh, one horse race with NVIDIA.

Google has like an almost NVIDIA scale internal thing making their own chips at this point for their own and for Google Cloud customers.

The TPU is a giant deal in AI in a way that I think a lot of people don't realize.

Yep.

This is one of the great ironies and maddening things to OpenAI and Elon Musk is that OpenAI gets founded in 2015 with the goal of, hey, let's shake all this talent out of Google and level the playing field.

And Google just accelerates.

Right.

They also build TensorFlow.

That's the framework that Google Brain built to enable researchers to build and train and deploy machine learning models.

And they built it in such a way that it doesn't just have to run on TPUs.

It's super portable without any rewrites to run on GPUs or even CPUs too.

So this would replace the old DIST belief system and kind of be their internal and external framework for enabling ML researchers going forward.

So, somewhat paradoxically, during these years after the founding of OpenAI,

yes, some amazing researchers are getting siphoned off from Google and Google Brain, but Google Brain is also firing on all cylinders during this timeframe.

Delivering on the business purposes for Google left and right.

Yes, and pushing the state of the art forward in so many areas.

And then in 2017, a paper gets published from eight researchers on the Google brain team.

Kind of quietly.

These eight folks were obviously very excited about the paper and what it described and the implications of it, and they thought it would be very big.

Google itself,

cool.

This is like the next iteration of our language model work.

Great.

Which is important to us.

But are we sure this is the next Google?

No.

No, there are a whole bunch of other things we're working on that seem more likely to be the next Google.

But this paper and its publication would actually be what gave OpenAI the opportunity to build the next Google.

To grab the ball and run with it and build the next Google.

Because this is the Transformer paper.

Okay, so where did the Transformer come from?

Like what was the latest thing that language models had been doing at Google?

So coming out of the success of Franz Ox's work on Google Translate and the improvements that happened there.

In like the late 2000s-ish, 2007?

Yeah, mid to late 2000s.

They keep iterating on translate.

And then once Jeff Hitton comes on board and AlexNet happens, they switch over to a neural network-based language model for translate.

Which was dramatically better and like a big, crazy cultural thing because you've got these researchers.

parachuting in, again, led by Jeff Dean, saying, I'm pretty sure our neural networks can do this way better than the classic methods that we've been using for the last 10 years.

What if we take the next several months and do a proof of concept?

They end up throwing away the entire old code base and just completely wholesale switching to this neural network.

There's actually this great New York Times magazine story that ran in 2016 about it.

And I remember reading the whole thing with my jaw on the floor, like, wow, neural networks are a big effing deal.

And this was the year before the Transformer paper would come out.

Before the Transformer Transformer paper.

Yes.

So they do the rewrite of Google Translate, make it based on recurrent neural networks, which were state-of-the-art at that point in time.

And it's a big improvement.

But as teams within Google Brain and Google Translate keep working on it, there's some limitations.

And in particular, a big problem was that they quote unquote forgot things too quickly.

I don't know if it's exactly the right analogy, but you might say in sort of like today's Transformer world speak, you might say that their context window was pretty short.

As these language models progressed through text, they needed to sort of remember everything they had read so that when they need to change a word later or come up with the next word, they could have a whole memory of the body of text to do that.

So one of the ways that Google tries to improve this is to use something called long short-term memory networks or LSTMs as the acronym that people use for this.

And basically, what LSTMs do is they create a persistent or long

short-term memory.

You got to use your brain a little bit here for the model so that it can keep context as it's going through a whole bunch of steps.

And people were pretty excited about LSTMs at first.

People are thinking, like, oh, LSTMs are what are going to take language models and large language models mainstream.

Right.

And indeed, in 2016, they incorporated into Google Translate these LSTMs.

It reduces the error rate by 60%.

Huge jump.

Yep.

The problem with LSTMs, though, they were effective, but they were very computationally intensive.

And they didn't parallelize that great.

All the efforts that are coming out of AlexNet and then the TPU project of parallelization.

This is the future.

This is how we're going to make AI really work.

LSTMs are a bit of a a roadblock here.

Yes.

So a team within Google Brain starts searching for a better architecture that also has the attractive properties of LSTMs, that it doesn't forget context too quickly, but can parallelize and scale better.

To take advantage of all these new architectures.

Yes.

And a researcher named Jakob Oskarait had been toying around with the idea of broadening the scope of quote-unquote attention in language processing.

What if, rather than focusing on the immediate words, instead, what if you told the model, hey, pay attention to the entire corpus of text, not just the next few words.

Look at the whole thing, and then based on that entire context and giving your attention to the entire context, Give me a prediction of what the next translated word should be.

Now, by the way, this is actually how professional human translators translate text.

You don't just go word by word.

I actually took a translation class in college, which was really fun.

You read the whole thing of the original in the original language, you get and understand the context of what the original work is,

and then you go back and you start to translate it with the entire context of the passage in mind.

So, it would take a lot of computing power for the model to do this,

but it is extremely parallelizable.

So Jakob starts collaborating with a few other people on the brain team.

They get excited about this.

They decide that they're going to call this new technique the transformer.

Because one, that is literally what it's doing.

It's taking in a whole chunk of information, processing, understanding it, and then transforming it.

And B, they also love transformers as kids.

That's not why they named it the Transformer.

And it's taking in the giant corpus of text and storing it in a compressed format, right?

Yeah.

I bring this up because that is exactly how you pitched the micro kitchen conversation with Noam Shazir in 2000, 2001, 17 years earlier.

Who is a co-author on this paper?

Yes.

Well, so speaking of Noam Shazir, he learns about this project and he decides, hey, I've got some experience with this.

This sounds pretty cool.

LSTMs definitely have problems.

This could be promising.

I'm going to jump in and work on it with these guys.

And it's a good thing he did.

Because before Gnome joined the project, they had a working implementation of the Transformer, but it wasn't actually producing any better results than LSTMs.

Gnome joins the team, basically pulls a Jeff Dean,

rewrites the entire code base from scratch.

And when he's done, the Transformer now crushes the LSTM-based Google Translate solution.

And it turns out that the bigger they make the model, the better the results get.

It seems to scale really, really, really well.

Stephen Levy wrote a piece in Wired about the history of this.

And there are all sorts of quotes from the other members of the team just littered all over this piece with things like, Gnome is a magician.

Gnome is a wizard.

Noam took the idea and came back and said, it works now.

Yeah.

And you wonder why Noam and Jeff Dean are the ones together working on the next version of Gemini now.

Yes.

Noam and Jeff Dean are definitely two peas in a pot here.

Yes.

So we talked to Greg Corrado from Google Brain, one of the founders of Google Brain, and it was a really interesting conversation because he underscored how elegant the Transformer was.

And he said it was so elegant that people's response was often, this can't work.

It's too simple.

Transformers are barely a neural network architecture.

Right.

It was another big change from the AlexNet, Jeff Hinton lineage neural networks.

Yeah.

It actually has changed the way that I look at the world because he pointed out that in nature, this is Greg, the way things usually work is the most energy-efficient way they could work, almost from an evolution perspective, that the most simple, elegant solutions are the ones that survive because they are the most efficient with their resources.

And you can kind of port this idea over to computer science too, that he said he's developed a pattern recognition inside of the research lab to realize that you're probably onto the right solution when it's really simple and really efficient versus a complex idea.

It's very clever.

I think it's very true.

You know how when you sit around and you have a thorny problem and you debate and you whiteboard and you come up with all, and then you're like, oh my God.

Oh my God, it's so simple.

And that ends up being the right answer.

Yeah, there's an elegance to the transformer.

Yes.

And that other thing that you touched on there, this is the beginning of the modern AI.

Just feed it more data.

The famous piece, The Bitter Lesson by Rich Sutton, wouldn't be published until 2019.

For anyone who hasn't read it, it's basically, we always think as AI researchers, we're so smart and our job is to come up with another great algorithm.

But effectively, in every field from language to computer vision to chess, you just figure out a scalable architecture and then the more data wins.

Just these infinitely scaling more data, more compute, better results.

Yes.

And this is really the start of when that starts to be like, oh, we have found the scalable architecture that will go so far for, I don't know, close to a decade of just more data in, more energy, more compute, better results.

So the team and Noam.

They're like, yo, this thing has a lot, a lot of potential.

This is more than better translate.

We can really apply this.

Yeah, this is going to be more than better Google Translate.

The rest of Google, though, definitely slower to wake up to the potential.

They build some stuff.

Within a year, they build BERT, the large language model.

Yes, absolutely true.

It is a false narrative out there that Google did nothing with the Transformer after the paper was published.

They actually did a lot.

In fact, BERT was one of the first LLMs.

Yes, they did a lot with Transformer-based large language models after the paper came out.

What they didn't do was treat it as a wholesale technology platform change.

Right.

They were doing things like BERT and MOM, this other model, you know, they could work it into search results quality.

And I think that did meaningfully move the needle, even though Google wasn't bragging about it and talking about it.

They got better at query comprehension.

They were working it into the core business, just like every other time Google brain came up with something great.

Yep.

So, in perhaps one of the greatest decisions ever for value to humanity, and maybe one of the worst corporate decisions ever for Google, Google allows this group of eight researchers to publish the paper under the title, Attention is All You Need, obviously, a nod to the classic Beatlesong about love.

As of today, in 2025,

This paper has been cited over 173,000 times in other academic papers, making it currently the seventh most cited paper of the 21st century.

And I think all of the other papers above it on the list have been out much longer.

Wow.

And also, of course, within a couple of years, all eight authors of the Transformer paper had left Google to either start or join AI startups, including OpenAI.

Brutal.

And of course, Noam starting Character AI, which what are we calling it, a hacquisition?

He would end up back at Google via some strange licensing and IP and hiring agreement on the few billion dollars order.

Very, very expensive mistake on Google's part.

It is fair to say that 2017 begins the five-year period of Google not sufficiently seizing the opportunity that they had created.

With the Transformer.

Yes.

So speaking of seizing opportunities, what is going on at OpenAI during this time?

And does anyone think the Transformer is a big deal over there?

Yes.

Yes, they did.

But here's where history gets really, really crazy.

Right after

Google publishes the Transformer paper, in September of 2017,

Elon gets really, really fed up with what's going on at OpenAI.

There's like seven different strategies.

Are we doing video games?

Are we doing competitions?

What's the plan?

What is happening here, as best as I can tell, all you're doing is just trying to copy DeepMind.

Meanwhile, I'm here building SpaceX and Tesla.

Self-driving is becoming more and more clear as critical to the future of Tesla.

I need AI researchers here and I need great AI advancements to come out to help what we're doing at Tesla.

OpenAI isn't cutting it.

So he makes an ultimatum to Sam and the rest of the OpenAI board.

He says, I'm happy to take full control of OpenAI and we can merge this into Tesla.

I don't even know how that would be possible to merge a nonprofit into Tesla.

But in Elon Land, if he takes over as CEO of OpenAI, it almost doesn't matter.

We're just treating it as if it's the same company anyway, just like we do with the deals with all of my companies.

Right.

Or he's out completely, along with all of his funding.

And Sam and the rest of the board are like, no.

And as we know now, they're sort of calling capital into the business.

It's not like they actually got all the cash up front.

Right.

So they're only $130 million-ish into the billion dollars of commitment.

They don't reach a resolution.

And by early 2018, Elon is out,

along with him, the main source of OpenAI's funding.

So either this is just a really, really, really bad misjudgment by Elon.

Or

the sort of panic that this throws OpenAI into is the catalyst that makes them reach for the transformer and say,

all right, we got to figure things out.

Necessity is the mother of invention.

Let's go for it.

It's true.

I don't know if during this personal tension between Elon and Sam, if they had already decided to go all in on Transformers or not, because the thing you very quickly get to, if you decide, Transformers, language models, we're going all in on that.

You do quickly realize you need a bunch of data, you need a bunch of compute, you need a bunch of energy, and you need a bunch of capital.

And so if your biggest backer is walking away, the 3D chess move is, oh, we got to keep him because we're about to pivot the company and we need his capital for this big pivot we're doing.

The 4D chess is, if he walks away, maybe I can turn it into a for-profit company.

and then raise money into it and eventually generate enough profits to fund this extremely expensive new direction we're going in.

I don't know which of those it was.

Yeah.

I don't know either.

I suspect the truth is it's some of both.

Yes.

But either way, how nuts is it that A, these things happened at the same time and B, the company wasn't burning that much cash and then they decided to go all in on

we need to do something so expensive that we need to be a for-profit company in order to actually achieve this mission because it's just going to require hundreds of billions of dollars for the far foreseeable future.

Yep.

So in June of 2018, OpenAI releases a paper describing how they have taken the Transformer and developed a new approach of pre-training them on very large amounts of general text on the internet and then fine-tuning that general pre-training to specific use cases.

And they also announced that they have trained and run the first proof of concept model of this approach, which they are calling GPT1,

Generatively Pre-Trained Transformer Version 1.

Which, we should say, is right around the same time as BERT and right around the same time as another large language model based on the Transformer out of here in Seattle, the Allen Institute.

Yes.

Indeed.

So it's not as if this is heretical and a secret.

Other AI labs, including Google's own, is doing it.

But from the very beginning, OpenAI seemed to be taking this more seriously, given the cost of it would require betting the company if they continued down this path.

Yep, or betting the nonprofit, betting the entity.

Yes.

We're going to need some new terminology here.

Yes.

So Elon's just walked out the door.

Where are they going to get the money for this?

Sam turns to one of the other board members of OpenAI, Reid Hoffman.

Reed, just a year or so earlier, had sold LinkedIn to Microsoft, and Reed is now on the board of Microsoft.

So Reed says, hey, why don't you come talk to Satya about this?

Do you know where he actually talks to Satya?

Oh, I do.

Oh, I do.

In July of 2018, they set a meeting for Sam Altman and Satya Nadella to sit down while they're both at the Allen and Company Sun Valley conference in Sun Valley, Idaho.

That's perfect.

And while they're there, they hash out a deal for Microsoft to invest $1 billion

into OpenAI in a combination of both cash and Azure cloud credits.

And in return, Microsoft will get access to OpenAI's technology, get an exclusive license to OpenAI's technology for use in Microsoft's products.

And the way that they will do this is OpenAI, the nonprofit, will create a captive for-profit entity called OpenAI LP, controlled by the nonprofit OpenAI Inc.

And Microsoft will invest into the captive for-profit entity.

Reid Hoffman joins the board of this new structure along with Sam, Ilya, Greg Brockman, Adam D'Angelo, and Tasha Macaulay.

And thus the modern OpenAI for-profit, non-profit question mark is created.

The thing that's still being figured out, even today, here in 2025, is created.

This is like the complete history of AI.

This is not just the Google AI episode.

Well, these things are totally inextricable.

And I was just going to say, this is the Google Part 3 episode.

Microsoft, they're back.

Microsoft is Google's mortal enemy.

Yes.

That in our first episode on the founding of Google and search, and then in the second episode on Alphabet and all the products that they made, the whole strategy at Google was always about Microsoft.

They finally beat them on every single front.

And here they are showing up again saying, what was Satya's line?

We just want to see them dance.

I think the line that would come a couple of years later is, we want the world to know that we made Google dance.

Oh, man.

But this is all still pre-Chat GPT.

This is just Sam lining up the financing he needs for what appears to be a very expensive scaling exercise they're about to embark on with GPT2 and onward.

Yep.

And this is the right time to talk about why,

from OpenAI's perspective, Microsoft is the absolute perfect partner.

It's not just that they have a lot of money.

Although that helps.

I mean, that helps.

That helps a lot.

But more important than money, they have a really, really great public cloud.

Azure.

Yes.

OpenAI is not going to go buy a bunch of NVIDIA GPUs and then build their own data center here at this point in 2018.

That's not the scale of company that they are.

They need a cloud provider in order to actually do all the compute that they want to do.

If they were back at Google and these researchers are doing it, great.

Then they have all the infrastructure.

But OpenAI needs to tie themselves to someone with the infrastructure.

And there's basically only two non-Google options.

They're both in Seattle.

And hey, one of them in Microsoft is really interested, also has a lot of cash.

This seems like a great partnership.

That's true.

I wonder if they did talk to AWS at all about it.

Because I think, this is a crazy Easter egg.

I hesitate to say it out loud, but I think AWS was actually in the very first investment with Elon in OpenAI.

Oh, wow.

And I don't know if it was in the form of credits or what the deal was, but I'd seen it reported a couple places that AWS actually was in that nonprofit round.

Yeah, in the nonprofit funding, the donations to

the early OpenAI.

Anyway, Microsoft, OpenAI, they end up tying up.

A match made in heaven.

Satya and Sam are on stage together talking about how this amazing partnership and marriage has come together and they're off to model training.

Yeah.

And this paves the way for the GPT era of OpenAI.

But before we tell that story.

Yes.

Now is a great time to thank one of our favorite companies, Shopify.

Yes.

And this is really fun because we have been friends and fans of Shopify for years.

We just had Toby on ACQ2 to talk about everything going on in AI and everything that has happened at Shopify in the six years now since we covered the company on acquired.

It's been a pretty insane transformation for them.

Yeah.

So back at their IPO, Shopify was the go-to platform for entrepreneurs and small businesses to get online.

What's happened since is that is still true.

And Shopify has also become the world's leading commerce platform for enterprises of any size, period.

Yeah.

So what's so cool about the company is how they've managed to scale without losing their soul.

Even though companies like Everlane and Vouri and even older established companies like Mattel are doing billions of revenue on Shopify, the company's mission is still the same as the day Toby founded it, to create a world where more entrepreneurs exist.

Oh, yeah.

Ben, you got to tell everyone your favorite enterprise brand that is on Shopify.

Oh, I'm saving that for next episode.

I have a whole thing planned for episode two of this season.

Okay, okay, great.

Anyway, the reason enterprises are now also using Shopify is simple, because businesses of all sizes just sell more with Shopify.

They've built this incredible ecosystem where you can sell everywhere.

Obviously, your own site, that's always been true.

But now with Shopify, you can easily sell on Instagram, YouTube, TikTok, Roblox, Roku, ChatGPT, Perplexity, anywhere.

Plus, with ShopPay, their accelerated checkout, you get amazing conversion and it has a built-in user base of 200 million people who have their payment information already stored with it.

Shopify is the ultimate example of not doing what doesn't make your beer taste better.

Even if you're a huge brand, you're not going to build a better e-commerce platform for your product.

But that is what Toby and Shopify's entire purpose is.

So you should use them.

Yes.

So whether you're just getting started or already at huge scale, head on over to shopify.com/slash acquired.

That's S-H-O-P-I-F-Y dot com slash acquired.

And just tell them that Ben and David sent you.

All right.

So what are we in GPT2?

Is that what's being trained right here?

Yes.

GPT2.

This was the first time I heard about it.

Data scientists around Seattle were talking about this cool.

Right.

So after the first Microsoft partnership, the first billion dollar investment, in 2019, OpenAI releases GPT-2,

which is still early, but very promising, that can do a lot of things.

A lot of things, but it required an enormous amount of creativity on your part.

You kind of had to be a developer to use it.

And if you were a consumer, there was a very heavy load put on you.

You had to go write a few paragraphs and then paste those few paragraphs into the language model.

And then it would suggest a way to finish what you were writing based on the source paragraphs.

But it wasn't interactive.

Yes, it was not a chat interface.

Yes.

There was no interface essentially for it.

It was an API.

But it can do things like obviously translate text.

I mean, Google's been doing that for a long time, but GPT-2, you could do stuff like make up a fake news headline and give it to GPT-2, and it would write a whole article.

You would read it and you'd be like, uh,

sounds like it was written by a bot.

Yeah.

But again, there was no front door to it for normal people.

You had to really be willing to wait in the muck to use this thing.

So then the next year in June of 2020, GPT-3 comes out.

Still no front door, you know, user interface to the model, but it's very good.

GPT-2 showed the promise of what was possible.

GPT-3,

it's starting to be in the conversation of can this thing pass the Turing test?

Oh, yeah.

You have a hard time distinguishing between articles that GPT wrote and articles that humans wrote.

It's very good, and there starts to be a lot of hype around this thing.

And so, even though consumers aren't really using it, the broader awareness is that there's something interesting on the horizon.

I think the number of AI pitch decks that VCs are seeing is starting to tick up around this time.

As is the NVIDIA stock price.

Yes.

So, then in the next year, in the summer of 2021,

Microsoft releases GitHub Copilot using GPT-3.

This is the first, not just Microsoft product that comes out with GPT baked into it, but first productization.

Product anywhere.

Yeah, first productization of GPT.

Yes, of any open AI technology.

Yeah, it's big.

This starts a massive change in how software gets written in the world.

Slowly then all at once.

It's one of these things where at first, just a few software engineers and there was a lot of whispers of, how cool is this?

It makes me a little bit more efficient.

And now you get all these comments like 75% of all companies' code is written with AI.

Yep.

So after that, Microsoft invests another $2 billion in OpenAI, which seemed like a lot of money at the time.

So that takes us to the end of 2021.

There's an interesting kind of context shift that happens around here.

Yeah, the bottom falls out on tech stocks, crypto, the broader markets, really.

Everyone suddenly goes from risk on to risk off.

And part of it was war in Ukraine, but a lot of it was interest rates going up.

And Google gets hit really hard.

The high watermark was November 19th of 2021.

Google was right at $2 trillion of market cap.

About a year after that slide began, they were worth a trillion dollars, nearly a 50% drawdown.

Wow.

So towards the end of 2022, leading up to the launch of ChatGPT, people, I think, are starting to realize Google's slow.

They're slow to react to things.

It feels like they're a old, crusty company.

Are they like the Microsoft of the 2000s where they haven't had a breakthrough product in a while?

People are not bright on the future of Google.

And then ChatGPT comes out.

Yeah.

Wow.

Which means if you were bullish on Google back then and contrarian, you could have invested at a trillion dollar market cap.

Which is interesting.

Like in October of 21, the market was saying that the forthcoming AI wave will not be a strength for Google.

Or maybe what it was saying is we don't even know anything about a forthcoming AI wave because people are talking about AI, but they've been talking about VR and they've been talking about crypto and they've been talking about all this frontier tech.

And like, that's not the future at all.

This company just feels slow and unadaptive.

And slow and unadaptive at that point in history, I think would have been a fair characterization.

They had an internal chatbot, right?

Yes, they did.

All right.

So

before we talk about ChatGPT,

Google had a chatbot.

So Noam Shazir, incredible engineer, re-architected the Transformer, made it work, one of the lead authors of the paper, storied career within Google, has all of this sway, should have all of this sway within the company.

After the Transformer paper comes out, he and the rest of the team are like, guys,

we can use this for a lot more than Google Translate.

And in fact, the last paragraph of the paper.

Are you about to read the Transformer paper?

Yes, I am.

We are excited about the future of attention-based models and plan to apply them to other tasks.

We plan to extend the Transformer to problems involving input and output modalities other than text and to investigate large inputs and outputs such as images, audio, and video.

This is in the paper.

Wow.

Google obviously does not do any of that for quite a while.

GNOME, though, immediately starts advocating to Google leadership, hey, I think this is going to be so big, the Transformer, that we should actually consider just throwing out the search index and the 10 blue links model and go all in on transforming all of Google into one giant transformer model.

And then Gnome actually goes ahead and builds a chatbot interface to a large transformer model.

Is this Lambda?

This is before Lambda, Mina is what he calls it.

And there is a chatbot in the like late teens, 2020 timeframe that Gnome has built within Google that arguably is pretty close to chat GPT.

Now, it doesn't have any of the post-training safety that chat GPT does.

So it would go off the rails.

Yeah, someone told us that you could just ask it who should die and it would come up with names for you of of people that should die.

It was not a shippable product.

It was a very

raw, not safe, not post-trained chatbot and model.

Right.

But it existed within Google and they didn't ship it.

And technically, not only did it not have post-training, it didn't have RLHF either, this very core component of the models today, the reinforcement learning with human feedback that ChatGPT, I don't know if it had it in three, but it did in 3.5 and it did for the launch of ChatGPT.

Realistically, it wasn't launchable, even if it was an open AI thing, because it was so bad.

But a company of Google stature certainly could not take the risk.

So strategically, they have this working against them.

But aside from the strategy thing, there's two business model problems here.

One, if you're proposing to drop the 10 blue links and just turn google.com into a giant AI chatbot, Revenue drops when you provide direct answers to questions versus showing advertisers and letting people click through to websites.

That upsets the whole Apple cart.

Obviously, they're thinking about it now, but until 2021, that was an absolute non-starter to suggest something like that.

Two, there were legal risks of sitting in between publishers and users.

I mean, Google at this point had spent decades fighting the public perception and court rulings that they were disintermediating publishers from readers.

So there was like a very high bar internally, culturally, to clear if you were going to do something like this.

Even those info boxes that popped up that took until the 20 teens to make it happen, those really were mostly on non-monetizable queries anyway.

So anytime that you were going to say, hey, Google's going to provide you an answer instead of 10 blue links, you had to have a bulletproof case for it.

Yep.

And there was also a brand promise and trust issue too.

Consumers trusted Google so much.

For us, even today, you know, when I'm doing research for acquired, we need to make sure we get something right.

I'm going to Google.

I look something up in Claude.

Yeah.

It gives me an answer.

I'm like, that's a really good answer.

And then I verify by searching Google that I can find those facts too if I can't click through the sources on Claude.

That's my workflow.

Which sort of sounds funny today, but it's important.

If you're going to propose replacing the 10 blue links with a chatbot, you need to be really damn sure that it's going to be accurate.

Yes.

And

in 2020, 2021, that was definitely not the case.

Arguably, still isn't the case today.

And there also wasn't a compelling reason to do it because

nobody was really asking for this product.

Right.

No knew and people in Google knew that you could make a chatbot interface to a transformer-based LLM.

And that was a really compelling product.

The general public didn't know.

OpenAI didn't even really know.

I mean, GPT was out there.

Do you know the story of the launch of ChatGPT?

Well, I think I do.

I have it in my notes here.

All right.

So they've got GPT 3.5.

It's becoming very, very useful.

Yeah, this is late 2022.

They've got 3.5.

But there's still this problem of how am I supposed to actually use it?

How does it productized?

And Sam just kind of says, we should make a chat bot.

That seems like a natural interface for this.

Can someone just make a chat?

And within like a week, internally, someone makes a chat.

They just turn calls to the chat GPT 3.5 API into a product where you're just chatting with it.

And every time you kick off a chat message, it just calls GPT 3.5 on the API.

And that turns out to be this magic product.

I don't think they expected it.

I mean, servers are tipping over.

They're working with Microsoft to try to get more compute.

They're cutting deals with Microsoft in real time to try to get more investment, to get more Azure credits or get advances on their Azure credits in order to handle the incredible load in November of 2022 that's coming in of people wanting to use this thing.

They also just throw up a paywall randomly because they thought that the business was going to be an API business.

They thought that the projections were all about how much revenue they were going to do through B2B licensing deals.

And then they just realized, oh, there's all these consumers trying to use this, put up a paywall to at least dampen the most expensive use of this thing so we can kind of offset the costs or slow the rollout.

Right.

This isn't a Google search, you know, 89% gross margin stuff here.

Right.

So they end up having incredibly fast revenue takeoff just from the quick Stripe paywall that they threw up over a weekend to handle all the demand.

So to say that OpenAI had any idea what was coming would also be completely false.

They did not get that this would be the next big consumer product when they launched it.

Ben Thompson loves to call OpenAI the accidental consumer tech company, right?

Yes.

It was definitely accidental.

Now, there is actually

another slightly different version of the motivation for launching the chat.

Is this the Dario interface?

Yeah, the Dario and Anthropic version.

So Anthropic was working on what would become Claude and

rumors were out there and people at OpenAI got wind of like, oh, hey, Anthropic and Dario are working on a chat interface.

We should probably do one too.

And if we're going to do one, we should probably launch it before they launch theirs.

So I think that had something to do with the timing.

But again, I don't think anybody, including OpenAI, realized what was going to happen, which is, Ben, you alluded to it, but to give the actual numbers, on November 30th, 2022, basically Thanksgiving.

OpenAI launches a research preview of an interface to the new GPT 3.5 called ChatGPT.

That morning on the 30th, Sam Altman tweets, today we launched ChatGPT.

Try talking with it here.

And then a link to chat.openai.com.

Within a week, less than a week, actually, it gets 1 million users.

By the end of the year, so, you know, one month later, December 31st, 2022, it has 30 million users.

By the end of the next month, by the end of January 23, so two months after launch, it crosses 100 million registered users, the fastest product in history to hit that milestone.

Completely insane.

Completely insane.

Before we talk about what that unleashes within Google, which is the famous code red, To rewind a little bit back to Noam and the chatbot within Google, Mina, Google does keep working on Mina.

They develop it into something called Lambda, which is also a chatbot, also internal.

I think it was a language model.

At this point in time, they still differentiated between the underlying model brand name and the application name.

Yes, Lambda was the model.

And then there also was a chat interface to Lambda that was internal for Google use only.

Gnome is still advocating to leadership.

We got to release this thing.

He He leaves in 2021 and founds a chatbot company, Character AI, that still exists to this day.

And they raise a lot of money, as you would expect.

And then Google ultimately in 2024, after ChatGPT launches, pays $2.7 billion, I think, to do a licensing deal with Character AI, the net of which GNOME comes back to Google.

Yeah, I think Larry and Sergei were like,

if we're going to compete seriously, we kind of need Gnome back and blank check to go get him.

Yep.

So throughout 2021, 2022, Google's working on the Lambda model and then the chat interface to it.

In May of 2022, they do release something that is available to the public called AI Test Kitchen, which is a AI product test area where people can play around with Google's internal AI products, including the Lambda chat interface.

Yep.

And all fairness predates Chat GPT.

Do you know what they do to nerf chat so that it doesn't go too far off the rails?

This is amazing.

No.

For the version of Lambda chat that is in AI Test Kitchen, they stop all conversations after five turns.

So you can only have five turns of conversation with the chatbot, and then it's just, and we're done for today.

Thank you.

Goodbye.

Oh, wow.

And the reason they did that was for safety of like, you know, if the more turns you had with it, the more likely it would start to go off the rails.

And honestly, it was a fair concern.

I mean, this thing was not for public consumption.

And if you remember back a few years before, Microsoft released Tay,

which was this crazy racist chat bot.

Yeah, they launched it as a Twitter bot, right?

And it was going off the rails on Twitter.

This was in 2016, I think.

Right.

Maximal impact of badness.

Yeah.

And so despite Google, all the way back in 2017, Sundar declared we are an AI-first company, is being understandably very cautious in real public AI launches, especially on consumer-facing things.

Yep.

And as far as anyone else is concerned before ChatGPT, they are an AI-first company and they're launching all this amazing AI stuff.

It's just within the vector of their existing products.

Right.

So ChatGPT comes out, becomes the fastest product in history to 100 million users.

It is immediately obvious to Sundar, Larry, Sergei, all of Google leadership, that this is an existential threat to Google.

Chat GPT is a better user experience to do the same job function that Google search does.

And to underscore this, so if you didn't know it in November of 22, you sure knew it by February of 23, because good old Microsoft, our biggest, scariest enemy.

Oh, yeah.

announces a new bing powered by open ai and satya has a quote it's a new day for search the race starts today

there's an announcement of a new AI-powered search page.

He says, we want to rethink what search was meant to be in the first place.

In fact, Google's success in the initial days came by reimagining what could be done in search.

And I think the AI era we're entering gets us to think about it.

This is the worst possible thing that could happen to Google that.

Now Microsoft can actually challenge Google on their own turf, intent on the internet, with a legitimately different, better, differentiated product vector.

Not what Bing was trying to do, copycat.

This is the full leapfrog, and they have the technology partnership to do it.

Or so everybody thinks at the moment.

Oh my God, terrifying.

This is when Satya says the quote in an interview around this launch with Bing.

I want people to know that we made Google dance.

Oh boy.

Well, hey, if you come at the king, you'd best not miss.

Right.

And this big launch kind of misses.

Yes.

So what happens in Google December 2022, even before the big launch, but after the chat GPT moment, Sundar issues a code read within the company.

And what does that mean?

Up until this point, Google and Sundar and Larry and everyone had been thinking about AI as a sustaining innovation in Clay Christensen's terms.

This is great for Google.

This is great for our products.

Look at all these amazing things that we're doing.

It further entrenches incumbents.

It further is entrenching our lead in all of our already leading products.

We can deploy more capital in a predictable way to either drive down costs or make our product experiences that much better than any startup could make.

Get more monetized that much better, all the things.

Once ChatGPT comes out, on a dime overnight, AI shifts from being a sustaining innovation to a disruptive innovation.

It is now an existential threat.

And many of Google's strengths from the last 10, 15, 20 years of all the AI work that's happened in the company are now liabilities.

They have a lot of existing castles to protect.

That's right.

They have to run everything through a lot of filters before they can decide if it's a good idea to go try to out open AI, open AI.

Yep.

So this code red that Sundar issues to the company is actually a huge moment because what it means and what he says is we need to build and ship real

native AI products ASAP.

This is actually what you need to do in the textbook response to a disruptive innovation as the incumbent.

You need to not bury your head in the sand and you need to say, okay, we need to like actually go build and ship products that are comparable to these disruptive innovators.

And you need to be laser operationally in all the details to try and figure out where is it that the new product is actually cannibalizing our old product and where is it that the new product can be complementary and just lean into all the ways in which you can be complementary in all the different little scenarios.

And really what they've been trying to do, this ballet from 2022 onward, is protect the growth of search while also creating the best AI experiences they can.

And so it's very clever the way that they do AI overviews for some but not all queries.

And they have AI mode for some but not all users.

And then they have Gemini, the full AI app, but they're not redirecting Google.com to Gemini.

It's this like very delicate dance of protecting the existing franchise while also building a hopefully non-cannibalizing as much as we can new franchise.

Yep.

And you see them really going hard and, I think, building leading products in non-search cannibalizing categories like video.

Right.

VO3 or Nano Banana.

These are things that don't in any way.

cannibalize the existing franchise.

They, in fact, use some of Google's strength, all the YouTube training data and stuff like that.

Yeah.

So what happens next?

As you might expect, it gets worse before it gets better.

Code RED goes out December 2022.

Bard, baby.

Launch Bard.

Oh, boy.

Well, even before that, January 23,

when OpenAI hits 100 million registered users for ChatGPT, Microsoft announces they are investing another $10 billion in OpenAI and says that they now own 49% of the for-profit entity.

Incredible in and of itself.

But then now think about this from the Google lens of Microsoft, our enemy.

They now arguably own, obviously in retrospect here, they don't own OpenAI.

But it seems at the time like, oh my God, Microsoft might now own OpenAI, which is our first true existential threat in our history as a company.

Not great, Bob.

So then February 2023, the Bing integration launches.

Satya has the quote about wanting to make Google dance.

Meanwhile, Google is scrambling internally to launch AI products as fast as possible.

So the first thing they do is they take the Lambda model and the chatbot interface to it.

They rebrand it as BARD.

They ship that publicly.

And they release it immediately, February 2023.

Ship it publicly.

Available GA to anyone.

Which maybe was the right move, but God, it was a bad product.

It was really bad.

I didn't know the term at the time, RLHF, but it was clear it was missing a component of some magic that ChatGPT had, this reinforcement learning with human feedback, where you could really tune the appropriateness, the tone, the voice, the sort of correctness of the responses.

It just wasn't there.

Yep.

So to make matters worse, in the launch video for Bard, a video, this is a choreographed pre-recorded video where they're showing conversations with Bard.

Bard gives an inaccurate factual response to one of the queries that they include in the video.

This is one of the worst keynotes in history.

After the Bard launch and this keynote, Google's stock drops 8%

on that day.

And then, like we were saying, once the actual product comes out, it becomes clear.

It's just not good.

Yep.

And it pretty quickly becomes clear, it's not just that the chatbot isn't good, it's the model isn't good.

So in May, they replace Lambda with a new model from the brain team called Palm.

It's a little bit better, but it's still clearly behind not only GPT 3.5, but in March of 2023, OpenAI comes out with GPT-4, which is even better.

You can access that now through ChatGPT.

And here is where Sundar makes two really, really big decisions.

Number one, he says, we cannot have two AI teams within Google anymore.

We're merging brain and mind into one entity called Google DeepMind.

Which is a giant deal.

This is in full violation of the original deal terms of bringing deep mind in.

Yep.

And the way he makes it work is he says, Demis, you are now CEO of the AI division of Google, Google DeepMind.

This is all hands on deck.

And you and DeepMind are going to lead the charge.

You're going to integrate with Google Brain.

And we need to change all of the past 10 years of culture around building and shipping AI products within Google.

To further illustrate this, when Alphabet became Alphabet, they had all these separate companies, but things that were really core to Google, like YouTube, actually stayed a part of Google.

DeepMind was its own company.

That's how separate this was.

They're working on their own models.

In fact, those models are predicated on reinforcement learning.

That was the big thing that DeepMind had been working on the whole time.

And so reading in between the lines, it's Sundar looking at his two AI labs and going, look, I know you two don't actually get along that well.

But look, I don't care that you had different charters before.

I am taking the responsibility of Google Brain and giving it to DeepMind.

And DeepMind is absorbing the Google Brain team.

I think that's what you should sort of read into it, because as you look at where the models went from here, they kind of came came from DeepMind.

Yep.

There's a little bit of interesting backstory to this too.

So Mustafa, Suleiman, the third co-founder of DeepMind,

at some point before this, he became like the head of Google AI policy or something.

He had already shifted over to brain and to Google.

He stayed there for a little while and then he ended up getting close with who else?

Reed Hoffman.

Remember, Reed is on the ethics board for DeepMind.

And Mustafa and Reed leave and go found Inflection AI, which fast forward now into 2024, after the absolute insanity that goes down at OpenAI in Thanksgiving 2023 when Sam Altman gets fired over the weekend during Thanksgiving and then brought back by Monday when all the team threatened to quit and go to Microsoft.

Open AI loves Thanksgiving.

Can't wait for this year.

They love Thanksgiving.

Yeah, gosh.

After all that, which certainly strains the Microsoft relationship, remember, again, Reed is on the board of Microsoft.

Microsoft does one of these acquisition-type deals with Inflection AI and brings Mustafa in as the head of AI for Microsoft.

Crazy.

Wild, right?

Just wild.

Crazy turn of events.

Okay, so that first big decision that Sundar makes is unifying deep mind and brain.

That was huge.

Equally big, he says, I want you guys to go make a new model, and we're just going to have one model.

That is going to be the model for all of Google internally, for all of our AI products externally.

It's going to be called Gemini.

No more different models, no more different teams.

Just one model for everything.

This is also a huge deal.

It's a giant deal, And it's twofold.

It's push and it's pull.

It's saying, hey, if anyone's got a need for an AI model, you got to start using Gemini.

But two, it's actually kind of the Google Plus thing where they go to every team and they start saying, Gemini is our future.

You need to start looking for ways to integrate Gemini into your product.

Yes.

I'm so glad you brought up Google Plus.

This came up with a few folks I spoke to in the research.

Obviously, this is all playing out real time.

But the point a lot of people at Google made is the Gemini situation is very different than the Google Plus situation.

This is a technical thing, A, which has always been Google's wheelhouse.

But B, even more importantly, this is the rational business thing to do in the age of these huge models.

Even for a company like Google, there are massive scaling laws to models.

The more data you put in, the better it's going to get, the better all the outputs are going to be.

And

because of scaling laws, you need your models to be as big as possible in order to have the best performance possible.

If you're trying to maintain multiple models within a company, you're repeating multiple huge costs to maintain huge models.

You definitely don't want to do that.

You need to centralize on just one model.

Yeah, it's interesting.

There's also something to read into where at first it was the Gemini model underneath the BARD product.

Bard was still the consumer name.

And then at some point, they said, no, we're just calling it all Gemini, and Gemini became the user-facing name also.

This pulls in my quintessence from the alphabet episode.

I know it's a little bit woo-woo, but with Google saying we're actually going to name the consumer service the name of the AI model, they're sort of admitting to themselves, this product is nothing but technology.

There isn't productiness to do on top of it.

It's just like Gmail.

Gmail was was technology.

It was fast search.

It was lots of storage.

It was used in the web.

The productiness wasn't particular the way that like Instagram was all about the product.

Gemini the model, Gemini the chatbot says, we're just exposing our amazing breakthrough technology to you all, and you get to interface directly with it.

Anthropologically looking from afar, it kind of feels like it's that principle at work.

I totally agree.

I think it's actually a really important branding point and sort of rallying point to Google and Google culture to do this.

Right.

All right.

So this is all the stuff going on in Google 2023-ish in AI.

Before we catch up to the present, I have a whole other branch of alphabet that has been a real bright spot for AI.

Can I go there?

Can I take this off-ramp, if you will?

Can you take the wheel, so to speak?

May I take the wheel?

May I investigate another bet?

Yeah.

Please tell us the Waymo story.

Awesome.

So we got to rewind back all the way to 2004, the DARPA Grand Challenge, which was created as a way to spur research into autonomous ground robots for military use.

And actually, what it did for our purposes here today is create the seed talent for the entire self-driving car revolution 20 years later.

So the competition itself is really cool.

There is a 132-mile race course.

Now, mind you, this is 2004.

In the Mojave Desert that the cars have to race on, it is a dirt road.

No humans are allowed to be in or interact with the cars.

They are monitored 100% remotely, and the winner gets $1 million.

$1 million,

which was a break from policy.

Normally, these are grants, not prize money.

So this needs to be authorized by an act of Congress.

The $1 million eventually felt comical.

So the second year, they raised the pot to $2 million.

It's crazy thinking about what these researchers are worth today, that that was the prize for the whole thing.

So the first year in 2004 went fine.

There were some amazing tech demonstrations on these really tight budgets, but ultimately zero of the 100 registered teams finished the race.

But the next year in 2005 was the real special year.

The progress that the entire industry made in those first 12 months from what they learned is totally insane.

Of the 23 finalists that were entering the competition, 22 of them made it past the spot where the furthest team the year before had made it.

The amount that the field advanced in that one year is insane.

Not only that, five of those teams actually finished all 132 miles.

Two of them were from Carnegie Mellon, and one was from Stanford, led by a name that all of you will now recognize, Sebastian Thrun.

Indeed.

This is Sebastian's origin story before Google.

Now, as we said, Sebastian Sebastian was kind enough to help us with prep for this episode, but I actually learned most of this from watching a 20-year-old Nova documentary that is available on Amazon Prime Video.

Thanks to Brett Taylor for giving us the tip on where to find this documentary.

Yes, the hot research tip.

So what was special about this Stanford team?

Well, one, there is a huge problem with noisy data that comes out of all of these sensors.

You know, it's in a car in the desert getting rocked around.

It's in the heat.

It's in the sun.

So common wisdom and what Carnegie Mellon did was to do as much as you possibly can on the hardware to mitigate that.

So, things like custom rigging and gimbals and giant springs to stabilize the sensors, Carnegie Mellon would essentially buy a Hummer and rip it apart and rebuild it from the wheels up.

We're talking like welding and real construction on a car.

The Stanford team did the exact opposite.

They viewed any new piece of hardware as something that could fail.

And so, in order to mitigate risks on race day, they used all commodity cameras and sensors that they just mounted on a nearly unmodified Volkswagen.

So they only innovated in software and they figured they would just kind of come up with clever algorithms to help them clean up the messy data later.

Very googly, right?

Very googly.

The second thing they did was an early use of machine learning to combine multiple sensors.

They mounted laser hardware on the roof, just like what other teams were doing.

And this is the way that you can measure texture and depth of what is right in front of you.

And the data, it's super precise, but you can't drive very fast because you don't really know much about what's far away since it's this fixed field of view.

It's very narrow.

Essentially, you can't answer that question of how fast can I drive?

Or is there a turn coming up?

So on top of that, the way they solved it was they also mounted a regular video camera.

That camera can see a pretty wide field of view, just like the human eye, and it can see all the way to the horizon, just like the human eye.

And crucially, it can see color.

So what it would do, this is like really clever, they would use a machine learning algorithm in real time in 2005.

This computer is like sitting in the middle of the car.

They would overlay the data from the lasers on top onto the camera feed.

And from the lasers, you would know if the area right in front of the car was okay to drive or not.

Then the algorithm would look up in the frames coming off the camera, overlaid, what color that safe area was, and then extrapolate by looking further ahead at other parts of the video frame to see where that safe area extended to.

So you could figure out your safe path through the desert.

That's awesome.

It's so awesome.

I'm imagining like a Dell PC sitting in the middle of this car in 2005.

It's not far off in the email that we send out.

We'll share some photos of it.

It could then drive faster with more confidence and it knew when turns were coming up.

Again, this is real time on board the camera.

2005 is wild on that tech.

So ultimately, both of these bets bets worked and the Stanford team won in super dramatic fashion.

They actually passed one of the Carnegie Mellon teams autonomously through the desert.

It's like this big dramatic moment in the documentary.

So you would kind of think, so then Sebastian goes to Google and builds Waymo.

No,

as we talked about earlier, he does join Google through that crazy, please don't raise money from Benchmark and Sequoia and we'll just hire you instead.

But he goes and works on Street View and Project Ground Truth and co-founds Google X.

David, as you were alluding to earlier, this project chauffeur that would become Waymo is the first project inside Google X.

And I think the story, right, is that Larry came to Sebastian and was like, yo, that's self-driving car stuff.

Like, do it.

And Sebastian was like, no, come on.

That was a DARPA challenge.

And Larry was like, no, no, you should do it.

He's like, no, no, no, that won't be safe.

There's people running around cities.

I'm not just going to put multi-ton killer robots on roads and go and potentially harm people.

And Larry finally comes to him and says, why?

What is the technical reason that this is impossible?

And Sebastian goes home, has a sleep on it, and he comes in the next morning and he goes, I realize what it was.

I'm just afraid.

Such a good moment.

So they start.

He's like, there's not a technical reason.

As long as we can take all the right precautions and hold a very high bar on safety, let's get to work.

So Larry then goes, great, I'll give you a benchmark.

So that way you know if you're succeeding.

He comes up with these 10 stretches of road in California that he thinks will be very difficult to drive.

It's about a thousand miles.

And the team starts calling it the Larry 1000.

And it includes driving to Tahoe, Lombard Street in San Francisco, Highway 1 to Los Angeles, the Bay Bridge.

This is the bogey.

Yep.

If you can autonomously drive these stretches of road, pretty good indication that you can probably do anything.

Yep.

So they start the project in 2009.

Within 18 months, this tiny team, I think they hired, I don't know, it's like a dozen people or something.

They've driven thousands of miles autonomously, and they managed to succeed in the full Larry 1000 within 18 months.

Totally unreal how fast they did it.

And then also totally unreal.

How long it takes after that to productize and create the Waymo that we know today.

Right.

It's like the the first 99% and then the second 99% that takes 10 years.

Yeah.

Self-driving is one of these really tricky types of problems where it's surprisingly easy to get started, even though it seems like it would be an impossible thing.

But then there's edge cases everywhere, weather, road conditions, other drivers, novel road layouts, night driving.

So it takes this massive amount of work for a production system to actually happen.

So then the question is, what business do we build?

What is the product here?

And there was what Sebastian wanted, which was highway assist, assist, sort of the lowest stakes, most realistic.

Let's make a better cruise control.

There's what Eric Schmidt wanted, which is crazy.

He proposed, oh, let's just go buy Tesla and that'll be our starting place.

And then we'll just put all of our self-driving equipment on all the cars.

David, do you know what it would have cost to buy Tesla at the time?

I think at the time that negotiations were taking place between Elon and Larry and Google, this was in the depths of the Model S production scaling woes.

I think Google could have bought the company for $5 billion.

That's what I remember.

It was $3 billion.

$3 billion.

Oh, my goodness.

Obviously, that didn't happen, but what a crazy alternative history that could have been.

Right.

I mean, I think if that had happened, DeepMind would not have gone down in the same way, and probably OpenAI would not have gotten founded.

Who?

That's probably right.

I think that is obviously unprovable.

Right.

The counterfactuals that we always come up with on this show, you can't know.

Yeah.

Seems more likely than not to me that at a minimum, OpenAI would not exist.

Right.

So then there was what Larry wanted to do.

Option three, build robo-taxis.

Yeah.

And ultimately, that is, at least right now, what they would end up doing.

So we could do a whole episode about this journey, but we will just hit some of the major points for the sake of time.

The big thing to keep in mind here, neither Google nor the public really knew if self-driving was something that could happen in the next two years from any given point or take another 10.

And just to illustrate it, for the first five years of Project Chauffeur, it did not use deep learning at all.

They did the Larry 1000 without any deep learning and then went another three and a half years.

Wow.

That's crazy.

Yeah.

And yet, totally illustrates you never know how far away the end goal is.

And this is a field that comes from the only way progress happens is through these series of breakthroughs.

And you don't know, A, how far the next breakthrough is, because at any given time, there's lots of promising things in the field, most of which don't work out.

And then B, when there is a breakthrough, actually how much lift that will give you over existing methods.

So anytime people are forecasting, oh, in AI, we're going to be able to do X, Y, Z, and X years, it's a complete fool's errand.

Even the experts don't know.

Here are the big milestones.

2013, they started using convolutional neural nets.

They could identify objects.

They got much better perception capabilities.

This 2013, 2014 period is when Google found religion around deep learning.

So this is like right after the 40,000 GPUs rolled out.

So they've actually got some hardware to start doing this on now.

2016, they've seen enough technology proof that they think, let's commercialize this.

We can actually spin this out into a company.

So Waymo becomes its own.

subsidiary inside of alphabet it's no longer a part of google x anymore 2017 obviously the transformer comes out they incorporate some learnings from the transformer especially around prediction and planning.

March of 2020, they raise $3.2 billion from folks like Silver Lake Canada Pension and Investment Board, Mubatola, andreessen Horowitz.

And of course, the biggest check, I think, Alphabet.

And I think they're always the biggest check because Alphabet is still the majority owner, even after a bunch more fundraises.

In October of 2020, they launched the first public commercial no-human behind the driver's seat thing in Phoenix.

It's the first in the world.

This is 11 years after succeeding in the Lowry 1000.

And this is nuts.

I had given up at this point.

I was like, that's cute that Waymo and all these other companies are trying to do self-driving.

Seems like it's never going to happen.

And then they actually were doing a large volume of rides safely with consumers and charging money for it in Phoenix.

Then they bring it to San Francisco, where for me and lots of people in San Francisco, it is a huge part of life in the city here now.

It's amazing.

Yeah.

Every time I'm down, I love taking them.

They're launching in Seattle soon.

I'm pumped.

Interestingly, they don't make the hardware.

So they use a Jaguar vehicle that, from what I can tell, is only in Waymo's.

Like, I don't know if anybody else drives that Jaguar or if you can buy it, but they're working on a sort of van next.

They have some next generation hardware.

For anyone who hasn't taken it, it's an Uber, but with no driver.

And that launched in June of 24.

Along the way there, they raised their quote-unquote Series B another 2.5 billion.

Then, after the San Francisco rollout, they raised their quote-unquote Series C 5.6 billion.

This year, in January, they were reportedly doing more in gross bookings than Lyft in San Francisco.

Wow.

I totally believe it.

I mean, it is the number one option in San Francisco that I and everybody I know too always goes to for right-hailing.

It's like, try to get a Waymo.

If there's not a Waymo available anytime soon, then go down the stack.

Like, we're living in the future and how quickly we fail to appreciate it.

Yeah.

And what's cool, I think for people who it hasn't come to their city and is not part of their lives yet, it's not just that it's a cool experience to not have a driver behind them.

Like pretty quickly, that just fades.

It's actually a different experience.

So if I need to go somewhere with my older daughter, I don't mind hailing a Waymo, bringing the car seat, installing the car seat in the Waymo and driving with my daughter.

And she loves it.

We call it a robot car.

And she's like, a robot car.

I'm so excited.

Huh.

I would never do that with an Uber.

That's interesting.

To my dog.

Whenever I need to go with my dog, like it's super awkward to hail an Uber and be like, hey, I got my dog.

You know, can the dog come in?

Not a big deal with a Waymo.

And then when you're in town.

Yeah, we can actually have sensitive conversations in the car.

You can have phone calls.

It really is a different experience.

Yeah, that's so true.

Yeah, so may as well catch up to today.

They're operating in five cities, Phoenix, San Francisco, LA, Austin, and Atlanta.

They have hundreds of thousands of paid rides every week.

They've now driven over 100 million miles with no human behind the wheel, growing at 2 million every week.

There's over 10 million paid rides across 2,000 vehicles in the fleet.

They're going to be opening a bunch more cities in the U.S.

next year.

They're launching in Tokyo, their first international city, slowly and then all at once.

I mean, that's kind of the lesson here.

The technology, they really continued with that multi-sensor approach all the way from the DARPA Grand Challenge.

Camera, LiDAR, they added radar.

And actually, they use audio sensing as well.

And their approach is basically any data that we can gather is better because that makes it safer.

So they have 13 cameras, four LIDAR, six radar, and the array of external microphones.

This is obviously way more expensive of a solution than what Tesla is just doing with cameras.

But Waymo's party line is they believe it is the only path to full autonomy to hit the safety bar and regulatory bar that they're aiming for.

It seems like a really big line in the sand for them anytime you talk to somebody in that organization.

Yeah.

And look, as a regular user of both products, you know, happy owner and driver of a Model Y in addition to regular Waymo user, at least with the current instantiation of full self-driving on my Tesla, vastly different products.

Full self-driving on my Model Y is great.

I use it all the time on the freeway, but I would never not pay attention.

Whereas every time I get in a Waymo, it's almost like Google search, right?

It's like I just trust that, oh, this is going to be completely and totally safe.

And I'm sitting in the back seat and I can totally tune out.

I think I trust my Model Y FSD more than you do, but I get what you're saying.

And frankly, regulatorily, you are required to still pay attention in Tesla and not in the Waymo.

The safety thing is super real, though.

I mean, if you look at the numbers, over a million motor vehicle crashes cause fatalities every year, or there's over a million fatalities.

In the U.S.

alone, over 40,000 deaths occur per year.

So if you break that down, that's 120 every day.

I mean, that's like a giant cause of death.

Yes.

The study that Waymo just released last month showed that they have 91% fewer crashes with serious injuries or worse compared to the average human driver, even controlled for the fact that Waymos right now are only driving on city surface streets.

So they controlled it apples to apples with human driving data.

And it's a 91% reduction in those serious either fatality or serious injury things.

Why aren't we all talking about this all the time every day?

This is going to completely change the world and a giant cause of death.

Yeah.

So while we're in Waymo land, what do you think about doing some quick analysis?

Great.

Because I've been scratching my head here of what is this business?

And then I promise we'll go back to the rest of Google AI and catch up to today.

It is super expensive to operate, especially at early scale.

The training is high, the inference is high, the hardware is high, et cetera, et cetera, et cetera.

Also, the operations are expensive.

Yes.

And in fact, they're experimenting.

Some cities, they actually outsource the operations.

So the fleet is managed by, there's a rental car company in Texas that manages it, or they've partnered, I believe, with Lyft and with Uber and different.

So they're trying all sorts of O ⁇ O versus partnership models to operate it.

Yeah.

And the operations are like, these are electric cars.

They need to be charged.

They need to be cleaned.

They need to be returned to depots.

They need to be checked out.

They need to have sensors replaced.

So the question is, what is the potential market opportunity?

How big could this business be?

And there's a few different ways you could try to quantify it.

One total market size thing you could do is try to sum the entire automaker market cap today.

And that would be 2.5 trillion globally if you include Tesla or 1.3 trillion without.

But Waymo is not really making cars.

So that's probably the wrong way to slice it.

You could look at all the ride-sharing companies today, which might be a better comp because that's the business that Waymo is actually in today.

That's on the order of 300 billion, most of which is Uber.

So that's addressable market cap today with ride-sharing.

Waymo's ambitions, though, are bigger than that.

They want to be...

in the cars that you own.

They want to be in long-hauled trucking.

So they believe they can grow the share of transportation because there's blind people that could own a car.

There's elderly people who could get where they need to go on their own without having a driver, that sort of thing.

So the most squishy, but I think the most interesting way to look at it is what is the value from all of the reduction in accidents?

Because that's really what they're doing.

It's a product.

to replace accidents with non-accidents.

I think that's viable.

But again, I would say as a regular user of the product, it is a different and expanding product to human rideshare.

So your argument is whatever number I come up with for reducing accidents, it's still a bigger market than that because there's additional value created in the product experience itself.

Yeah.

Scoping just to ride share, now that we have Waymo in San Francisco, I use Waymo in scenarios where I would never use an Uber or a Lyft.

Yeah.

Makes sense.

So here's the data we have.

The CDC released a report saying deaths from crashes in 2022 in the U.S.

resulted in $470 billion in total costs, including medical costs and the cost estimates for lives lost, which is crazy that the CDC has some way of putting the costs on human life, but they do.

So if you reduce crashes 10x, which is what Waymo seems to be saying in their data, at least for the serious crashes, that's over $420 billion a year in total costs that we would save as a nation.

Now, it's not totally apples to apples.

I recognize this, but that cost savings is more than Google does today in revenue in their entire business.

You could see a path to a Google-sized opportunity for Waymo as a standalone company just through this analysis, as long as they figure out a way to get costs down to the point where they can run this as a large and profitable business.

Yeah, it is an incredible

20-plus year success story.

within Google.

The way I want to close it is the investment so far actually hasn't been that that large.

When you consider this opportunity, they have burned somewhere in the neighborhood of $10 to $15 billion.

That's sort of why I was listing all the investments to get to this point.

Jump change compared to foundational models.

Dude, also, let's just keep it scoped in this sector.

That's one year of Uber's profits.

Wow.

Seems like a good bet.

I used to think this was like some wild goose chase.

It now looks really, really smart.

Yep.

Totally agree.

Also, that cost 10 to 15 billion is the profits that Google made last month.

Google.

Well, speaking of Google, should we catch us up to today with Google AI?

Yes.

So I think where you were is the Gemini launch.

So Sundar makes these.

two decrees mid-2023.

One, we're merging brain and deep

into one team for AI within Google.

And two, we're going to standardize on one model, the future Gemini and

DeepMind slash brain team.

You go build it.

And then everybody in Google, you're going to use it.

Not to mention, apparently, Sergey Bran is like now back as an employee working on Gemini.

Yes.

Employee number.

Got his new badge back.

Yeah, got his badge back.

So once Sundar makes these decisions, Jeff Dean and Aurel Vinyalis from Brain go over and team up with the DeepMind team and they start working on Gemini.

I'm a believer now, by the way.

You got Jeff Dean working on it?

I'm in.

If you got Jeff Dean on it, it's probably going to work.

If you weren't a believer, yeah, wait till I'm going to tell you next.

Once they get Noam back, when they do the deal with character AI, bring him back into the fold.

Noam joins the Gemini team, and Jeff and Noam are the two co-technical leads for for Gemini now.

So let's go.

Let's go.

So they actually announced this very quickly at the Google IO keynote in May 2023.

They announced Gemini.

They announced the plans.

They also launch AI overviews in search, first as a labs product, and then later that becomes just standard for everybody using Google search.

Which is crazy, by the way.

The number of Google searches that happen is unfathomably large.

I'm sure there's a number for it, but just think about that's about the highest level of computing scale that exists other than like high bandwidth things like streaming.

But just think about the instances of Google searches that happen.

They are running an LLM inference on all of those, or at least as many as they're willing to show AI overviews on, which I'm sure is not every query, but many.

A subset, yeah.

But still a large, large number of Google searches.

I mean, I see them all the time.

Yep.

This is really Google immediately deciding to operate at AI speed.

I mean, Chat GPT happened in November 30th, 2022.

We're now in May 2023.

All of these decisions have been made.

All of these changes have happened and they're announcing things at IO.

And they're really flexing the infrastructure that they've got.

I mean, the fact that they can go, like, oh, yeah, sure, let's do inference on every query.

We're Google.

We can handle it.

So a key part of this new Gemini model that they announce in May 2023 is it's going to be multimodal.

Again, this is one model for everything, text, images, video, audio, one model.

They release it for early public access in December 2023.

So also crazy, six months.

They build it.

They train it.

They release it.

That is amazing.

Wild.

February 2024, they launched Gemini 1.5 with a 1 million token context window, much, much larger context window than any other model on the market.

Which enables all sorts of new use cases.

There's all these people who were like, oh, I tried to use AI before, but it couldn't handle my XYZ use case.

Now they can.

Yep.

The next year, February 2025, they released Gemini 2.0.

March of 2025, one month later, they launched Gemini 2.5 Pro in experimental mode, and then that goes GA in June.

This is like NVIDIA pace, how often they're shipping.

Yeah, seriously.

And also in March of 2025, they launched AI mode.

So you can now switch over on google.com to chatbot mode.

And they're split testing, auto-opting some people into AI mode to see what the response is.

This is the golden goose.

Yeah, the elephant is tap dancing here.

Yep.

Then there's all the other AI products that they launched.

So Notebook LM comes out during this period.

AI generated podcasts.

Which, does that sound like us to you?

It feels a little trained.

The number of texts that we got when that came out of this must be trained on Acquired.

I do know that a bunch of folks on the notebook LM team are Acquired fans.

So I don't know if they trained on us.

And then there's the video, the image stuff, VO3, Nano Banana, Genie 3 that just came out recently.

Genie, this is insane.

This is a world builder based on prompts and videos.

Yeah.

You haven't actually used it yet, right?

You watched that hype video.

Yeah, I watched the video.

I haven't actually used it.

Yeah.

I mean, if it does that, that's unbelievable.

It's a real-time generative world builder.

World builder.

Yeah.

You look right and it invents stuff to your right.

I mean, you combine that with like a Vision Pro hardware.

You're just living in a fantasy land.

So they announced there are now 450 million monthly users of Gemini.

Now, that includes everybody who's accessing Nano Banana.

Yeah, I can't believe this stat.

This is insane.

Even with recently being number one in the app store, it still feels hard to believe.

Google's saying it, so it must be true.

But I just wonder, what are they counting as use cases of the Gemini app?

Right.

Certainly everybody who's using Nanobanana is using Gemini.

But is it counting AI overviews or is it counting AI mode?

Or is it counting something where I'm like accidentally like Meta said that crazy high number of people using Meta AI?

And

that was complete garbage.

That was people searching Instagram who accidentally hit a llama model that made some things happen.

And they were like, like, oh, go away.

I actually am just looking for a user.

Is it really 450 million or is it 450 million?

Yeah.

Good question.

Either way, going from zero is crazy impressive in the amount of time that they've done.

Especially given revenue is at an all-time high.

They seem to so far be,

at least in this squishy, early phase, able to figure out how to keep the core business going.

while

doing well as a competitor in the cutting edge of AI.

Yep.

And to foreshadow a little bit to we're going to do a bull and bear here in a minute.

As we talked about in our alphabet episode, Google does have a history of navigating platform shifts incredibly well in the transition to mobile.

It's true.

Definitely a rockier start here in the AI platform shift.

Much rockier.

But hey, look, I mean, if you were to lay out a recipe for

how to respond, given the rocky start, it would be hard to come up with a much better slate of things than what they've done over the last two years.

Yeah.

All right.

Should I give us the snapshot of the business today?

Give us the snapshot of the business today.

Oh, yeah.

Also, by the way, the federal government decided they were a monopoly and then decided not to do anything about it because of AI.

Yeah, so between the time when we shipped our alphabet episode and here with our Google AI episode, or our part two and part three, for those who prefer simpler naming schemes, yeah,

there was a US versus Google antitrust case.

The judge first ruled that Google was a monopoly in internet search and then did not come up with any material remedies.

I mean, there are some, but I would call them immaterial.

They did not need to spin off Chrome and they did not need to stop sending tens of billions of dollars to Apple and others.

In other words, yes, Google's a monopoly and the cost of doing anything about that would have too many downstream consequences on the the ecosystem.

So we're just going to let them keep doing what they're doing.

And one of the reasons that the judge cited of why they weren't going to really take these actions is because of the race in AI, that because tens of billions of dollars of funding have gone into companies like OpenAI and Anthropic and Perplexity, Google essentially has this new war to fight and we're going to leave it to the free market to do its thing where it creates viable competition on its own and we're not going to hamstring Google.

Personally, I think this argument is a little bit silly.

I mean, none of these AI companies are generating net income.

And just because they've raised a huge amount of money, it doesn't mean that will last forever.

They'll all burn through their existing cash in a pretty short period of time.

And if the spigots ever dry up, Google doesn't have any self-sustaining competition right now, whether in their old search business or in AI.

It is all

dependent on people believing that the opportunity is so large that they keep pouring tens of billions of of dollars into these competitors.

Yeah, plenty of other folks have made the sort of glib comment, but there's merit to it of, hey, as flat-footed as Google was when ChatGPT happened, if the outcome of this is they avoid a Microsoft-level distraction and damage to their business from a U.S.

federal court monopoly judgment, worth it.

Well, there's a funny meme here that you could draw.

You know that meme of someone pushing the domino and it knocking over some big wall later?

Yeah.

There's the domino of Ilya leaving Google to start OpenAI.

And the downstream effect is Google is not broken up.

Yeah, right, exactly.

It actually saves Google.

It actually saves Google.

It's totally wild.

Totally wild.

All right.

So here's the business today.

Over the last 12 months, Google has generated $370 billion

in revenue.

On the earnings side, they've generated 140 billion over the last 12 months, which is more profit than any other tech company.

And the only company in the world with more earnings is Saudi Aramco.

Let's not forget, Google is the best business ever.

And we also made the point at the end of the Health Abit episode.

Even in the midst of all of this AI era and everything that's happened over the last 10 years, the last five years, Google's core business has continued to grow 5x since the end of our alphabet episode in 2015, 2016.

Yeah.

Market cap, Google surged past their old peak of 2 trillion and just hit that 3 trillion mark earlier this month.

They're the fourth most valuable company in the world behind NVIDIA, Microsoft, and Apple.

It's just crazy.

On their balance sheet, actually, I think this is pretty interesting.

I normally don't look at balance sheet as a part of this exercise, but it's useful.

And here's why in this case, they have 95 billion in cash and marketable securities.

And I was about to stop there and make the point, wow, look how much cash and resources they have.

I'm actually surprised it's not more.

So it used to be 140 billion in 2021.

And over the last four years, they've massively shifted from this mode of accumulating cash to deploying cash.

And a huge part of that has been the capex of the AI data center build out.

So they're very much playing offense in the way that Meta, Microsoft, and Amazon are in deploying that CapEx.

But the thing that I can't quite figure out is the largest part of that was actually buybacks and they started paying a dividend.

So if you're not a finance person, the way to read into that is, yes, we still need a lot of cash for investing in the future of AI and data centers.

But we still actually had way more cash than we needed and we decided to distribute that to shareholders.

That's crazy.

Best business of all time, right?

That illustrates what a crazy business their CoreSearch ads business is if they're saying the most capital-intense race in business history is happening right now.

We intend to win it.

And we have tons of extra cash lying around on top of what we think, plus a safety cushion for investing in that CapEx race.

Yeah.

Yes.

Wow.

So there are two businesses that are worth looking at here.

One is Gemini to try to figure out what's happening there.

And two is a brief history of Google Cloud.

I want to tell you the cloud numbers today, but it's probably worth actually understanding how did we get here on cloud.

Yep.

First on Gemini, because this is Google and they have, I think, the most obfuscated financials of any of the companies we've studied.

They anger me the most in being able to hide the ball in their financial statements.

Of course, we don't know Gemini-specific revenue.

What we do know is there are over 150 million paying subscribers to the Google One bundle.

Most of that is on a very low tier.

It's on like the $5 a month, $10 a month.

The AI stuff kicks in on the $20 a month tier where you get the premium AI features, but I think that's a very small fraction of the 150 million today.

I think that's what I'm on.

But two things to note.

One, it's growing quickly, that 150 million is growing almost 50% year over year.

But two is Google has a subscription bundle that 150 million people are subscribed to.

And so I've kind of had it in my head that AI doesn't have a future as a business model that people pay money for, that it has to be ad supported like search.

But hey, that's not nothing.

That's like a...

That's almost half of America.

I mean, how many subscribers does Netflix have?

Netflix is in the hundreds of millions.

Yeah.

Spotify is now a quarter billion, something like that.

We now live in a world where there are real scaled consumer subscription services.

I owe this insight to Shashir Moroda.

We chatted actually last night because I name dropped him in the last episode and then he heard it.

And so we reached out and we talked.

And that's made me do a 180.

I used to think if you're going to charge for something, your total addressable market shrunk by 90 to 99%.

But he kind of has this point that if you build a really compelling bundle and Google has the digital assets to build a compelling bundle.

Oh my goodness.

YouTube Premium, NFL Sunday Ticket.

Yes.

Stuff in the Play Store, YouTube Music, all the Google One storage stuff.

They could put AI in that bundle and figure out through clever bundle economics a way to make a paid AI product that actually reaches a huge number of paying subscribers.

Totally.

So we really can't figure out how much money Gemini makes right now.

Probably not profitable anyway.

So what's the point of even analyzing it?

Yep.

But okay, tell us the cloud story.

So we intentionally did not include cloud in our alphabet episode.

Google Part 2, effectively.

Google Part 2, yes.

Because

it is a new product and now very successful one within Google that was started during the same time period as all the other ones that we talked about doing, Google Part 2.

But it's so strategic for AI.

Yes, it is a lot more strategic now in hindsight than it looked when they launched it.

So just quick background on it.

It started as Google App Engine.

It was a way in 2008 for people to quickly spin up a backend for a web or soon after a mobile app.

It was a platform as a service.

So you had to do things in this very narrow, googly way.

It was very opinionated.

You had to use this SDK.

You had to write it in Python or Java.

You had to deploy exactly the way they wanted you to deploy.

It was not a thing where they would say, hey, developer, you can do anything you want.

Just use our infrastructure.

It was opinionated.

Super different than what AWS was doing at the time and what they're still doing today, which the whole world eventually realized was right, which is cloud should be infrastructure as a service.

Even Microsoft pivoted Azure to this reasonably quickly, where it was like, you want some storage, we got storage for you.

You want a VM, we got a VM for you.

You want some compute?

You want a database?

We got you.

Fundamental building blocks.

So eventually, Google launches their own infrastructure as a service in 2012.

Took four years.

They launched Google Compute Engine that they would later rebrand Google Cloud Platform.

That's the name of the business today.

The knock on Google is that they could never figure out how to possibly interface with the enterprise.

Their core business, they made really great products for people to use that they loved polishing.

They made them all as self-serve as possible.

And then the way they made money was from advertisers.

And let's be honest.

There's no other choice but to use Google search.

Right.

It didn't necessarily need to have a great enterprise experience for their advertising customers because they were going to come anyway.

Right.

And so they've got this self-serve experience.

Meanwhile, the cloud is a knife fight.

These are commodities.

All about the enterprise.

It's the lowest possible price and it's all about enterprise relationships and clever ways to bundle and being able to deliver a full solution.

You say solution, I hear grows margin.

Yes.

But yes.

So Google out of their natural habitat in this domain.

And early on, they didn't want to give away any crown jewels.

They viewed their infrastructure as this is our secret thing.

We don't want to let anybody else use it.

And the best software tools that we have on it, that we've written for ourselves, like Big Table or Borg, how we run Google, or Disk Belief, these are not services that we're making available on Google Cloud.

Yep.

These are competitive advantages.

Yes.

And then they hired the former president of Oracle, Thomas Kirion.

Yes.

And everything kind of changed.

So 2017, two years before he comes in, they had $4 billion in revenue 10 years into running this business.

2018 is their first very clever strategic decision.

They launched Kubernetes.

The big insight here is if we make it more portable for developers to move their applications to other clouds, the world is kind of wanting multi-cloud here.

Right.

We're the third place player.

We don't have anything to lose.

Yes.

So we can offer this tool and kind of counter position against AWS and Azure.

We shift the developer paradigm to use these containers.

They orchestrate on our platform.

And then, you know, we have a great service to manage it for you.

It was very smart.

So this kind of becomes one of the pillars of their strategy is you want multi-cloud.

We're going to make that easy.

And you can, sure, choose AWS or Azure too.

It's going to be great.

So David, as you said, the former president of Oracle, Thomas Kurian, is hired in late 2018.

You couldn't ask for a better person who understands the needs of the enterprise than the former president of Oracle.

This shows up in revenue growth right away.

In 2020, they crossed 13 billion in revenue, which was nearly tripling in three years.

They hired like 10,000 people into the go-to-market organization.

I'm not exaggerating that.

And that's on a base of 150 people when he came in, most of which were seated in California, not regionally distributed throughout the world.

The funniest thing is Google kind of was a cloud company all along.

They had the best engineers building this amazing infrastructure.

Right.

They had the products.

They had the infrastructure.

They just didn't have the go-to-market organization.

Right.

And the productization was all like googly-y.

It was like for us, for engineers, they didn't really build things that let enterprises build the way that they wanted to build.

This all changes.

2022, they hit 26 billion in revenue.

2023, they're like a real viable third cloud.

They also flipped to profitability in 2023.

And today, they're over $50 billion in annual revenue run rate.

It's growing 30% year over year.

They're the fastest growing of the major cloud providers, 5X in five years.

And it's really three things.

It's finding religion on how to actually serve the enterprise.

It's leaning into this multi-cloud strategy and actually giving enterprise developers what they want.

And three, AI has been such a good tailwind.

for all hyperscalers because these workloads all need to run in the cloud because it's giant amounts of data and giants amount of compute and energy.

But in Google Cloud, you can use TPUs, which they make a ton of.

And everyone else is desperately begging NVIDIA for allocations to GPUs.

So if you're willing to not use CUDA and build on Google Stack, they have an abundant amount of TPUs for you.

This is why we saved cloud for this episode.

There are two aspects of Google Cloud that I don't think they foresaw back when they started the business with App Engine, but are hugely strategically important to Google today.

One is just simply that cloud is the distribution mechanism for AI.

So if you want to play in AI today, you either need to have a great application, a great model, a great ship, or a great cloud.

Google is trying to have all four of those.

Yes.

There is no other company that has,

I think,

more than one.

I think that's the right call.

Think about the big AI players.

Nvidia

kind of has a cloud, but not really.

They just have chips.

And they have the best chips and the chips everyone wants, but chips.

And then you just look around the rest of the big tech companies.

Meta right now, only an application.

They're completely out of the race for the frontier models at the moment.

We'll see what their hiring is pre-yields.

You look at Amazon, infrastructure, they have application.

Maybe, I don't actually know if Amazon.com.

I'm sure it benefits from LLMs in a bunch of ways.

Mainly, it's cloud.

Yes, cloud.

And cloud leader.

Microsoft?

Cloud.

It's just cloud, right?

They make some models, but...

I mean, they've got applications, but yeah, cloud.

Cloud.

Apple?

Nothing.

Nothing?

AMD, just chips.

Yep.

Open AI, model.

Anthropic, model.

Yep.

Yep.

These companies don't have their own data centers.

They are like making noise about making their own chips, but not really.

And certainly not at scale.

Google has scale data center, scale chips, scale usage of model.

I mean, even just from Google.com queries now on AI overviews.

And scale applications.

Yes.

Yeah.

They have all of the pillars of AI.

And I don't think any other company has more than one.

And they have the very most net income dollars to lose.

Right.

So then there's the chip side specifically of this.

If Google didn't have a cloud, it wouldn't have a chip business.

It would only have an internal chip business.

The only way that external companies, users, developers, model researchers could use TPUs would be if Google had a cloud to deliver them, because there's no way in hell that Amazon or Microsoft are going to put TPUs from Google in their clouds.

We'll see.

We'll see, I guess.

I think within a year, it might happen.

There are rumors already that some Neo clouds in the coming months are going to have TPUs.

Mmm, interesting.

Nothing announced, but TPUs are likely going to be available in NeoClouds soon, which is an interesting thing.

Why would Google do that?

Are they trying to build an NVIDIA type business where they make money selling chips?

I don't think so.

I think it's more that they're trying to build an ecosystem around their chips the way that CUDA does.

And you're only going to credibly be able to do that if your chips are accessible and anywhere that someone's running their existing workloads.

Yep.

Be very interesting if it happens.

And, you know, look, you may be right.

Maybe there will be TPUs in AWS or Azure someday,

but I don't think they would have been able to start there.

If Google didn't have a cloud and there weren't any way for developers to use TPUs and start wanting TPUs,

would Amazon or Microsoft be like, eh, you know, all right, Google.

We'll take some of your TPUs, even though no developer out there uses them.

Right.

All right.

Well, with that, let's move into analysis.

I think we need to do bull and bear on this one.

You have to this time.

Got to bring that back.

For these episodes in the present, it seems like we need to paint the possible futures.

Yes, bringing back bull and bear.

I love it.

Then we'll do playbook, powers, quintessence, bring it home.

Perfect.

All right.

So here's my set of bull cases.

Google has distribution.

to basically all humans as the front door to the internet.

They can funnel that however they want.

You've seen it with AI overviews.

You've seen it with AI mode.

Even though lots of people use ChatGPT for lots of things, Google's traffic, I assume, is still essentially an all-time high and it's a default behavior.

Yep.

Powerful.

So that is a bet on implementation that Google figures out how to execute and build a great business out of AI, but it is still theirs to lose.

Yep.

And they've got a viable product.

It's not clear to me that Gemini is any worse than

OpenAI or Anthropics products.

No, I completely agree.

This is a value creation, value capture thing.

The value creation is there in spades.

The value capture mechanism is still TBD.

Google's old value capture mechanism is one of the best in history.

So that's the issue at hand.

Let's not get confused that it's not like a good experience.

It's a great experience.

Yeah, yeah.

Yeah.

Okay.

So we've talked about the fact that Google has all the capabilities to win an AI and it's not even close.

Foundational model, chips, hyperscaler, all this with self-sustaining funding.

I mean, that's the other crazy thing: you look at the clouds have self-sustaining funding, Nvidia has self-sustaining funding.

None of the model makers have self-sustaining funding.

So they're all dependent on external capital.

Yeah.

Google is the only model maker who has self-sustaining funding.

Yes.

Isn't that crazy?

Yeah.

Basically, all the other large-scale usage foundational model companies are effectively startups.

Yes.

And Google's is funded by a money funnel so large that they're giving extra dollars back to shareholders for fun.

Yeah.

Again, we're in the bull case.

Well, when you put it that way, yeah.

A thing we didn't mention, Google has incredibly fat pipes connecting all of their data centers.

After the dot-com crash in 2000, Google bought all that dark fiber for pennies on the dollar, and they've been activating it over the last decade.

They now have their own private backhaul network between data centers.

No one has infrastructure like this.

Yep.

Not to mention, that serves YouTube.

They're fat pipes.

Which, in and of itself, is its own bull case for Google in the future.

That's a great point.

Yep.

Ben Thompson had a big article about this yesterday at the time of recording.

Yeah, that was like a mega bull case that Ben Thompson published this week.

That it was an interesting point.

A text-based internet is kind of the old internet.

It's the first instantiation of the internet because we didn't have much bandwidth.

The user experience that is actually compelling is

high resolution video everywhere all the time.

We already live in the YouTube internet.

Right.

And not only can they train models on really the only scale source of UGC media across long form and short form, but they also have that as the number two search engine, this massive destination site.

So they previewed things like you'll be able to buy AI labeled or AI determined things that show up in videos.

And if they wanted to, they could just go label every single product in every single video and make it all instantly shoppable.

It doesn't require any human work to do it.

They could just do it and then run their standard ads model on it.

That was a mind-expanding piece that Ben published yesterday, or I guess if you're listening to this a few weeks ago, about that.

And then there's also all the video AI applications that they've been building, like Flow and VO.

What is that going to do for generating videos for YouTube that will increase engagement and add dollars for YouTube?

Yep.

Going to work real well.

Yep.

They still have an insane talent bench, even though they've bled talent here and there and lost people.

They have also shown they're willing to spend billions for the right people and retain them.

Unit economics.

Let's talk about the unit economics of chips.

Everyone is paying NVIDIA 75, 80% gross margins, implying something like a four or 5X markup on what it costs to make the chips.

A lot of people refer to this as the Jensen tax or the NVIDIA tax.

You can call it that, you can call it good business, you can call it pricing power, you could call it scarcity of supply, whatever you want.

But that is true.

Anyone who doesn't make their own chips is paying a giant, giant premium to NVIDIA.

Google has to still pay some margin to their chip hardware partner, Broadcom, that handles a lot of the work to actually make the chip, interface with TSMC.

I have heard that Broadcom has something like a 50% margin when working with Google on the TPU versus Nvidia's 80%,

but that's still a huge difference to play with.

A 50% gross margin from your supplier or an 80% gross margin from your supplier is the difference between a 2x markup and a 5x markup.

Yeah, I guess that's right.

When you frame it that way, it's actually a giant difference of the impact to your cost.

So you might wonder appropriately, well, are chips actually the big part of the cost of like the total cost of ownership ownership of running one of these data centers or training one of these models?

Chips are the main driver of the cost.

They depreciate very quickly.

I mean, this is at best a five-year depreciation because of how fast we are pushing the limits of what we can do with chips, the needs of next-generation models, how fast TSMC is able to produce.

Yeah.

I mean, even that is ambitious, right?

If you think you're going to get five years of depreciation on AI chips, five years ago, we were still two two years away from Chat GPT.

Right.

Or think about what Jensen said at, um, we were at GTC this year.

He was talking about Blackwell and he said something about Hopper and he was like, eh, you don't want Hopper.

My sales guys are going to hate me, but like, you really don't want Hopper at this point.

I mean, these were the H100s.

This was the hot chip just when we were doing our most recent NVIDIA episode.

Yes.

Things move quickly.

Yes.

So.

I've seen estimates that over half the cost of running an AI data center is

the chips and the associated depreciation.

The human cost, that R ⁇ D is actually a pretty high amount because hiring these AI researchers and all the software engineering is meaningful, call it 25 to 33%.

The power is actually a very small part.

It's like 2% to 6%.

So when you're thinking about the economics of doing what Google's doing, it's actually incredibly sensitive to how much margin are you paying your supplier in the chips because it's the biggest cost driver of the whole thing.

So I was sanity checking some of this with Gavin Baker, who's the partner at Atreides Management to prep for this episode.

He's like a great public equities investor who's studied the space for a long time.

We actually interviewed him at the NVIDIA GTC pregame show.

And he pointed out normally, like in historical technology eras, it hasn't been that important to be the low-cost producer.

Google didn't win because they were the lowest cost search engine.

Apple didn't win because they were the lowest cost.

You know, that's not what makes people win.

But this era might actually be different because these AI companies don't have 80% margins the way that we're used to in the technology business, or at least in the software business.

At best, these AI companies look like 50% gross margins.

So Google being definitively the low-cost provider of tokens because they operate all their own infrastructure and because they have access to low markup hardware.

It actually makes a giant difference and might mean that they are the winner in producing tokens for the world.

Very compelling bull case there.

That's a weirdly winding analytical bullcase, but it's kind of the, if you want to really get down to it, they produce tokens.

Yep.

I've got one more bullet point to add to the bullcase for Google here.

Everything that we talked about in part two, the alphabet episode, all of the other products within Google, Gmail, Maps, Docs, Chrome, Android, that is all personalized data about you that Google owns that they can use to create personalized AI products for you that nobody else has.

Another great point.

So really the question to close out the bull case is, is AI a good business to be in compared to search?

Search is a great business to be in.

So far, AI is not.

But in the abstract, again, we're in the bullcase, so I'll give you this.

It should be.

With traditional web search, you type in two to three words.

That's the average query length.

length.

And I was talking to Bill Gross, and he pointed out that in AI chat, you're often typing 20 plus words.

So there should be an ad model that emerges, and ad rates should actually be dramatically higher because you have perfect precision.

Right.

You have even more intent.

Yes.

You know the crap out of what that user wants.

So you can really decide to target them with the ad or not.

And AI should be very good at targeting with the ad.

So it's all about figuring out the user interface, the mix of paid versus not, exactly what this ad model is.

But in theory, even though we don't really know what the product looks like now, it should actually lend itself very well to monetization.

And since AI is such a amazing, transformative experience, all these interactions that were happening in the real world or weren't happening at all, like answers to questions and being on a time spent, is now happening in these AI chats.

So it seems like the pie is actually bigger for digital interactions than it was in the search era.

So again, monetization should kind of increase because the pie increases there.

And then you've got the bull case of Waymo could be its own Google-sized business.

I was just thinking that, yeah, that's scoping all of this to a replacement to the search market.

Waymo and potentially other applications of AI beyond the traditional search market could add to that.

Right.

And then there's the like Galaxy brain bullcase, which is if Google actually creates AGI, none of this even matters anymore.

And like, of course, it's the most valuable thing.

That feels out of the scope for an acquired episode.

It's disconnected.

Yes, agree.

Bear case.

So far, this is all fun to talk about, but then the product shape of AI has not lent itself well to ads.

So despite more value creation, there's way less value capture.

Google makes something like $400-ish dollars per user per year, just based on some napkin math in the US.

That's a free service that everyone uses and they make $400-ish dollars a year.

Who's going to pay $400 a year for access to AI?

It's a very thin slice of the population.

Some people certainly will, but not every person in America.

Some people will pay 10 million, but right.

So if you're only looking at the game on the field today, I don't see the immediate path to value capture.

And think about when Google launched in 1998, it was only two years before they had AdWords.

They figured out an amazing value capture mechanism instantly.

Very quickly.

Yep.

Another bare case, think back to Google launch in 1998.

It was immediately, obviously, the superior product.

Yes.

Definitely not the case today.

No, there's four, five great products.

Google's dedicated AI offerings in chatbot was initially the immediately, obviously inferior product.

And now it's arguably on par with several others.

Right.

They own 90% of the search market.

I don't know what they own of the AI market, but it ain't 90%.

Is it 25%?

I don't know.

But at steady state, it probably will be something like 25%, maybe up to 50%.

But this is going to be a market with several big players in it.

So even if they monetized each user as great as they monetize it in search, they're just going to own way less of them.

Yep.

Or at least it certainly seems that way right now.

Yes.

AI might take away the majority of the use cases of search.

And even if it doesn't, I bet it takes away a lot of the highest value ones.

If I'm planning a trip, I'm planning that in AI.

I'm no longer searching on Google for things that are going to land Expedia ads in my face.

Or health, another

huge vertical.

Hey, I think I might have something that reminds me of mesothelioma.

Is it that or not?

Right.

Oh, where are you going to put the lawyer ads?

Maybe you put them there.

Maybe it's just an ad product thing, but these are very high-value queries, former searches that those feel like some of the first things that are getting siphoned off to AI.

Yep.

Any other bear cases?

I think the only other bear case I would add is that they have the added challenge now of being the incumbent this time around.

And people and the ecosystem isn't necessarily rooting for them in the way that people were rooting for Google when they were a startup.

And in the way that people were still rooting for Google in the mobile transition, I think the startups have more of the hearts and minds these days.

Right.

So I don't think that's quantifiable, but is just going to make it all a little harder path to row this time around.

Yep.

You're right.

They had this incredible PR and public love tailwind the first time around.

Yep.

And part of that's systemic too.

Like all of tech and all of big tech is just generally more out of favor with the country and the world now than it was 10 or 15 years ago.

They're just more important.

It's just big infrastructure.

It's not underdogs anymore.

Yep.

And that affects the OpenAIs and the Anthropics and the startups too, but I think to a lesser degree.

Yeah, they had to start behaving like big tech companies really early in their life compared to Google.

I mean, Google gave a Playboy interview during their quiet period of their IPO.

Times have changed.

Well, I mean, given all the drama at OpenAI, I don't know that I characterize them as acting like a mature company.

Fair.

Fair.

Company entity, whatever they are.

Yes.

Yeah.

But point taken.

Well, I worked most of my playbook into the story itself.

So you want to do power?

Yeah, great.

Let's move on and do power.

Hamilton Helmer's seven powers analysis of Google here in the AI era.

And the seven powers are scale economies, network economies, counter positioning, switching costs, branding, cornered resource, and process power.

And the question is: which of these enables a business to achieve persistent differential returns?

What entitles them to make greater profits than their nearest competitor sustainably?

Normally, we would do this on the business all up.

I think for this episode, we should try to scope it to AI products.

Yes, I agree.

Usage of Gemini, AI mode, and AI overviews versus the competitive set of Anthropic, OpenAI, Perplexity, Grok,

Meta MetaAI, et cetera,

scale economies for sure.

Even more so in AI than traditionally in tech.

Yeah, they're just way better.

I mean, look, they're amortizing the cost of model training across every Google search.

I'm sure it's some super distilled down model that's actually happening for AI overviews, but think about how many inference tokens are generated for the other model companies and how many inference tokens are generated by Gemini.

They just are amortizing that fixed training cost over a giant, giant amount of inference.

I saw some crazy charts.

We'll send it out to email subscribers.

In April of 24, Google was processing 10 trillion tokens across all their services.

In April of 25,

that was almost 500 trillion.

Wow.

That's a 50x increase in one year of the number of tokens that they're vending out across Google services through inference.

And between April of 25 and June 25, it went from a little under 500 trillion to a little under one quadrillion tokens, technically 980 trillion.

But they are now, because it's later in the summer, definitely sending out maybe even multiple quadrillion tokens.

Wow.

Wow.

So among all the other obvious scale economies things of amortizing all the costs of their hardware, they are amortizing the cost of training runs over a massive amount of value creation.

Yeah.

Scale economies must be the biggest one.

I find switching costs to be relatively low.

I use Gemini for some stuff, then it's really easy to switch away.

That probably stops being the case when it's personal AI to the point that you're talking about integrating with your calendar and your mail and all that stuff.

Yeah, the switching costs have not really

come out yet in AI products, although I expect they will.

Yes.

They have within the enterprise for sure.

Yep.

Network economies.

I don't think if anyone else is a Gemini user, it makes it better for me because they are sucking up the whole internet, whether anyone's participating or not.

Yep.

I agree.

I'm sure AI companies will develop network economies over time.

I can think of ways it could work, but yeah, right now, no, and arguably for the foundational model companies, can't think of obvious reasons right now.

Where does Hamilton put distribution?

Because that's a thing that they have right now that no one else has.

Despite ChatGPT having the Kleenex brand, Google distribution is still unbelievable.

I don't know, is that a cornered resource?

Cornered resource, I guess.

Yeah.

Definitely of that.

Yeah, Google search is a cornered resource for sure.

Certainly don't have counter positioning.

They're getting counter positioned.

Yep.

I don't think they have process power unless they were like coming up with the next transformer reliably, but I don't think we're necessarily seeing that.

There's great research being done at a bunch of different labs.

Branding, they have.

Yeah.

Branding is a funny one, right?

Well, I was going to say it's a little bit to my bear case point about they're the incumbent.

It cuts both ways, but I think it's net positive.

Yeah, probably.

For most people, they trust Google.

Yeah, they probably don't trust these who knows AI companies, but I trust Google.

I bet that's actually stronger than any downsides, as long as they're willing to still release stuff on the cutting edge.

Yep.

So to sum it up, it's scale economies is the biggest one.

It's branding, and it's a cornered resource.

And potential for switching costs in the future.

Yep.

Sounds right to me.

But it's telling that it's not all of them.

You know, in search, it was like very obviously all of them or most of them.

Yep.

Quite telling.

Well, I'll tell you, after hours and hours spending multiple months learning about this company, my quintessence, when I boil it all down, is just that this is the most fascinating example of the innovator's dilemma ever.

I mean, Larry and Sergei control the company.

They have been quoted repeatedly saying that they would rather go bankrupt than lose at AI.

Will they really?

If AI isn't as good a business as search, and it kind of feels like, of course it will be, of course it has to be.

It's just because of the sheer amount of value creation.

But if it's not, and they're choosing between two outcomes, one is fulfilling our mission.

of organizing the world's information and making it universally accessible and useful and

having the most profitable tech company in the world, which one wins?

Because if it's just the mission, they should be way more aggressive on AI mode than they are right now and full flip over to Gemini.

It's a really hard needle a thread.

I'm actually very impressed at how they're managing to currently protect the core franchise, but it might be one of these things where it's being eroded away at the foundation in a way that just somehow isn't showing up in the financials yet.

I don't know.

Yep.

I totally agree.

And in fact, perhaps influenced by you, I think my quintessence is a version of that too.

I think if you look at all the big tech companies, Google, as unlikely as it seems, given how things started, is probably doing the best job of trying to thread the needle with AI right now.

And that is incredibly commendable to Sundar and their leadership.

They are making hard decisions like we're unifying deep mind and brain.

We're consolidating and standardizing on one model.

And we're going to ship this stuff real fast

while at the same time not making rash decisions.

It's hard.

Rapid, but not rash, you know?

Yes.

And obviously we're still in early innings of all this going on, And we'll see in 10 years where it all ends up.

Yeah.

Being tasked with being the steward of a mission and the steward of a franchise with public company shareholders is a hard dual mission.

And Sundar and the company is handling it remarkably well, especially given where they were five years ago.

Yep.

And I think this will be one of the most fascinating examples in history to watch it play out.

Totally agree.

Well, thus concludes our Google series for now.

Yes.

All right, let's do some carve-outs.

All right.

Let's do some carve-outs.

Well, first off, we have a very, very fun announcement to share with you all.

The NFL called us.

We're going to the Super Bowl, baby.

Acquired is going to the Super Bowl.

This is so cool.

It's the craziest thing ever.

The NFL is hosting a innovation summit the week of the Super Bowl, the Friday before Super Bowl Sunday.

The Super Bowl is going to be in San Francisco this year in February.

And so it's only natural coming back to San Francisco with the Super Bowl that the NFL should do an innovation summit.

Yep.

And we're going to host it.

That's right.

So the Friday before, there's going to be some great on-stage interviews and programming.

Most of you, you know, we can't fit millions of people in a tidy auditorium in San Francisco the week of the Super Bowl when every other venue has tons of stuff too.

So there will be an opportunity to watch that streaming online.

And as we get closer to that date in February, we will make sure that you all know a way that you can tune in and watch the MCing, interviewing, and festivities at hand Super Bowl week.

It's going to be an incredible, incredible day leading up to an incredible Sunday.

Yes.

Well, speaking of sport, my carve-out is I finally went and saw F1.

It is great.

I highly recommend anyone go see it, whether you're an F1 fan or not.

It is just beautiful cinema.

Amazing.

Did you see it in the theater or I did see it in the theater?

Yeah.

Wow.

I unfortunately missed the IMAX window, but it was great.

It was my first time being in a movie theater in a while.

And whether you watch it at home or whether you watch it in the theater, I recommend the theater, but it's going to be a great surround-sound experience wherever you are.

Oh, I haven't been to the movie theater since the Eras tour.

Ah.

Which I think is just more about the current state of my family life with two young children.

Yes.

My second one, some of you are going to laugh, is the Travel Pro suitcase.

Ah.

This is the brand that pilots and flight attendants use, right?

Maybe.

I think I've seen some of them use it.

Usually they use something higher end, like a Briggs and Riley or a Toomey or, you know, Travel Pro is not the most high-end suitcase, but I bought two really big ones for some international travel that we were doing with my two-year-old toddler.

And I must say, they're robust.

The wheels glide really well.

They're really smooth.

They have all the features you would want.

They're soft shell, so you can like really jam it full of stuff, but it's also a thick amount of protection.

So even if you do jam it full of stuff, it's probably not going to break.

This is approximately the most budget suitcase you could buy.

I mean, I'm looking at the big Honkin international check bag version.

It's $416 on Amazon right now.

I've seen it cheaper.

They have great sales pretty often.

Everything about this suitcase checked lots of boxes for me.

And I completely thought I would be the person buying the Ramoa suitcase or the something very high-end.

And this is just perfect.

So I think I may be investing in more Travel Pro suitcases.

More Travel Pro.

Nice, nice.

Well, I mean, hey, look, for family travel, you don't want nice stuff.

Yeah.

I mean, I bought it thinking like, I'll just get something crappy for this trip.

But it's been great.

I don't understand why I wouldn't have a full lineup of Travel Pro gear.

So

this is my like budget pick gone right that I highly recommend for all of you.

I love how Acquired is turning into the wire cutter here.

That's it for me today.

Great.

All right.

I have two carve-outs.

I have one carve-out and then I have a update in my ongoing Google Carve Out saga.

But first, my actual carve-out.

It is the Glue Guys podcast.

Oh, it's great.

Those guys are awesome.

So great.

Our buddy Ravi Gupta, partner at Sequoia, and his buddies, Shane Badier, the former basketball player, and Alex Smith, the former quarterback for the 49ers and the Kansas City Chiefs and the Redskins.

Their dynamic is so great.

They have so much fun.

Half of their episodes, like us, are just them, and then half of their episodes are with guests.

Ben and I, we went on it a couple of weeks ago.

That was really fun.

When we were on it, we were talking about this dynamic of some episodes do better than others and pressure for episodes and whatnot.

And the guys brought up this interview they did with a guy named Wright Thompson.

And they said, Look, this is an episode.

It's got like 5,000 listens.

Nobody's listened to it.

It's so good.

And the mentality that we have about it is not that we're embarrassed that nobody listened to it.

It's that we feel sorry for the people who have not yet listened to it because it's so good.

I was like, that is the way to think about

your episode.

So here you are.

You're giving everyone the gift of.

I'm giving everyone the gift because then I was like, all right, well, I got to go listen to this episode.

Ray Thompson, I didn't know anything about him before.

I probably read his work in magazines over the years without realizing it.

He's the coolest dude.

He has the same accent as Bill Gurley.

So listening to him sounds like listening to

like if Bill Gurley, instead of being a VC, only wrote about sports and basically dedicated his whole life to understanding the mentality and psychology of athletes and coaches.

It's so cool.

It's so cool.

It's a great episode.

Highly, highly, highly recommend.

All right.

Legitimately, I'm queuing that up right now.

Great.

That's my carve out.

And then my ongoing family video gaming saga.

In Google part one, I said I was debating between the switch 2 and the Steam Deck.

Well, that's right.

First, you got the Steam Deck because you decided your daughter actually wasn't old enough to play video games with you.

So you just got the thing for you.

The update was I went with the Steam Deck for that reason.

I thought if it's just for me, it would be more ideal.

I have an update.

You also got a switch.

No, not yet.

Okay.

Okay.

But the most incredible thing happened.

My daughter noticed this device that appeared in our house that dad plays every now and then.

And we were on vacation.

And I was playing the Steam Deck.

And she was like, what's that?

Well, let me tell you.

And I was playing, I've been playing this really cool indie old school style RPG called Sea of Stars.

It's like a Chrono Trigger style, Super Nintendo style RPG.

I'm playing it.

And my daughter comes up and she's like, can I watch you play?

And I'm like, hell yeah, you can watch me play.

I get to play video games and you sit here and snuggle with me.

And like, you know, amazing.

I get to play video games and call it parenting.

Then it gets even better.

Probably like two weeks ago, we're playing and she's like, hey, dad, can I try?

I'm like, absolutely you can try.

I hand her the Steam Deck and

it was the most incredible experience.

One of the most incredible experiences I've had as a parent because she doesn't know how to play video games.

And I'm watching her learn how to like use a joystick and hit the button.

Supervised learning.

Yeah, yeah, yeah.

It's supervised learning.

I'm telling her what to do.

And then within two or three nights, she got it.

She doesn't even know how to read yet, but she figured it out.

It's awesome.

I'm watching her in real time.

And so now the last week, it's turned to mostly she's playing.

And I'm like helping her asking questions of like, well, what do you think you should do here?

Like, you know, should you go here?

I think this is the goal.

I think this is where it's so, so fun.

So I think I might actually pretty soon, her birthday's coming up, end up getting a Switch so that we can play, you know, together on the Switch.

Right.

But unintentionally, the Steam Deck was the gateway drug for my soon-to-be four-year-old daughter.

That's awesome.

There you go.

Parent of the year right there, getting to play video games.

And

oh, honey, I got it.

I'll take it.

Oh, yeah, I got it.

I got it.

All right, well, listeners, we have lots of thank yous to make for this episode.

We talked to so many folks who were instrumental in helping put it together.

First, a thank you to our partners this season: JP Morgan Payments, trusted, reliable payments infrastructure for your business, no matter the scale.

That's jpmorgan.com/slash acquired.

Sentry, the best way to monitor for issues in your software and fix them before users get mad.

That's sentry.io slash acquired.

Work OS, the best way to make your app enterprise ready, starting with single sign-on in just a few lines of code, workos.com.

And Shopify, the best place to sell online, whether you're a large enterprise or just a founder with a big idea, shopify.com slash acquired.

The links are all in the show notes.

As always, all of our sources for this episode are linked in the show notes.

Yes.

First, Stephen Levy at Wired and his great classic book on Google in the Plex, which has been an amazing source for all three of our Google episodes.

Definitely go buy the book and read that.

Also, to Parmi Olson at Bloomberg for her book, Supremacy about Deep Mind and Open AI, which was a main source for this episode.

And I guess also to Cade Metz, right?

For genius makers.

Yeah.

Yeah.

Great book.

Our research thank yous.

Max Ross, Liz Reed, Josh Woodward, Greg Corrado, Sebastian Thrun, Anna Patterson, Brett Taylor, Clay Bavour, Demis Hasabis, Thomas Kurian, Sundar Pachai.

A special thank you to Nick Fox, who is the only person we spoke to for all three Google episodes for research.

We got the hat trick.

Yeah.

To Arvin Navarotnam at Worldly Partners for his great write-up on alphabet linked in the show notes.

To Jonathan Ross, original team member on the TPU and today the founder and CEO of Grok.

That's Grok with a Q, making chips for inference.

To the Waymo folks, Dmitri Doglov and Suzanne Fillion.

To to Gavin Baker from Atreides Management, to MG Siegler, writer at Spyglass.

MG is just one of my favorite technology writers and pundits.

OG TechCrunch writer.

That's right.

To Ben Idelson for being a great thought partner on this episode and his excellent recent episode on the Step Change podcast on the history of data centers.

I highly recommend it if you haven't listened already.

It's only episode three for them of the entire podcast, and they're already getting, I don't know, 30, 40,000 listens on it.

I mean, this thing is taking off.

Amazing.

Dude, that's way better than we were doing on episode three.

It's way better than we were doing.

And if you like Acquired, you will love the Step Change podcast.

And Ben is a dear friend.

So highly recommend checking it out.

To Korai Kovakchalu from the DeepMind team, building the core Gemini models.

To Shashir Maroda, the CEO of Grammarly, formerly ran product at YouTube.

To Jim Gao, the CEO of Phaedra and former DeepMind team member.

Chaithing Putagunta, partner at Benchmark, Dwarkash Patel, for helping me think through some of my conclusions to draw.

draw, and to Brian Lawrence from Oak Cliff Capital for helping me think about the economics of AI data centers.

If you like this episode, go check out our episode on the early history of Google and the 2010s with our alphabet episode, and of course, our series on Microsoft and NVIDIA.

After this episode, go check out ACQ2 with Toby Lutke, the founder and CEO of Shopify, and come talk about it with us in the Slack at acquired.fm slash Slack.

And don't forget, our 10th anniversary celebration of Acquired.

We are going to do an open Zoom call, an LP call, just like the days of yore.

With anyone, listeners, come join us on Zoom.

It's going to be on October 20th at 4 p.m.

Pacific Time.

Details are in the show notes.

And with that, listeners, we'll see you next time.

We'll see you next time.

Who got the truth?

Is it you?

Is it you?

Is it you?

Who got the truth now?

Huh?