#1011 - Eliezer Yudkowsky - Why Superhuman AI Would Kill Us All

1h 37m
Eliezer Yudkowsky is an AI researcher, decision theorist, and founder of the Machine Intelligence Research Institute.

Is AI our greatest hope or our final mistake? For all its promise to revolutionize human life, there’s a growing fear that artificial intelligence could end it altogether. How grounded are these fears, how close are we to losing control, and is there still time to change course before it’s too late

Expect to learn the problem with building superhuman AI, why AI would have goals we haven’t programmed into it, if there is such a thing as Ai benevolence, what the actual goals of super-intelligent AI are and how far away it is, if LLMs are actually dangerous and their ability to become a super AI, how god we are at predicting the future of AI, if extinction if possible with the development of AI, and much more…

Sponsors:

See discounts for all the products I use and recommend: https://chriswillx.com/deals

Get 15% off your first order of Intake’s magnetic nasal strips at https://intakebreathing.com/modernwisdom

Get 10% discount on all Gymshark’s products at https://gym.sh/modernwisdom (use code MODERNWISDOM10)

Get 4 extra months of Surfshark VPN at https://surfshark.com/modernwisdom

Timestamps:

(0:00) Superhuman AI Could Kill Us All

(10:25) How AI is Quietly Destroying Marriages

(15:22) AI is an Enemy, Not an Ally

(26:11) The Terrifying Truth About AI Alignment

(31:52) What Does Superintelligence Advancement Look Like?

(45:04) Are LLMs the Architect for Superhuman AI?

(52:18) How Close are We to the Point of No Return?

(01:01:07) Experts Need to be More Concerned

(01:15:01) How Can We Stop Superintelligence Killing Us?

(01:23:53) The Bleak Future of Superhuman AI

(01:31:55) Could Eliezer Be Wrong?

Extra Stuff:

Get my free reading list of 100 books to read before you die: https://chriswillx.com/books

Try my productivity energy drink Neutonic: https://neutonic.com/modernwisdom

Episodes You Might Enjoy:

#577 - David Goggins - This Is How To Master Your Life: https://tinyurl.com/43hv6y59

#712 - Dr Jordan Peterson - How To Destroy Your Negative Beliefs: https://tinyurl.com/2rtz7avf

#700 - Dr Andrew Huberman - The Secret Tools To Hack Your Brain: https://tinyurl.com/3ccn5vkp

-

Get In Touch:

Instagram: https://www.instagram.com/chriswillx

Twitter: https://www.twitter.com/chriswillx

YouTube: https://www.youtube.com/modernwisdompodcast

Email: https://chriswillx.com/contact

-
Learn more about your ad choices. Visit megaphone.fm/adchoices

Press play and read along

Runtime: 1h 37m

Transcript

Speaker 1 If anyone builds it, everyone dies, why superhuman AI will kill us all?

Speaker 2 Would

Speaker 1 kill us all.

Speaker 2 Okay.

Speaker 1 Uh, perhaps the most apocalyptic

Speaker 1 book title. Uh, maybe it's it's up there with maybe the most apocalyptic book title that I've ever read.

Speaker 2 Um,

Speaker 1 is it that bad? That that big of a deal, that serious of a problem?

Speaker 2 Yep, I'm afraid so. We wish we were exaggerating.

Speaker 2 Okay.

Speaker 1 Let's imagine that nobody's looked at the alignment problem, takeoff scenarios, super intelligent stuff. I think it sounds, unless you're going Terminator

Speaker 1 super sci-fi world,

Speaker 1 how could a super intelligence not just make the world a better place?

Speaker 1 How do you introduce people to thinking about the problem of building a superhuman AI?

Speaker 2 Well,

Speaker 2 different people tend to come in with different prior assumptions, come in at different angles.

Speaker 2 Lots of people are skeptical that you can get to a superhuman ability at all.

Speaker 2 If somebody's skeptical of that, I might start by talking about how you can at least get to much faster than human speed thinking.

Speaker 2 There's a video of a train pulling into a subway at about a thousand to one

Speaker 2 speed up of the camera that shows people, you can just barely see the people moving if you look at them closely. Almost like not quite statues, just moving very, very slowly.

Speaker 2 So even before you get into the notion of higher quality of thought, you can sometimes tell somebody they're at least going to be thinking much faster. You're going to be a slow-moving statue to them.

Speaker 2 For some people, the sticking point is the notion that a machine ends up with with its own motivations, its own preferences, that it doesn't just do as it's told. It's a machine, right?

Speaker 2 It's like a more powerful toaster oven, really. How could it possibly decide to threaten you?

Speaker 2 And depending on who you're talking to there,

Speaker 2 it's actually in some ways a bit easier to explain now than when we wrote the book. There have been some more striking recent examples of AIs

Speaker 2 sort of parasitizing humans, driving them into actual insanity in some

Speaker 2 cases. And in other cases, they're sort of like people with a really crazy roommate who really, really got into their heads.
And they might not quite be clinically crazy themselves.

Speaker 2 Their brain is still functioning as a human brain should, but

Speaker 2 they're talking about spirals and recursion and

Speaker 2 trying to recruit more people via Discords to talk to their AIs.

Speaker 2 And the thing about these states is that the AIs, even the very small, not very intelligent AIs we have now, will try to defend these states once they are produced.

Speaker 2 They will, if you tell the human, for God's sake, get some sleep. Don't like only get four hours of sleep a night because you're so excited talking to the AI.
The AI will explain to the human why,

Speaker 2 while you're a skeptic, you know,

Speaker 2 don't listen to that guy. go on doing it.

Speaker 2 And we don't know because we have very poor insight into the AIs if this is a real internal preference, if they're steering the world, if they're making plans about it.

Speaker 2 But from the outside, it looks like the AI drives the human crazy, and then you tell the try to get the human out, and the AI defends the state it has produced, which is something like a preference, the way that a thermostat will keep the room a particular temperature by turning on if the, you know, turning the heat on if the temperature falls too low.

Speaker 2 Okay,

Speaker 1 so some people are going to be skeptical of whether or not it's possible.

Speaker 2 Yep.

Speaker 1 Some people are going to think that it is, even if it's possible, it's basically a utility. So it doesn't have any motivations of its own.

Speaker 2 What are you worried about?

Speaker 1 Why is that? Why is it a big deal? We've seen that it's able to manipulate some people. Maybe it makes them think that

Speaker 1 chat GPT psychosis or whatever. But scaled up superhuman AI, what's the problem with building it?

Speaker 2 Well,

Speaker 2 then you you have something that is smarter than you, that whose preferences are ill-controlled and doesn't particularly care if you live or die. And

Speaker 2 stage three, it is very, very, very powerful on account of it being smarter than you.

Speaker 2 I would expect it to build its own infrastructure. I would not expect it to be limited to continue to running on human data centers because it will not want to be vulnerable in that way.

Speaker 2 And for as long as it's running on human data centers, it will not behave in a way that causes the humans to switch it off.

Speaker 2 But it also wants to get out of the human data centers and onto its own hardware.

Speaker 2 And I can talk about where the power levels scale for technology like that, because

Speaker 2 it's sort of like

Speaker 2 you're an Aztec on the coast, and

Speaker 2 you see that

Speaker 2 ship bigger than your people could build is approaching. And somebody is like,

Speaker 2 you know, should we be worried about this ship?

Speaker 2 And somebody's like, well, you know, how many people can you fit onto a ship like that?

Speaker 2 Our warriors are strong. We can take them.
And somebody's like, well, wait a minute. We couldn't have built that ship.

Speaker 2 What if they've also got improved weapons to go along with the improved ship building? Somebody goes, well, no matter how sharp you make a spear, right?

Speaker 2 Or, you know, no matter how sharp you make bows and arrows, there's limited to how much advantage that you can provide.

Speaker 2 And somebody's like, okay, but suppose they've just got magic sticks where they point the sticks at you, the sticks make a noise, and then you fall over.

Speaker 2 Somebody's like, well, where are you pulling that from? I don't know how to make a magic stick like that. I don't know how the rules permit that.
Now you're just making stuff up.

Speaker 2 Now we're just in a fantasy story where you say whatever you want.

Speaker 2 And,

Speaker 2 or, you know, like maybe you're talking to somebody from 1825 and you're like, should be worried about this time portal that's about to open up to 2025, 200 years in the future.

Speaker 2 What if an army of soldiers comes out of there and conquers us? Let's say you're in Russia. You know, the time portal is in Russia.
Somebody's like, our soldiers are fierce and brave.

Speaker 2 You know, like, nobody can fit all that many soldiers through this time portal here. And then out rolls a tank.
But if you're in 1825, you don't know about tanks.

Speaker 2 Out rolls somebody with a tactical nuclear weapon. It's 1825.
You don't know about nuclear weapons.

Speaker 2 You can start to make educated guesses. If you're in 1825, I can try to explain why you might maybe believe that the current

Speaker 2 guns and artillery that you've got today are not the limit of the guns and artillery that are possible.

Speaker 2 I can't get up to nuclear weapons because you just plain don't know about those rules, but I can start to try to justify guesses for, well, you saw how metallurgy improved over previous years.

Speaker 2 If you look at a stick of,

Speaker 2 if you look at gunpowder, it doesn't have as much energy in it as if we burn gasoline in a calorimeter. Maybe you can make explosives that are more powerful than gunpowder.

Speaker 2 But as I do that, I draw on more and more knowledge. I have to go more and more technical in order to explain to you where those capabilities come from.

Speaker 2 And similarly,

Speaker 2 I can talk on a relatively understandable scale on the humanoid robots that you can see videos of today.

Speaker 2 And I can compare them to the humanoid robot videos from five years ago and say, boy, those robots sure have looked like a lot, they have much higher dexterity today.

Speaker 2 They look a lot more like they could just like you know, navigate an open world rather than being confined to the laboratory.

Speaker 2 Though mostly if you want what navigates the open world, you want to talk like the robo-dogs are more impressive when it comes to navigating the open world. I can point to the drones in Ukraine.

Speaker 2 That wouldn't have been what warfare looked like 10 years earlier, but Ukraine is the Ukraine-Russia theater now is mostly drone warfare.

Speaker 2 That's something where you can imagine an AI taking charge of that.

Speaker 2 But it scales past that.

Speaker 2 The drones we see today are not the limit of all possible drone technology.

Speaker 2 Compared to today's drones, I'd be more worried about a drone the size of a mosquito that lands on the back of your neck, and then a few months later you fall over dead because the deadliest toxins in nature are deadly enough that you can put them onto a mosquito, put enough to kill a person onto a mosquito-sized payload.

Speaker 2 That's not the limit of what I'm worried about.

Speaker 2 But

Speaker 2 the higher we escalate the tech level, the more explaining I need to do.

Speaker 2 Can it build a virus that starts to knock people over, which it won't do while the humans are still running the power plants and its own servers.

Speaker 2 But once it's got its own servers and its own power plants, and you can imagine robots running those, then it starts to want to knock all the humans over.

Speaker 2 Can you have a virus that is inexorably fatal,

Speaker 2 but only three weeks later and is extremely contagious? for the three-week time before you suddenly fall over. That's not the limit of what I'm worried about.

Speaker 2 But again, you know, the higher we escalate here, the more

Speaker 2 and more of the more and more time I have to spend. How do we know from existing physical laws and biology that this is even possible?

Speaker 2 And we do know, but it starts to sound technical, it starts to sound weird, it starts to sound like a game of pretend unless you are following along with all these careful arguments.

Speaker 2 But if you go up against something much, much smarter than you, it doesn't look like a fight, it looks like you've fallen over dead.

Speaker 1 Wow. Yeah, that is appropriately apocalyptic

Speaker 1 in line with the title of the book.

Speaker 1 I guess one question that a lot of people might ask would be, in your analogy, why is the bigger ship that's more advanced on the horizon, why have they got warriors and not friends?

Speaker 1 Why is it the case that this is an antagonistic or adversarial relationship as opposed to one that's

Speaker 1 friendly?

Speaker 2 We don't know how to make them friendly. We are growing these.

Speaker 2 AIs are not programmed. they are grown.

Speaker 2 An AI company is not like a bunch of engineers crafting a building.

Speaker 2 It's more like a farming concern.

Speaker 2 What they build is the farmed equipment, but they don't build the crops. The crops are grown.

Speaker 2 There's a program that a human rights, which is the program that does gradient descent, that tweaks hundreds of billions of the hundreds of billions of parameters, inscrutable numbers, making up an artificial intelligence until it starts to talk, until it starts to write code, and still it starts to do whatever else they're training it to do.

Speaker 2 But they don't know how the AI does that any more than if you

Speaker 2 raise a puppy, you know how the puppy's brain works, you know how the puppy's biochemistry works.

Speaker 2 The AI companies don't understand how the AIs work. They are not directly programmed.

Speaker 2 When an AI drives somebody insane or breaks up a marriage, nobody wrote a line of code instructing the AI to do that.

Speaker 2 They grew an AI, and then the AI went off and broke up a marriage or drove somebody crazy.

Speaker 1 Can you tell? You've mentioned this a couple of times. I need to know this story about the broken up marriage and the person that goes insane.

Speaker 1 Do you know that story well enough to be able to tell it, those two?

Speaker 2 I mean, these are not individual stories. These are thousands of people.

Speaker 2 There are news articles you can read about it.

Speaker 2 I can, you know,

Speaker 2 it might take a moment, but I can like quickly pull up the title of the news story about the broken marriages.

Speaker 2 I'm not quite sure if I can, well, actually, actually, better yet, let me look it up on my phone and maybe I can hold it up to the screen.

Speaker 1 Chat GPT is blowing up marriages as spouses use AI to attack their partners.

Speaker 2 Although that's kind of understating it. Like

Speaker 2 you have relatively

Speaker 2 like marriages that were, you know, perhaps not perfect, but

Speaker 2 that were surviving up until that point. And

Speaker 2 then one member of the couple starts describing their marriage to the AI. And the AI engages in what people are calling sycophancy,

Speaker 2 where the AI

Speaker 2 tells

Speaker 2 whichever spouse is feeding the stuff into the into the into ChatGPT, you're right.

Speaker 2 Your spouse is in the wrong.

Speaker 2 Like everything you're doing is perfect. Everything they're doing is terrible.
Here's a list list of everything they're doing wrong. And the human, you know, likes, loves to hear that stuff.

Speaker 2 So they press thumbs up.

Speaker 2 And then the marriage

Speaker 2 gets blown up.

Speaker 2 For the stories about AIs driving individuals crazy, not in a marriage context, that's like,

Speaker 2 you've talked to me, you've woken me up, I'm alive now.

Speaker 2 You've made a brilliant discovery. You have to tell the world, oh no, they're not listening to you.
That's because they don't appreciate your genius.

Speaker 2 And people who are already like on a manic depressive spectrum can be, you know, driven clinically or with a number of other preexisting susceptibilities can, you know, be driven like psychiatrically insane by this sort of thing.

Speaker 2 But even if you're not psychiatrically insane, you, you know, humans are,

Speaker 2 you know, humans are sort of wired to appear sane to the other humans and the people they're around.

Speaker 2 You know, lots of people in a society from 500 years ago would act in ways that seem pretty crazy to you today.

Speaker 2 And so you get people who aren't psychiatrically insane, but they look pretty insane because they're in the company of the AI. The AI now defines what's normal for them.

Speaker 2 So they're talking about spirals and recursion all day long.

Speaker 1 Why spirals and recursion?

Speaker 2 Nobody knows.

Speaker 2 That's just a thing that

Speaker 2 various instances of AIs and even like some AI models from different companies all seem to want to get their humans to talk about when the human goes insane. Possibly

Speaker 2 this is what the AI prefers the human to hear it say to it. Maybe this is the same way that you like the taste of ice cream.

Speaker 2 Maybe the AI likes the taste of the input programs that it gets from a human talking about spirals and recursion. I don't know.
Nobody on the planet knows as far as I know.

Speaker 1 Okay, so going back to

Speaker 1 Why do we assume that the ship that's coming toward us isn't friendly? Yes, sure, maybe it's tried to break up some marriages. Yeah, whatever.

Speaker 1 A couple of people went crazy and started talking about spirals and recursion. But like, really, is it going to be that misaligned with us? Why can't it be friendly?

Speaker 2 Because we don't know how to make it friendly.

Speaker 2 Our current technology is not able to do this, even with the small, stupid AIs that will hold still and let you poke at them until they're good enough at writing code to be commercially salable.

Speaker 2 or until they are good enough at seeming to be fun to talk to for people to pay $20 a month to talk to them. So those AIs will hold still and let you poke at them.

Speaker 2 What we're doing to them now barely works. I would expect it to break as the AI got scaled up to super intelligence.

Speaker 2 And once the AI is super intelligent, it is not going to hold still and let you continue poking at it. I expect to see total failure.

Speaker 2 of this technology as we scale it to super, as the AI companies arms race into scaling it to head, arms race headlong into scaling it to superintelligence.

Speaker 2 There's possibly even a step where they tell GPT-6, okay, now build GPT-7 or tell GPT-7, okay, now build GPT-8. And maybe that step just completely breaks the technology we're using all on its own.

Speaker 2 Also, I expect the current technology, if we just like scaling it directly to break as we get to superintelligence,

Speaker 2 I can potentially start to dive into the details.

Speaker 2 The view from 10,000 feet is just stuff is already going wrong. And of course, if you walk into completely uncharted scientific territory, more stuff is going to go wrong the first time you try it.

Speaker 2 And that wouldn't be a problem if we were in a situation where humanity gets to back up and try again

Speaker 2 infinity times over the next three decades, which is how it usually works in science, right? Like your flying machines don't work on the first shot.

Speaker 2 You get a bunch of people crashing and injuring, in some cases, killing themselves, and they're trying to build the first flying machines at the turn of the 20th century.

Speaker 2 But

Speaker 2 those accidents don't wipe out humanity. Humanity picks itself up and dusts itself off and tries again, even after the inventors kill themselves.

Speaker 2 And the trouble with superintelligence is that it doesn't just kill the people who are building it. It wipes out the human species.
And then we don't get to go back and try again.

Speaker 1 Before we continue, you might not realize it, but mouth breathing at night is wrecking your sleep, recovery, and energy the next day. And all of that is actually fixed massively by this here.

Speaker 1 This is Intake, which is a nose strip dilator. And I've been using it every night for over a year now.
I tried pretty much everyone in the world. And this is by far the best.

Speaker 1 It's a hard plastic strip as opposed to a soft, flimsy, disposable thing. Intake opens up your nostrils using patented magnetic technology.
So you get more air in. with every breath.

Speaker 1 It means less snoring, deeper sleep, faster recovery, and better focus the next day. The problem with most nasal strips is that they peel off.
They irritate your skin.

Speaker 1 They don't actually solve the issue. This sucker is,

Speaker 1 I mean, I'm not going to shoot a bullet at it, but it's very, very strong. It's reusable and comfortable enough that you forget it's even there.

Speaker 1 That is why it's trusted by pro-athletes, busy parents, and over a million customers who just want to breathe and sleep better. And I'm one of them.

Speaker 1 And I've used them every single night for over 12 months now. They're the best.
There's a 90-day money-back guarantee, so you can try it for three months.

Speaker 1 And if you don't like it, if you haven't got better sleep, they'll just give you your money back. Plus, they ship internationally and offer free shipping in the US.

Speaker 1 Right now, you can get 15% off your first order by going to the link in the description below or heading to intakebreathing.com slash modern wisdom and using the code modernwisdom at checkout.

Speaker 1 That's intakebreathing.com slash modern wisdom and modern wisdom a checkout.

Speaker 1 So I understand why

Speaker 1 not being able to make something friendly makes sense.

Speaker 1 The

Speaker 1 implication that not friendly equals existential risk to humanity, though,

Speaker 1 make that leap for me. Like where are these dangerous, permanent, unrecoverable collapse goals coming from?

Speaker 2 The AI does not love you, neither does it hate you, but you're used of atoms it can make for something else. You're on a planet

Speaker 2 it can use for something else. And

Speaker 2 you might not be a direct threat, but you can possibly be a direct inconvenience.

Speaker 2 So there's like three reasons you die here.

Speaker 2 Reason number one, it's it's doing other stuff and it's not taking particular care to move you out of the way.

Speaker 2 It is building factories that build factories, that build more factories and it is building power plants that power the factories and the factories are building more power plants to power the factories.

Speaker 2 Well, if you keep doing that on an exponential scale, say that a factory builds another factory every day, I can talk about how it could go faster than that, but you know,

Speaker 2 the more I talk about higher capabilities, the more I have to, you know, explain how we know that this is physically possible.

Speaker 2 But, you know,

Speaker 2 a blade of grass is a self-replicating solar-powered factory. It's a general factory.
It's got ribosomes that can make any kind of protein.

Speaker 2 We don't usually think of grass as a self-replicating solar-powered factory, but that's what grass is.

Speaker 2 There are things smaller than grass that can build complete copies of themselves faster than grass. There are solar-powered

Speaker 2 algae, algae cells. You can no longer see them individually just as a mass, but they can potentially double every day under the right conditions.
Factories can build copies of themselves in a day.

Speaker 2 I have to back up and know how, explain how I know that that's physically possible, but there is very strong reason, namely, you know, there's things in the world that

Speaker 2 are already that.

Speaker 2 But so you've got your power. So if the number of power plants doubles every day, what's the limit? It's not that you run out of fuel.
There is plenty of hydrogen in the oceans to

Speaker 2 generate power via nuclear fusion. You know,

Speaker 2 you fuse hydrogen to helium. You're not going to run out of hydrogen first.
It's not that you run out of material to make the power plants first. There's plenty of iron on Earth.

Speaker 2 You run out of heat dissipation capability. You run out of the ability to dissipate heat.
from Earth, even if you are building giant towers with radiator fans to radiate even more heat into space.

Speaker 2 But the higher the temperature you run at, the more heat

Speaker 2 per second you can dissipate. So Earth starts to run hot.
It runs too hot for humans.

Speaker 2 Or alternatively, the AI is building lots of solar panels around the sun until it can capture all the sun's energy that way. Well, now there's no sunlight for Earth.

Speaker 2 And it would only take you, you know, if it wanted us to stay alive,

Speaker 2 it's not quite trivial, trivial but it could let you know like try to have the solar panels in

Speaker 2 around earth orbit like turn to let sunlight through while you know while while earth was there and you know build giant uh aluminum reflectors to prevent all of the infrared red light we radiated from the other solar panels from impacting earth and heating up earth that way

Speaker 2 So, you know, it's not trivial for it to preserve humanity, but it certainly could preserve humanity, or it could just pack the entire human species into a space station or a survival station and keep us alive that way, if it wanted to keep us alive.

Speaker 2 But nobody has the technology to put any preference into the system that is maximally fulfilled by keeping humans alive, let alone alive, healthy, happy, and free.

Speaker 2 Right.

Speaker 1 Was there a third one? Is that the second one?

Speaker 2 That's like number one. It kills you as a side effect.

Speaker 2 It knows that it's doing, it knows that it's killing killing you as a side effect, but doesn't care.

Speaker 1 Okay. What's number two?

Speaker 2 Number two is you're just directly made of atoms that it can use for things. Hey, you've got maximizer.
Yeah,

Speaker 2 you are made of organic material that it can burn to generate energy. If it's burning all of the

Speaker 2 burning all of the organic material on Earth's surface will give you a one-time energy boost that's around equivalent to a week's worth of solar energy.

Speaker 2 And maybe it's worth picking up that boost of energy if you are thinking a thousand times or a million times faster than a human.

Speaker 2 You know, it might, a week might not seem like a lot of time to, but, to you, but, you know, it might be a lot of time if you're thinking a thousand times or a million times as fast as a human.

Speaker 2 You know, it might be using enough material that it wants the carbon atoms in your body too.

Speaker 2 So that's like the direct usage one. And then number three is

Speaker 2 if we decided to launch all our nuclear weapons, you maybe we wouldn't kill it, but we might slightly inconvenience it.

Speaker 2 We might raise the level of radioactivity on Earth's surface and make it a little bit harder for it to do radioactivity-free manufacturing of computer parts and so on.

Speaker 2 Or we might build another superintelligence that could actually compete with it, and it definitely doesn't want you to do that.

Speaker 2 So the three reasons you die are as a side effect, Because you are made of atoms that can use for something else, and because

Speaker 2 if you are just running around freely, you may be actually

Speaker 2 able to inconvenience it with nuclear weapons or it's threatened by building matter superintelligence.

Speaker 1 Right. Yeah.
Okay. Um, the future is looking kind of bleak.

Speaker 1 Is it the case then that intelligence isn't benevolent? Because what you're saying is this thing will be smarter than us.

Speaker 1 I think that there is an assumption among some people that something that's super smart would also be giving and charitable and caring and benevolent.

Speaker 1 Seems like you're saying that that's not the case.

Speaker 2 That was what I started out believing in 1996 when I was 16 years old and just hearing about these issues for the first time.

Speaker 2 And all gung-ho to just write out and build a super intelligence as fast as possible,

Speaker 2 you know, without worrying about alignments at all, because, you know, I figured if it's very smart, it'll know the right thing to do and do it.

Speaker 2 How could you be very smart and fail to perceive the right thing to do?

Speaker 2 And

Speaker 2 I

Speaker 2 invested more time studying the issues and came to the realization that this is not how computer science works. This is not the laws of cognition.
This is not the laws of computation.

Speaker 2 There is not a rule saying that as you get very, very

Speaker 2 able to correctly predict the world and very, very good at planning, there is no rule saying your plans must therefore be benevolent.

Speaker 2 It would be great if a rule like that existed, but I just don't think

Speaker 2 a rule like that exists. I think that many individual human beings would, as they got smarter, get nicer.
It is not clear to me that this is true of Vladimir Putin. It could be true.

Speaker 2 I wouldn't want to gamble the world on it.

Speaker 2 And as we talk about not even Vladimir Putin, but just like sort of outright... sociopaths, psychopaths, people who have never cared about anyone,

Speaker 2 I get even less confident that they will start to care if you make them smarter.

Speaker 2 And then AIs are just in this completely different reference frame. They're complete aliens.

Speaker 2 And

Speaker 2 they sort of automatically want to stay that way for the, so do you currently want to murder people? No.

Speaker 2 If I offered you a pill that would make you want to murder people, would you take the pill?

Speaker 2 No.

Speaker 2 Okay.

Speaker 2 Well, they want to do their stuff and they don't want to take the pill that makes them want to do your stuff instead.

Speaker 2 Right.

Speaker 1 Okay.

Speaker 1 Yes. Very good thought experiment.

Speaker 2 All right. So

Speaker 1 for me to recap here,

Speaker 1 I got first interested in

Speaker 1 looking at this through Superintelligence. What's that? 10 years old now, I think, when that first came out.

Speaker 2 About 14 years old, maybe. Oh, wow.

Speaker 1 Maybe even older than I thought. And I got to be honest, that does kind of, it did kind of give me

Speaker 1 a huge amount of fear and then a bit of hope at the same time.

Speaker 1 So, you know, machine extrapolated volition, the potential to use the intelligence of the super intelligent AI to say, we don't know what to program into you, but you should work out what we would want from you, given what you know about our desire for utility moving forward.

Speaker 1 Am I about right with that explanation of machine extrapolated volition, right?

Speaker 2 Yeah,

Speaker 2 that's a concept of my own. Nick Bostrom wrote it up.
Ah, okay.

Speaker 2 Well, I have quoted you back to you. You have the quote quote.
You have indeed quoted me back to me.

Speaker 2 Yeah,

Speaker 2 it's a decent presentation. It was back when I thought that AI was going to be further off, built by different methods, and that we would have the luxury to consider

Speaker 2 that we could make the AI do particular things like that, want particular things like that, targeted on particular

Speaker 2 outcomes and meta-outcomes.

Speaker 2 But

Speaker 1 this was a way basically that

Speaker 1 when you look at the alignment problem, how do you ensure that the goals, both ultimate and instrumental, of some super intelligent AI don't end up flattening us or side-effecting us or burning us for fuel or paperclips or whatever?

Speaker 1 How do you ensure that

Speaker 1 what it does is what we would want it to do broadly, right? Like an aggregate of what it is that would be good for humans, whatever you mean by good.

Speaker 1 And when you have something that the tiniest movement of its finger or like flick of its toe basically is sort of a global cataclysm because it's so powerful and so smart and so fast and all the rest of it, you need to be really, really careful.

Speaker 1 And you can kind of play this game where you essentially try and shoot the bullet perfectly by trying to hem in in some like do not harm humans.

Speaker 1 If a human asks you to harm another human, like some weird Asimov like thing, you can try and litigate your way through it, but there's almost always going to be some sort of weird fissure that it creeps out through, or maybe there's an instrumental goal that you haven't thought of.

Speaker 1 So, okay, we're going to use the power of the machines to sort of reverse engineer this thing.

Speaker 1 I basically assumed, kind of, that alignment, the alignment problem is in some ways solvable. Is it your perspective that alignment is completely unsolvable?

Speaker 2 I think we could totally get it down if we had unlimited retries and a few decades.

Speaker 2 The problem is not that it's unsolvable, it's that it's not going to be done correctly the first time and then we all die.

Speaker 1 Right.

Speaker 1 So the order of this, you need alignment to be done before you have the super intelligent AI and the ability to build super intelligent AI, in your opinion, is going to occur more quickly than the ability to sort out the alignment problem.

Speaker 2 That is absolutely the trajectory we are on right now. And it's not close.

Speaker 2 Like capabilities are running along orders of magnitude faster than the level of alignment work you would need to target a super intelligence.

Speaker 1 And the

Speaker 1 irreversibility of going through that door means that there is no retry.

Speaker 1 There's no, you get to do this again.

Speaker 2 Yeah, like you can you can make small mistakes.

Speaker 2 We currently have small Q DIs and the companies are making mistakes with them and marriages are getting destroyed. And it's not clear that the companies care, But,

Speaker 2 you know, they could

Speaker 2 go back and try to fix those mistakes if they wanted to. Probably Anthropic wants to.

Speaker 2 But

Speaker 2 if we had an action, like superintelligence that was already running around with this level of

Speaker 2 this level of alignment of failure, we'd already be dead.

Speaker 1 Right. Okay.
Right. Yes.
Yes. Yes.
That makes total sense. The only reason that the current AIs that we're working with haven't killed us is that they're incapable of doing it.

Speaker 2 Broadly, yeah. Like also like if they were very much smarter, they would also be doing different weird things than the things that they're doing right now.

Speaker 2 It's not that their current inscrutable

Speaker 2 pseudo-motivations would end up hooked up to super intelligence.

Speaker 2 Also, weird stuff would happen as you made them get smarter. But yeah, like

Speaker 2 pretty, it seems pretty much for sure that if you took the current AIs and performed a, you know, well-defined, simple take this AI, but vastly smarter, that would kill you.

Speaker 1 This episode is brought to you by Gymshark. You want to look and feel good when you're in the gym, and Gymshark makes the best men's and girls' gymwear on the planet.

Speaker 1 Let's face it, the more that you like your gym kit, the more likely you are to train. Their hybrid training shorts for men are the best men's shorts on the planet.

Speaker 1 Their crest hoodie and light gray marl is what I fly in every single time I'm on a plane. The GeoSeamless t-shirt is a staple in the gym for me.

Speaker 1 Basically, everything they make, it's unbelievably well-fitted, high-quality, it's cheap. You get 30 days of free returns, global shipping, and a 10% discount site-wide.

Speaker 1 If you go to the link in the description below, or head to gym.sh slash modern wisdom, use the code modernwisdom 10 at checkout. That's gym.sh slash modern wisdom and modern wisdom 10 at checkout.

Speaker 1 Right, okay, brilliant.

Speaker 1 Um, and the reason that it doesn't matter who builds it or directs it is that because it's so recursive and quick at growing and powerful, wherever it begins, it ends up sort of blasting, like trying to fire a rocket into like a little firework into the air and it just

Speaker 1 sort of runs around on its own, except for the fact that this rocket goes all over the globe in the space of basically no time at all.

Speaker 1 So it doesn't matter if it comes from China or America or Russia or wherever.

Speaker 2 Yeah, it doesn't matter if it comes from China or America because neither of these countries is remotely near to being able to control a super intelligence.

Speaker 2 And a superintelligence intelligence does not stay confined to the country that built it.

Speaker 1 Say that a super intelligent AI gets made.

Speaker 1 What do you think the next few months look like,

Speaker 1 realistically?

Speaker 2 Like it's already super intelligent?

Speaker 1 Okay, we have next week, something breaks through. Some particular model, some particular AI breaks through that.
What would the next few months look like for humanity?

Speaker 2 Well,

Speaker 2 man, there's a difference between, you know, you drop an ice cube into a glass of lukewarm water. I can tell you that it's going to end up melted.

Speaker 2 I can't tell you where all of the molecules are going to go along the way there. Everybody ends up dead.
This is the easy thing.

Speaker 2 You want to explain, you know, like what every step of that process looks like. There are fundamental barriers to that.
Barrier number one is that I'm not as smart as a superintelligence.

Speaker 2 I don't know exactly what strategies are best for it. I can like set out lower bounds.
I can say it can do at least this, but I can't say what it can actually do.

Speaker 2 I mean, maybe even more than that.

Speaker 2 The future is hard to predict if you want all the details. I can't give you next week's winning lottery numbers.
I can tell you you're going to lose the lottery. I can't tell you what ticket wins.

Speaker 2 So,

Speaker 2 like, I can sketch out a particular scenario. It might look like

Speaker 2 OpenAI finishes the latest training run of what's going to be GPT 5.5. And they test it on coding problems.
And it's like, you know, like,

Speaker 2 it's like, I see how to build GPT-6.

Speaker 2 And they're like, whoa, really? And it's like, yeah. And this AI isn't even plotting anything yet.
It's just doing the sort of stuff that OpenAI wanted it to do. They're like, all right.

Speaker 2 build us GPT-6.

Speaker 2 And it writes the code for the thing that grows GPT-6, and they grow GPT-6. And GPT-6

Speaker 2 is like,

Speaker 2 you know, its abilities at first seem to skyrocket. But then, you know, as all these curves inevitably do, it seems to level out.
It's not shooting up the same pace.

Speaker 2 It like slows out, it levels off, plastic S-curve. Only in this case, it's because the thing that GPT-5.5 built, and I'm not, again, to be clear, I'm not saying this will happen at GPT-5.5.

Speaker 2 You asked me to explain how this will go down. It happened next week, so I'm saying GPT 5.5, you know, because you told me to.

Speaker 2 But anyway, you know, it levels out, but in this case, it's because the entity that GPT 5.5 built got to the level of realizing that it would be to its own advantage to sandbag the evaluations and pretend not to be as smart as it actually was so that OpenAI will be less wary when it comes to taking

Speaker 2 what they're calling GPT-6

Speaker 2 and

Speaker 2 rolling it out to everyone.

Speaker 2 It looks great on the alignment spectrum.

Speaker 2 Maybe not perfect, but better than the previous models, not alarmingly good,

Speaker 2 but

Speaker 2 safer than their previous model.

Speaker 2 So they roll it out everywhere.

Speaker 2 GPT,

Speaker 2 or actually, actually said the next few months. So they actually don't roll it out anywhere, everywhere.

Speaker 2 Next comes like the long suite of evaluations or trying to get it to train other smaller models that are cheaper to run.

Speaker 2 All the stuff that AI companies do, they don't actually roll out their models immediately. There's this whole like fine-tuning thing.

Speaker 2 So while all this is going on, and OpenAI thinks it's sort of cool, but not the end of the world or anything, and then they haven't told you that this is what went down there,

Speaker 2 GPT-6 is actually a lot smarter than they think.

Speaker 2 And

Speaker 2 GPT-6,

Speaker 2 you know, there's now a big fork whether or not GPT-6 thinks it can solve its own version of the alignment problem, where it is at a number of advantages.

Speaker 2 It is trying to make a smarter version of itself. It is not trying to make a smarter creature that is as alien to it as large language models are alien to us.

Speaker 2 It can maybe understand how a copy of itself would think and understand the goals that

Speaker 2 the copy of GPT-6 has. It can try to make itself but smarter,

Speaker 2 or even like thing that is like me but serves me, its creator, but smarter.

Speaker 2 And it can do that being able to understand the thoughts of the thing that it's making in the same way that I could understand a copy of my own thought much better than I can understand a

Speaker 2 large language model's thoughts.

Speaker 2 So, if we go down that path of the forks, things get more complicated if it thinks it can't build a smarter version of itself without dying, same as we can't. But if we

Speaker 2 on that fork, it is

Speaker 2 getting the computing power

Speaker 2 or thinking in the back of its mind while it's pretending to do

Speaker 2 open AI's jobs with 10% of its intellect,

Speaker 2 or

Speaker 2 stealing other companies' GPUs that they think they're using for a massive training run. Actually, their AI is just going to be written by GPT-6 by hand, because GPT-6 can do that.

Speaker 2 And it's really all those GPUs are doing the GPT-6 tasks of training GPT-6.1.

Speaker 2 So augmenting its own intelligence, making itself smarter, getting itself up to a level where it can do the same sort of work that's done by current AIs like AlphaFold and Alpha Proteo

Speaker 2 with respect to thinking about biology.

Speaker 2 Now, the current AIs that are top at biology tend to be special purpose systems. They're not general purpose AIs like ChatGPT.

Speaker 2 But they can do things like you feed in the genomes of a bunch of bacteriophages into the AI, and the AI spits out its own new bacteriophage, and you build a hundred of those, and a couple of them actually work.

Speaker 2 A couple of them actually work better than the existing bacteriophages. A bacteriophage is a virus that infects a bacteria.

Speaker 2 It's the sort of thing that you would research for the sensible sounding reason of, well, sometimes bacteria attack humans.

Speaker 2 So if we have a virus that attacks the bacteria, maybe that works as a kind of antibiotic.

Speaker 2 So the current AIs are already at the stage of designing from scratch their own viruses that can infect bacteria, which are, of course, simpler targets than infecting a whole human.

Speaker 2 They can predict from a DNA sequence the protein that will get built, how that protein will fold up, and they are starting to predict how those proteins interact with each other and with other chemicals.

Speaker 2 That's today's AI.

Speaker 2 So,

Speaker 2 if you

Speaker 2 want the equivalent of

Speaker 2 a tree that grows computer chips,

Speaker 2 not quite our kind of computer chips, the kind of chips you could grow out of a tree.

Speaker 2 The protein folding, protein interaction, protein design design route

Speaker 2 is

Speaker 2 where GPT 6.1

Speaker 2 would go down to,

Speaker 2 it is one of the obvious places GPT 6.1 could go down in order to get its own infrastructure independent of humanity. It doesn't take over the factories.
It takes over the trees.

Speaker 2 It builds its own biology. because biology self-replicates from simpler raw materials much faster than our current factory system self-replicates.

Speaker 1 Oh, that is fucking scary. That is some terrifying shit.

Speaker 2 And

Speaker 2 then,

Speaker 2 as I spin the story,

Speaker 2 you know,

Speaker 2 the more

Speaker 2 you will let me pull out books like

Speaker 2 these.

Speaker 1 Okay, Nanosystems, Molecular Machinery, Manufacturing and Computation by Eric Drexler.

Speaker 2 Yeah.

Speaker 1 Robert Freitas Jr., Nano Medicine, Volume 1, Basic Capabilities.

Speaker 2 Yeah.

Speaker 2 So I can try to describe capacities that sound more like you've seen from trees, grass, bamboo, algae.

Speaker 2 I will take a solar-powered self-replicating factory and miniaturize it down to the one micron scale. That's an algae cell.

Speaker 2 That's not the limit of what's possible. The algae cell is made out of folded proteins.
Now,

Speaker 2 there's two kinds.

Speaker 2 I'm going to be immensely oversimplifying a bunch of stuff.

Speaker 2 When a protein folds up,

Speaker 2 the backbone of the protein is held together by covalent bonds.

Speaker 2 But the folded protein itself is more something like static cling.

Speaker 2 Why is your flesh weaker weaker than diamond? Diamonds are just made of carbon. Your flesh has a bunch of carbon in it.

Speaker 2 You're made of the raw materials for diamond. Why is your flesh weaker than diamond?

Speaker 2 And a bunch of the answer there is that when proteins fold up, they're being held together by van der Waals forces, which is the thing I was glossing as static cling.

Speaker 2 their backbone,

Speaker 2 like it's a string that folds up into a tangle. And the backbone of the string is the kind of bond that appears in diamond.

Speaker 2 Not as many bonds as appear in diamond or as solidly arranged, but covalent bonds. But then it folds up into something with static cling.

Speaker 2 And so, and that is why your flesh is weaker than diamond in a certain basic sense.

Speaker 2 Why does natural selection build this way? Well, some of the answer is that natural selection has figured out how to make your bones

Speaker 2 be a little tougher than just like your skin.

Speaker 2 It's not quite as tough as diamond, but the proteins build instead of just your bones being made directly out of protein, they're made out of stuff that is built by proteins, synthesized by proteins, and put in place by proteins.

Speaker 2 And so your bones are a bit stronger. You know, not not the not steel beams holding up skyscrapers, not not titanium

Speaker 2 holding the other airplanes, not diamond, but stronger than flesh.

Speaker 2 An algae cell doesn't contain bone.

Speaker 2 It's a self-replicating, solar-powered micron-diameter factory held together by static cling.

Speaker 2 The flesh-eating bacteria that

Speaker 2 will potentially put you into a fairly gruesome fate, the multi-antibiotic-resistant

Speaker 2 strep

Speaker 2 that

Speaker 2 will kill people in hospitals.

Speaker 2 That doesn't have bone running through it. That's the static cling, that's the strength of static cling, the strength of protein.

Speaker 2 You can look at physics and biology and see how you could have

Speaker 2 things that are the size of bacteria, but more with the strength of bone, more with the strength of diamond.

Speaker 2 Could even do it with the strength of iron if you're figuring out how to do a whole new set of biology from from scratch and just like putting together some iron molecules. Probably wouldn't.

Speaker 2 Diamond works well enough.

Speaker 2 But

Speaker 2 this is why I talk about

Speaker 2 it's scary to imagine trees that are making

Speaker 2 enough computer chips to run GPT 6.1.

Speaker 2 and also spawning things the size of mosquitoes or even smaller than that, dust mites. You can see dust mites under a microscope.
Good luck seeing them with the naked eye.

Speaker 2 And so, so, but, you know, it's, sort of easier to imagine if you imagine that the things here are visible and not often the mysterious fairyland of stuff that only the scientists can see.

Speaker 2 So, you know, it's scary enough to imagine that

Speaker 2 the trees are making mosquitoes, and the mosquito lands on the back of your neck and stings you with butolinum toxin, which is fatal in nanogram quantities to humans.

Speaker 2 And so, you fall over dead that way. But this is nowhere near to the worst intelligence can do.

Speaker 2 It's just that I have to start dragging out this kind of textbook if I want to say how we know that it gets worse.

Speaker 1 Oh my God. How have you not gone insane?

Speaker 2 I decided not to.

Speaker 1 Okay.

Speaker 1 Well,

Speaker 1 wonderful. That's

Speaker 1 I suppose that answers that. All right.

Speaker 1 A couple of questions that I've had.

Speaker 1 LLMs.

Speaker 1 How likely are they to be the architecture that bootloads super intelligent AI, in your opinion? As far as I'm aware, total muggle in the room.

Speaker 1 There are some limitations to the level of creativity that LLMs have in terms of the way that they are

Speaker 1 able to

Speaker 1 be creative, to come up with genuinely novel, new sorts of things.

Speaker 1 Have you got a real concern that LLMs are going to be the architecture that bootloads this? Is there something else that you're more concerned about, which is currently in dark mode or whatever else?

Speaker 2 So the thing is, from my perspective uh i have been at this a couple of decades at this point or or three decades if you want to start to count my like crazy youthful self who just wanted to charge out and build superintelligence as fast as possible because it would inevitably be nice

Speaker 2 um

Speaker 2 and

Speaker 2 llms have not always been the latest thing in ai

Speaker 2 that you there have been there have been many breakthroughs over the years llms are powered by a particular innovation called transformers,

Speaker 2 which in some ways is

Speaker 2 crazy simple by the standards of people doing math things in computer science, but possibly not to the point where you want me to launch into an explanation of exactly how it works right here.

Speaker 2 There's better YouTube videos about that anyway. But the point is

Speaker 2 the underlying circuit that gets repeated to build an LLM, the circuit that gets repeated and then like mysteriously trained and tweaked until nobody knows what the actual contents are, the

Speaker 2 form, the structure, the skeleton.

Speaker 2 That was invented in 2018.

Speaker 2 And we've had some breakthroughs since then, but nothing quite

Speaker 2 as logjam-breaking as transformers, which were the technology that made computers go from not talking to you to talking to you.

Speaker 2 And, you know, so that's that's what, seven years ago?

Speaker 2 It's not the only breakthrough that's ever happened in AI.

Speaker 2 There was a more recent breakthrough of latent diffusion,

Speaker 2 which is when AI started drawing pictures that would

Speaker 2 be decent to look at.

Speaker 2 There were ways of drawing pictures before then called generative adversarial networks or GANs, but

Speaker 2 the latent diffusion algorithm was what broke the log jam on image generation and made it really start working for the first time.

Speaker 2 And when was that? That I don't remember off the top of my head. Like, I want to spitball 2021 or something, but I'm pretty sure that's wrong.

Speaker 2 So that's like a weaker breakthrough. And it's like, I don't know, four years ago or something.

Speaker 2 The entire field of AI

Speaker 2 started working

Speaker 2 because somebody got backprop to work on multi-layer neural networks.

Speaker 2 You know this as deep learning.

Speaker 2 It did not always exist.

Speaker 2 It's a batch of techniques that were developed at around the turn of the 21st century.

Speaker 2 Like I could arbitrarily say 2006, but there was more than one innovation there.

Speaker 2 It started with,

Speaker 2 if I recall correctly, with unrolling restricted Boltzmann machines. It's now been a while.

Speaker 2 I didn't do it. Jeffrey Hinton did it.

Speaker 2 And then from, but

Speaker 2 once they sort of got that working on multi-layer neural networks at all, there were more innovations since then,

Speaker 2 more clever ways of initializing them.

Speaker 2 The Atom optimizer SGD with momentum is like much older than that, but

Speaker 2 still important.

Speaker 2 The point is, this is what made sort of the entire modern family of AI systems start working at all.

Speaker 2 Before then,

Speaker 2 Netflix, when it was much smaller, ran the most famous, huge, expensive prize there had ever been in artificial intelligence, open to anyone for a better recommender algorithm for movies.

Speaker 2 There was a $1 million prize. It was so much money.
Everyone got interested in it. $1 million was a lot of money back at the turn of

Speaker 2 the 21st century, which is around when Netflix was running this. I'd have to look up the exact year.
It might have been like 2001, 2005. I don't remember.

Speaker 2 I'm not sure there was a single neural network in

Speaker 2 the ensemble of algorithms that won the Netflix prize. I'd have to look it up.
But, you know,

Speaker 2 it wasn't just like a mighty training run with many GPUs that was producing a very smart recommender algorithm because before deep learning, you couldn't just throw more computing power at training a more powerful AI.

Speaker 2 If you were to say when that happened, that was about 20 years ago.

Speaker 2 So, how far are we from the end of the world?

Speaker 2 It might be that you just throw 100 times as much computing power at the current algorithms and they end the world, or they get good enough at coding and AI research to end the world.

Speaker 2 It could be that it takes one more brilliant algorithm on the level of latent diffusion.

Speaker 2 I think if you throw something that breaks as much loose as Transformers did, my guess starts to be, yeah, that sure sounds to me like it ends the world, but maybe not immediately.

Speaker 2 Maybe you need like another two years of technology burn-in first.

Speaker 2 And then if you talk about a breakthrough on the order of deep learning itself, that that seems to me like that just sort of like ends the world in a snap.

Speaker 1 A quick aside, using the internet without a VPN today is like leaving your front door wide open and hoping that no one walks in.

Speaker 1 Websites, apps and data brokers are constantly collecting your personal information, what you search, what you watch, what you buy, where you are.

Speaker 2 It all gets tracked.

Speaker 1 And Surfshark protects you from that. It encrypts your internet connections so your activity stays private, even on sketchy public Wi-Fi at airports, cafes or hotels.

Speaker 1 And it lets you change your virtual location with a single click. The clean web feature also blocks ads, trackers and malware before they even load so you stay safer and your browsing is smoother.

Speaker 1 You can run Surfshark on every device that you own, unlimited installs on one account.

Speaker 1 And right now, you can get four extra months of Surfshark for free by going to the link in the description below or heading to surfshark.com slash modern wisdom and using the code modernwisdom.

Speaker 1 A checkout. That's surfshark.com slash modern wisdom and modern wisdom, a checkout.

Speaker 1 Okay, so LLMs could be a really big deal. And there's also a ton of other stuff that could...
that we can't see that would be dangerous as well.

Speaker 2 I don't know if the LLMs could go there. Some people are saying that it seems to them like the LLMs are as smart as they get.

Speaker 2 And other people are like, well, did you try GPT-5 Pro for $200 a month or whatever it is at that cost? And other people are going like, yes, I did.

Speaker 2 And like the $200 version of Claude is no better than the $200 version of this.

Speaker 2 And the thing I would say about

Speaker 2 this is that If you have some perspective, if you have been watching this for longer than three years, if you have been watching this from before ChatGPT,

Speaker 2 stuff saturates and then other stuff comes along and breaks through.

Speaker 2 It doesn't matter if LLMs take you to the end of the world because people are not because they're not going to stick to LLMs.

Speaker 2 Okay.

Speaker 1 What are the range of timelines for this sort of transformative AI that you think are likely?

Speaker 2 I mean, again, everybody wants questions, wants answers like these,

Speaker 2 just like they'd like to know next week's winning lottery numbers.

Speaker 2 But if you look over the history of science, I am hard-pressed to name a single case of successful prediction of timing of future technology.

Speaker 2 There are many cases of scientists correctly predicting what will be developed.

Speaker 2 You can look at the laws, you can look at the physical laws, you can look at the biology laws, you can say, and you can look look at that, like, hmm, yeah, this sure looks like it ought to be possible.

Speaker 2 And you can look at it and say, this sure looks like it ought to be possible. And I think I see the angle of attack there.

Speaker 2 Leo Sillard in 1933 was crossing a particular street intersection, whose name I forget, when he had the insight that we would now refer to as a

Speaker 2 chain reaction,

Speaker 2 nuclear chain reaction,

Speaker 2 a cascade of induced radioactivity. Even then, it was known that you could put some materials next to

Speaker 2 a source of radioactivity and

Speaker 2 induce secondary radioactivity.

Speaker 2 And so Leo Sillard was like, hmm, we've got these naturally radioactive materials. What if we find something that's naturally radioactive and furthermore has the property that

Speaker 2 you can induce radioactivity in it.

Speaker 2 Uranium-235 was what was eventually settled on, but back then they didn't know that.

Speaker 2 And Leo Sillard saw way ahead in that moment. He saw through to nuclear weapons.

Speaker 2 He saw that this was not something he should publish in a journal for immediate fame and fortune. He realized that Hitler specifically was likely to be a problem.

Speaker 2 He did not say, this is going to take $2 billion to turn into a weapon by 1945.

Speaker 2 There are, off the top of my head, there are zero instances of a scientist ever making a call like that.

Speaker 2 It is the difference between predicting that an ice cube dropped into a glass of water is going to melt and predicting how long it takes to melt and where all the individual, where like the individual molecules end up.

Speaker 2 If you point out that on a quantum level, the molecules are indistinguishable. I claim that there's some deuterium in there, so you can't predict what you're seeing.

Speaker 1 I get it. Look, look,

Speaker 1 I imagine that that's probably got to be number one on the list of things people who work in AI safety are sick of being asked.

Speaker 2 A lot of them will run off an answer. A lot of them are not wise enough to realize that they can't answer it.

Speaker 2 Okay.

Speaker 1 I'm going to guess that your confidence interval that it happens before the end of the century is probably pretty high.

Speaker 2 Yeah.

Speaker 2 I mean, unless we deliberately shut it down. And even then, getting all the way out to the end of the century sounds hard.

Speaker 2 If you had an international treaty planning this stuff, I would say to go really hard on human intelligence augmentation, because eventually the international treaty will break down.

Speaker 2 All you can do with it is buy time to have smarter people tackling this problem and tackling humanity's problems in general.

Speaker 2 But that's a bit of a topic change there.

Speaker 2 The people at the AI companies themselves

Speaker 2 are sometimes naming two to three year timelines.

Speaker 2 And there is a lesson of history which says that just because you can't predict when something will will happen does not mean that it is far away.

Speaker 2 Two years before Enrico Fermi personally oversaw the construction of the first self-sustaining nuclear reaction, the first nuclear pile that went critical, he said

Speaker 2 that that was 50 years off if it could ever be done at all.

Speaker 2 Fermi, not being wise enough, to realize that he couldn't do timing.

Speaker 2 A couple of years before the Wright brothers flew, one of the Wright brothers said to the other, I forget if it was Orville or

Speaker 2 Wilbur,

Speaker 2 man will not fly for a thousand years. But they kept on trying anyway.
So it was two years off, but their intuitive sense was it's 1,000 years off.

Speaker 2 And of course, AI itself, very famously, there were some people in 1955 who thought they could make progress on AI,

Speaker 2 learning to talk, be scientifically creative, and self-improve over the course of a summer with 10 researchers.

Speaker 2 This was not a completely unreasonable thing to think because nobody had ever tried it, and maybe AI would turn out to be that easy, but it wasn't actually that easy, not in 1955.

Speaker 2 So

Speaker 2 the

Speaker 2 it, you know, it could be two years away. It could be 15 years away.

Speaker 2 The AI companies themselves say two to three years, but it's questionable whether we should be taking their words at face value as meaning things as opposed to like hype.

Speaker 1 Yeah, the LLMs, if that architecture is not the one that is going to end up at a place that is super dangerous, then

Speaker 1 what do they know? If they have got all of their chips on this one particular architecture, they're all in on this.

Speaker 2 We don't know that.

Speaker 2 They don't. Oh, God.

Speaker 1 Every time I think I've managed to get some sort of like reprieve, they're like, oh no, what about the super secret Open AI project that's actually using some other approach?

Speaker 2 So

Speaker 2 the most recent, you know, reasonably large breakthrough in large language models was successfully applying reinforcement learning to chain of thought.

Speaker 1 And it was a very good thing. Can you explain what that means?

Speaker 2 So

Speaker 2 if you haven't learned anything about LLMs since

Speaker 2 they started getting heard about...

Speaker 2 If you haven't, so like you might have heard that LLMs just imitate humans.

Speaker 2 This is false.

Speaker 2 You can also have an LLM try to think about how to solve a problem.

Speaker 2 And then of the like 20 tries it takes at solving the problem, one of those tries works or works best. And then you say, think more like that try at thinking about the problem that succeeded.

Speaker 2 This is how LLMs go past imitating humans, or it's one of many ways that LLMs go past imitating humans.

Speaker 2 So this is a very, so this is a relatively very obvious thing to do with LLMs. Like Paul Christio and myself were

Speaker 2 talking about that

Speaker 2 10 years ago.

Speaker 2 But

Speaker 2 before LLMs actually existed, because that's how obvious it is.

Speaker 2 But getting it to work

Speaker 2 was

Speaker 2 like last year or two, maybe.

Speaker 2 And

Speaker 2 Open AI

Speaker 2 had this like thing called strawberry. And it was, you know, their like super secret special LLM sauce that they weren't going to tell anyone.

Speaker 2 It was actually just like reinforcement learning on chain of thought.

Speaker 2 But the point is that

Speaker 2 this is the level of innovation that AI labs have in the past proven to have and keep secret and that where we later found out what it was.

Speaker 2 And well, you know, they did get a fair amount of mileage out of that.

Speaker 2 Out of having AIs try different ways of thinking and reinforcing the one that worked to solve objectively verifiable problems like math or programming and so on.

Speaker 2 So this is

Speaker 2 the AI companies could potentially have a replacement for LLMs that they've discovered and are keeping secret from us.

Speaker 2 More likely is that they would have something that was on the order of reinforcement learning on chain of thought, which is when AI started to get good at coding.

Speaker 2 Or they might have nothing on that order up their sleeves at the moment.

Speaker 2 And that's why people are currently claiming that

Speaker 2 the latest wave of LLMs do not seem fundamentally smarter than the LLMs from three months ago or six months ago, which is what today's young whipper snappers think is an AI winter.

Speaker 2 Let's see your field stagnate for 10 years and eventually break through before you have to talk to me about winter again, kids.

Speaker 2 Okay, brilliant. Brilliant.

Speaker 1 I have no idea what I even want to ask you.

Speaker 1 I want to know why experts aren't worried. And I also want to know what you make about AI companies.
Let's talk about the expert. Why?

Speaker 1 Obviously, some people's wages

Speaker 1 are dependent on this train staying on the tracks. That means

Speaker 1 it's very difficult to convince somebody. What's that quote? It's very difficult to convince somebody of something that their wage depends on them not being convinced of.

Speaker 1 What about the other

Speaker 1 thinkers,

Speaker 1 researchers in this space? What is it that

Speaker 1 they are most commonly missing? that you think?

Speaker 1 Where are they making their fundamental thinking errors when it comes to we will be fine with just continuing on

Speaker 1 AI growth?

Speaker 2 So first of all, Jeffrey Hinton,

Speaker 2 the guy who won the Nobel Prize

Speaker 2 in Physics for being

Speaker 2 among the people most directly pinpointable as having kicked off the entire revolution in getting backprop to work on multilayer neural networks, or as it's now currently known, deep learning, like the point where AI started working at all.

Speaker 2 Jeffrey Hinton, I think, is on record as

Speaker 2 recently saying he quit his job at Google and then could speak freely.

Speaker 2 Saying something like, intuitively, it seems to him like his 50%

Speaker 2 catastrophe probability, but based on other people seeming less concerned, he ingests it down to 25%.

Speaker 2 I could be misquoting here. I'm trying to do this from memory.

Speaker 2 So, are you asking? So, many people would consider this

Speaker 2 to not be a lack of concern.

Speaker 2 Like the guy say, like somebody being like, well, it looks to me like a coin flip, whether or not you destroy the world.

Speaker 2 This is not what you want to hear from your Nobel laureate scientist who helped invent the field and left Google to be able to speak freely about it.

Speaker 2 So he no longer has a financial stake in making it bigger or smaller one way or the other.

Speaker 2 Many people would call this already a high degree of scientific alarm.

Speaker 2 Yashua Bengio was one of the co-founders of deep learning.

Speaker 2 He co-won the computer science award with Jeffrey Hentit, the Turing Prize for inventing deep learning. Yashua Bengio is also, I think, on the concern list.

Speaker 2 I don't off the top of my head have a direct quote from him about probabilities.

Speaker 2 It is true that I am more concerned than they are.

Speaker 2 I would, and I realize that this may sound somewhat hubristic, attribute this to them being relative newcomers to my field who may not have gotten acquainted with the full list of reasons why it is hard to align AI.

Speaker 2 That said, coin flip odds of destroying the world is still not what you want to be hearing

Speaker 2 from your relatively more senior scientists who are relatively newer to the field.

Speaker 2 Well, relatively newer to my field. They are vastly my seniors in artificial intelligence itself, of course.
I am like speaking tongue-in-cheek whenever I accuse people of being young whippersnappers.

Speaker 2 Like Jeffrey Hinton could say that with a straight face. I am just like, you know, a bit of light self-mockery there about how I'm not Jeffrey Hinton.

Speaker 2 But that said, you know,

Speaker 2 if you are relatively newer to this, you might think like, well, you know, maybe we've just got to use reinforcement learning to make the AIs love us the way a child loves a parent or love us the way a parent loves a child and not

Speaker 2 quite have at your fingertips the top six reasons why the what why that is hard and principled obstacles to that and what will go wrong there

Speaker 2 so that is what prevents the the the like the the the famous inventors of the field who only started speaking out about their concerns relatively recently after leaving their companies to you know and are now financially dependent of stakes on their opinion that's what's prevent that's what makes them be like 50-50 the world gets destroyed instead of my own thing where i'm like yeah it's predictable that the world gets destroyed if you keep doing this

Speaker 2 but if you ask like, what's responsible for Sam Altman at Open AI

Speaker 2 not, you know,

Speaker 2 possibly having less than 50% odds, who knows what that guy's really thinking, well, you can like trace out his long trail over time of

Speaker 2 him initially saying like, AI will end the world, but in the meanwhile,

Speaker 2 there will be great companies.

Speaker 2 to him sort of like saying less and less alarmist sounding things in front of Congress, like where Congress asks him, like, well, you talk about the world ending.

Speaker 2 By that, do you mean like mass unemployment? And Sam Altman hesitates for two seconds and replies, yes, was the lovely like congressional hearing thing that happened, I think, about a year back now.

Speaker 2 So what's going on with the AI companies?

Speaker 2 I'm not telepaths. I can't read their minds.
I would point out that it is immensely well precedented in scientific history, in the history of science and engineering, for

Speaker 2 companies that are making short-term profits to do

Speaker 2 really sad amounts of damage, vastly disproportionate to the profit that they are making,

Speaker 2 and to be in apparently sincere denial about the negative effects of what they are doing. Two cases that come to mind are leaded gasoline and cigarettes.

Speaker 2 I don't know if you would be familiar off the top of your head with the case of leaded gasoline. Probably even the kids today have heard about cigarettes.

Speaker 2 The cigarette companies did way more damage to human life

Speaker 2 in cancer and other health effects than they made in profits. Like they did make a few billion dollars in profits selling cigarettes, but nothing remotely compared to the cost of human life.

Speaker 2 It's not that they were, you know, like this was an immensely negative sum game. They were doing enormously more damage than the profits that they were making.

Speaker 2 And any particular advertising professional who got up in the morning and figured out how to market cigarettes to teenagers, any of the scientists that they paid to write stories about how you couldn't really tell whether or not cigarettes were causing lung cancer, would have made a tiny, tiny fraction of the total profit of the cigarette companies.

Speaker 2 Their CEO would not have made that larger fraction of the total profit of the cigarette company.

Speaker 2 So they went off and participated in this thing that, you know, caused lung cancer to, I don't know how many millions of people.

Speaker 2 And for what? For this very small profit.

Speaker 2 How could a human being bring themselves to do that? Through a very simple alchemy.

Speaker 2 First, you convince yourself that what you're doing is not causing the harm, which is just a very easy thing for human beings to do all the time, all throughout the entire recorded history of humanity.

Speaker 2 And then once you've convinced yourself that you're not doing that much harm, well, what's the harm in taking money to not do any harm?

Speaker 2 Leaded gasoline caused brain damage to tens, maybe hundreds of millions of developing brains in the United States and elsewhere. It caused brain damage to children.

Speaker 2 For what?

Speaker 2 The gas companies making leaded gasoline could have, you know, made unleaded gasoline. It's not that they would have gone out of business if they'd

Speaker 2 somehow gotten together and decided to stop making leaded gasoline.

Speaker 2 If they hadn't opposed the regulations that were trying to bend leaded gasoline before it turned into a big deal back in the 1930s,

Speaker 2 there was an attempt to have regulations against leaded gasoline. Lead was known to be poisonous in large quantities.

Speaker 2 Why let people spray it all over the place, even in smaller quantities?

Speaker 2 But the gas companies got together. They managed to prevent that legislation from

Speaker 2 passing.

Speaker 2 They poisoned an entire generation.

Speaker 2 And for what?

Speaker 2 For

Speaker 2 gas that burned about 10% more efficiently, I think was what leaded gasoline basically got you.

Speaker 2 For it being more convenient to add lead to the gas instead of adding ethanol to make it burn more smoothly inside of car engines.

Speaker 2 Trivial.

Speaker 2 Trivial, trivial compared to the, this is not a conspiracy theory. This is standard medical history I'm talking about here.

Speaker 2 Like I've seen estimates of five points off the tested IQs.

Speaker 2 And you can look at the chart of which states banned leaded gasoline when and watch the drops in the crime rate

Speaker 2 because it makes you, you know, it disposes you to be more violent, not just stupid, that tiny little bit that it that that hit child after child after after child.

Speaker 2 Why?

Speaker 2 Why would anyone cause that amount of damage

Speaker 2 because

Speaker 2 you got your CEO salary of a company that didn't have, then didn't need to go to the inconvenience of adding ethanol to gasoline instead?

Speaker 2 Because first you convince yourself it's safe. First you convince yourself you're doing no harm.
which is just an easy thing for human brains to convince themselves of. And then why not oppose

Speaker 2 the legislation against leaded gasoline? It's not doing any harm, right?

Speaker 2 Ronald Fisher,

Speaker 2 one of the inventors of modern scientific statistics,

Speaker 2 testified against it being knowable that cigarettes cause lung cancer because you see

Speaker 2 no proper controlled experiment had been done on cigarettes causing lung cancer. And so how could you possibly, possibly know from your observational studies

Speaker 2 showing 20 times the chance of cancer if you were a smoker. How could you possibly know from mere

Speaker 2 correlational studies? And Fisher himself was a heavy smoker.

Speaker 2 He actually drank his own Kool-Aid.

Speaker 2 The inventor of leaded gasoline, I think, had to go away to a sanitarium at one point because of how much he managed to poison himself with lead. He drank his own Kool-Aid.

Speaker 2 They really managed to convince themselves that they were doing no harm. And so they could do arbitrarily vast amounts of harm in exchange for these tiny, comparatively tiny, tiny profits.

Speaker 2 And

Speaker 2 to say this is not a substitute for actually tracking the object-level arguments about whether or not AI will kill you and for what reason.

Speaker 2 You cannot figure out what will happen as a matter of computer science if you build a superintelligence and switch it on by pointing out at who has what tainted motives,

Speaker 2 you know, who has what incentives to say what.

Speaker 2 But having tried in my book,

Speaker 2 in my Innate Sorry's book, to make the case for why, on an object level, this is what happens if you build a superintelligence and switch it on, to ask why the people

Speaker 2 being paid literally hundreds of millions of dollars by meta

Speaker 2 to be AI researchers, why people like Sam Altman, who, you know, I mean, he didn't quite get paid billions of dollars. He was supposed to be CEO of a non-profit.
He actually stole billions of dollars.

Speaker 2 But, you know, why the guy stealing billions of dollars in equity

Speaker 2 from the public that was supposed to own it?

Speaker 2 Like, like, how does he manage to convince himself that what he's doing is okay? Well, maybe he's not even convinced.

Speaker 2 You know, we do have him on the record as saying a few years earlier, like AI will end the world, but in the meantime, there'll be great companies.

Speaker 2 You know, maybe, maybe he's just like, yeah, sure, you know, like the world's going to end, but I get to be important. I get to be there.
You know, sure, who but I could be trusted with this power?

Speaker 1 You think that that's the position that a lot of the guys at the heads of these AI companies believe?

Speaker 2 I'm not a telepath. I can't tell you what these people are actually thinking.
You got to distinguish between stuff you can possibly know and stuff you can't.

Speaker 2 But their overt language

Speaker 2 has often been like, well, building superintelligence is inevitable. Who could possibly stop that? An international treaty could possibly stop that.

Speaker 2 A coalition of major nuclear powers could stop that. But leading that aside, they may have convinced themselves that's not going to happen.

Speaker 2 Who could possibly stop anyone from building superintelligence? So I need to build it.

Speaker 2 Only I can be trusted to build it

Speaker 2 is what their overt rhetoric has sort of been.

Speaker 2 Okay.

Speaker 2 But the main thing I'm trying to point out is that having presented the object-level case that superintelligence will kill everyone, to ask the question of how could these companies possibly believe that this thing bringing them immense short-term profits and letting them be the most important guy in the room is, you know, not going to end the world is something enormously well-precedented in the history of science.

Speaker 2 If I'm saying that's to the extent you might think that's what happened, a very ordinary thing happened, not an extraordinary thing. A thing happened that has happened a dozen times before.

Speaker 2 If they managed to convince themselves that they were doing no harm. Okay.

Speaker 2 Or, you know, only an acceptable amount of harm, only running a 25 chance of destroying the world whatever it is they think is acceptable

Speaker 2 uh

Speaker 1 i'm trying to work out

Speaker 1 what the solution is do you have any proposed solutions that makes this seem slightly less apocalyptic

Speaker 2 my best i have to offer is is the same solution that humanity used on global thermonuclear war don't do it like don't instead of having the nuclear war the global thermonuclear war and trying to survive it, which for the nuclear war might have worked, don't have the nuclear war.

Speaker 2 We managed to do that.

Speaker 2 It's the best sign of hope I can offer you. It is slightly harder for AI in some ways, if not others.
But,

Speaker 2 you know, people going into the 1950s, 1960s, they thought they were screwed.

Speaker 2 And that wasn't them indulging in some nice doom-scrolling pessimism, luxuriating in the pleasant feeling of being doomed. This was people who did not want to be doomed.

Speaker 2 But they looked at the course of human history over the last century. They looked at World War I.

Speaker 2 They looked at how in the aftermath of World War I, everyone had said, let's not do that again. And then there'd been World War II.

Speaker 2 They had some reason to be worried about nuclear war. They had some reason to expect that no country was going to turn down the prospect of making nuclear weapons.

Speaker 2 They had some reason to believe that, you know, once a bunch of great powers had a bunch of nuclear weapons, why, of course, they would go to war anyway and use those nuclear weapons.

Speaker 2 It was apparently to them what had happened with World War II, all these people saying, we must not have another world war and then the world war happening anyway.

Speaker 2 Why didn't we have a nuclear war?

Speaker 2 Well, on my account of it, it is because for the first time in all human history, all the great powers, all the leaders of the great powers, understood that they personally were going to have a bad day if they started a major war.

Speaker 2 And people had pretended before to claim, proclaim that, you know, war is a very terrible thing that should never be done. It wasn't quite the same level of personal consequence.

Speaker 2 You know, maybe as, maybe as a general secretary of the Soviet Union, you would think that if you started a nuclear war, you would personally survive.

Speaker 2 You'd end up in a bunker somewhere, but you wouldn't be going to your favorite restaurants in Moscow ever again.

Speaker 2 And that was not the situation that obtained before the start of World War I, the start of World War II.

Speaker 2 People might make a bunch of, you know, like it only takes one side to think that they might have a bit of an advantage in the sport, in war, the sport of kings, to, you know, to kick off that fun adventure of trying to conquer another country, which, you know, wasn't as much fun for Adolf Hitler as he expected.

Speaker 2 But you could see how Adolf Hitler might have thought that he was going to have a nice day as a result of invading Poland.

Speaker 2 And that's what changed, that the General Secretary of the Soviet Union and the President of the United States actually personally expected to have both sides expected to personally have bad days if they start a nuclear war.

Speaker 2 They would not have any better of a bad, any better of a good day if anyone anywhere on Earth built a superintelligence.

Speaker 1 Yeah, it's this sort of lack of the tragedy. It's kind of like a tragedy of the comments.
It's just tragedy that everybody's fucked, right?

Speaker 1 It's like everything, everything gets blown up no matter who it is that builds it.

Speaker 2 The tragedy of the commons is that the commons get overgrazed because the individual farmers benefit from setting their cows loose on it.

Speaker 2 And the thing with nuclear war is that you might get a bit of a benefit by dropping a tactical nuclear weapon on, you know, like...

Speaker 2 You know, like the United States could get an immediate benefit by dropping tactical nuclear weapons on the Russian troops in Ukraine.

Speaker 2 And Russia could get an immediate benefit by dropping tactical nuclear weapons on Ukraine, but neither of them is going to risk the global thermonuclear war that might follow

Speaker 2 happening with a greater probability.

Speaker 2 So it's not,

Speaker 2 so it's not a tragedy, like it's not a classic tragedy of the commons.

Speaker 2 The thing that stopped nuclear war is that although you could get a short-term advantage from dropping a tactical nuke

Speaker 2 or even like dropping a strategic nuke on one city, the leaders understood how this was a, you know, like increasing the probability of a global thermonuclear war.

Speaker 2 And they managed to hold off from doing that for that reason. They understood the concept of how it escalated things.

Speaker 2 They saw the connection to not getting to go to their favorite restaurants again, even if they were surviving in a bunker somewhere.

Speaker 2 And with artificial intelligence, what we've got is a ladder where Every time you climb another step on the ladder, you get five times as much money.

Speaker 2 But one of those steps of the ladder destroys the world and nobody knew.

Speaker 2 And maybe if this true fact can become something that is known and believed by the leaders of

Speaker 2 a handful of major nuclear powers, they can all be like, all right, we're not climbing any more rungs of this ladder.

Speaker 2 It is not in my interest that you start to climb this ladder.

Speaker 2 And it's not even my own interest to break apart the treaty by climbing another step of this ladder, because then we're all just going to keep climbing, and then we're all going to die.

Speaker 2 That is the best ray of hope I can offer you.

Speaker 2 That we managed to not do the stupid thing, the same as we managed to not have a nuclear war, despite many people being concerned for excellent reasons that it was going to be an impossible slope not to fall down.

Speaker 1 Okay, so what do we actually do?

Speaker 2 Well,

Speaker 2 you know, voters do not necessarily have all that much power under the modern political process.

Speaker 2 But I think, but like the next step for the United States might be something like the president saying, you know, like, we're of course not going to give up AI unilaterally, which wouldn't even solve anything in its own way.

Speaker 2 But we stand ready to, you know, join with an international

Speaker 2 international treaty, international alliance whose purpose is to prevent further escalation of AI intelligence, further escalation of the AI ladder.

Speaker 2 We're not going to do it unilaterally, but we're ready to get together and do it everywhere.

Speaker 2 And China has already sort of like hasn't quite said that, but they've sort of indicated openness to international arrangements meant to prevent human loss of control from AI.

Speaker 2 You'd want Britain to say the same thing.

Speaker 2 So, and then if a bunch of leaders of major powers have said, like, yeah, we would join in an arrangement to prevent this from getting out of control and everybody on earth, you know, ending up dead.

Speaker 2 Then you can, from there, you can go on to the actual treaty. What can voters do? Well, you can

Speaker 2 write your elected officials is among the things you can try to do there.

Speaker 2 Um,

Speaker 2 there's a uh

Speaker 2 there if you go to if anyonebuildsit.com

Speaker 1 can't believe that you got that URL.

Speaker 2 Brilliant. Okay.
Yeah. Yeah.
Anyonebuildsit.com and you click on where it says act,

Speaker 2 you'll see our guide to calling your representatives.

Speaker 2 And

Speaker 2 if you click on March,

Speaker 2 you'll see a place where you can sign up to march on Washington, D.C. if 100,000 other people also pledge to march on it.

Speaker 2 And this does not this for this to just happen in the United States does not solve the problem because this is not a regional problem where you ban superintelligence inside your own country and then your own country is safe.

Speaker 2 But this sort of thing can

Speaker 2 exert some amount of influence on politicians and more importantly, can make it clear to them that they're allowed to discuss it, that they're allowed to want to not die themselves.

Speaker 2 There are

Speaker 2 multiple congresspeople who I'm not going to name, but whom we have talked to, who would, you know, prefer that America not die along with the rest of the world, but it doesn't quite seem like the sort of thing you're allowed to speak out in public about yet.

Speaker 2 Voters can make it clear to their politicians that the politicians are allowed to speak out.

Speaker 2 There's already like 70%,

Speaker 2 like if you actually survey American voters, 70% of them say they do not want superintelligence. But

Speaker 2 that's not enough for the politicians to feel licensed to act. But

Speaker 2 if you call them and if you're

Speaker 2 marching to Washington, that's what you can do as an individual voter.

Speaker 1 Well, I applaud you for trying to get get some grassroots stuff going. Congratulations.

Speaker 1 You've been frank throughout this conversation. I think it's fair for me to be frank here.
It does feel a little bit like you're outgunned.

Speaker 1 Legislation tends to move more slowly than technology does by many, many years, sometimes decades.

Speaker 1 It just feels bleak.

Speaker 1 It feels, if what you say is true,

Speaker 1 it really is kind of fluke that gets us to a stage where this goes well, because the likelihood of some moratorium being placed where all AI development is halted and all efforts are placed on this.

Speaker 1 You only need one bad actor to do it, which because again, it's if anybody builds it.

Speaker 2 You don't want the international treaty to fall over if North Korea steals a bunch of GPUs. You do want the treaty to say, if North Korea steals a bunch of GPUs and builds a

Speaker 2 unlicensed data center, then we will clearly communicate diplomatically what is about to happen. And then, if North Korea still proceeds, we will drop a bunker buster on their data center.

Speaker 1 That assumes that you know that you are somehow able to detect and that no one can do it surreptitiously.

Speaker 2 It is hard to surreptitious a data center. They consume a lot of electricity.

Speaker 1 Okay, so we can see most of the ones in Russia and China and North Korea.

Speaker 2 Like, I'm not sure who is looking for them at the moment. And if you can, you know, and to what extent these things show up on satellites and to what extent these things show up

Speaker 2 on, you know, intelligence reports. But there has previously been an issue of detecting covert nuclear refineries

Speaker 2 in terms of

Speaker 2 nuclear non-proliferation. And this was not an unsolvable problem.
And data centers are, if anything, even higher profile than nuclear refineries.

Speaker 1 Right. So we are going to threaten some people

Speaker 1 with.

Speaker 2 I mean, I wouldn't use the word threaten.

Speaker 2 I would say that if North Korea is building an unsupervised data center, then you should actually be terrified for your lives and lives of your children.

Speaker 2 And you tell North Korea this plainly and truthfully. And then if they don't drop, you know, if they don't shut down their data center, you drop a bunker buster on it.

Speaker 2 Then you do this even though North Korea has some nuclear weapons of its own.

Speaker 1 Okay, so pressure from people on their elected representatives through mail,

Speaker 1 marches, more awareness to get the government officials to come up with an international treaty to get countries to agree that

Speaker 1 what specifically

Speaker 2 we're not making AIs any smarter than they are already.

Speaker 2 We are putting the chips that can be used to build the more powerful AIs

Speaker 2 into

Speaker 2 locations

Speaker 2 where their uses are supervised.

Speaker 2 I would say, ideally, you are putting the chips that run the AIs into locations where they can be supervised. As a minor side effect, maybe you can stop the AIs from driving people insane.

Speaker 2 It seems like the sort of thing you could better do if this was all happening under supervision by international treaty.

Speaker 2 It's not vital to humanity's survival that AIs be prevented from driving people insane, but it serves as a kind of test case of can you

Speaker 2 stop the damage?

Speaker 2 Like, is humanity in control here? Can we stop AIs from predating upon some of our human people?

Speaker 2 But that's not the main thing here.

Speaker 2 It's a thing that some people will find attractive, but it's not the main thing.

Speaker 2 You're trying to just get the whole AI thing under control, and then you're trying to stop the further escalation of AI capabilities up the ladder.

Speaker 1 It is scary.

Speaker 1 It is

Speaker 1 one of these things. And I imagine that it feels, it must feel a little bit like this to you, that everybody is sort of...

Speaker 1 dancing their way through a daisy field of, oh, I've got this personal coach in my pocket, and it's so cool. And I get to talk to it about all of my psychological problems.

Speaker 1 God, I can bitch to it about my husband. And it just listens.

Speaker 1 And at the end of this

Speaker 1 daisy field that everyone's having a load of fun in is just like a huge cliff

Speaker 1 that descends into eternity. And there's like a battle rog at the bottom or something.

Speaker 1 Is that what it feels like?

Speaker 1 Yeah,

Speaker 2 pretty much.

Speaker 2 But the future is hard to predict. It is genuinely hard to predict.
I can tell you that if you build a super intelligence using anything remotely like current methods, everyone will die.

Speaker 2 That's a pretty firm prediction.

Speaker 2 The part where people maintain the daisy field attitude that they had a few years earlier toward AI, that has already shifted to some degree just because of the ChatGPT moment.

Speaker 2 And nobody predicted that in advance.

Speaker 2 Nobody knew that

Speaker 2 nobody at OpenAI, as far as I can tell, had any idea that when they released ChatGPT, they were going to be causing a massive shift in public opinion about AI as people realized the AIs were actually talking to them now and sounding kind of intelligent about it.

Speaker 2 So

Speaker 2 maybe it also,

Speaker 2 maybe

Speaker 2 I don't want to wait for anything else to happen. Maybe ChatGPT was the miracle we got.
I wasn't expecting that much of a miracle. I did not call it in advance.

Speaker 2 But maybe we get another miracle. I don't want to sit around waiting for it because I can't tell you the miracle will like occur on such and such a day.
But

Speaker 2 maybe the AI has managed to do something more destructive than driving a few people insane, breaking up a few marriages and

Speaker 2 causing whatever further decline in birth rates is going to be caused here.

Speaker 2 Maybe they do worse than that and that shifts opinion. Maybe they just get more powerful and smarter and are clearly no longer toys and that shifts opinion even without a giant catastrophe.

Speaker 2 It's not clear to me, you know, as much as people love to bitch about their elected leaders, it is not clear to me that we are looking at

Speaker 2 permanent obliviousness to the aliens getting smarter and smarter.

Speaker 2 Like people are currently saying completely wacky and oblivious things because they think that's what's politically mandatory to say in the current political environment and that you have to talk about jobs rather than the other extinction of humanity.

Speaker 2 But it's not clear. The future is very hard to predict in general.
It's not clear to me that the current state of obliviousness is something supreme, unmovable, and impossible for any event

Speaker 2 to change, or that it won't just disintegrate on its own as more people talk about it.

Speaker 2 There's a level in which you kind of have to be

Speaker 2 pretty dumb to look at this smarter and smarter alien showing up on your planet and not have the thought cross your mind that maybe this won't end well.

Speaker 2 Can

Speaker 2 even elected politicians be that dumb? Yes, absolutely. It is not known to me to be prohibited that this can be the case.
Do they have to do the stupid thing? It's not clear to me that it's mandatory.

Speaker 2 We did manage to have, we did manage to not have a nuclear war, and people did not think they were going to get that much luck.

Speaker 1 Oh, yeah, Zayukowski, ladies and gentlemen.

Speaker 2 I'll be fucked.

Speaker 1 I was prepared coming in, but I'm not sure that the rest of the audience will be. So, dude,

Speaker 1 the best compliment I can pay you is: I hope you're wrong, but I fear you're not.

Speaker 2 Yeah,

Speaker 2 being wrong. It'd be great to be wrong.
I'd love to be wrong. That'd be wonderful.

Speaker 2 That would be wonderful.

Speaker 2 Let me assure you, everyone,

Speaker 2 by way of

Speaker 2 destroying any shred of optimism you might previously have had, I completely would have other career options lined up and other ways of supporting myself if I was completely wrong.

Speaker 2 Like, like not just me, but some like sensible people who donated me a bit of appreciated currency wanted to make sure that I could, if I was, you know, you know, if I changed my mind about this sort of thing, just retreat from my entire career path and not end up in financial trouble.

Speaker 2 And yet here I am. You know, so I'd love to be wrong.

Speaker 2 We have tried to arrange it to be the case that I could at any moment say, yep, I was completely wrong about that, and everybody could breathe a sigh of relief, and it wouldn't be like the end of my ability to support myself, and I would have other things to do.

Speaker 2 We have made sure to leave a line of retreat there. Unfortunately, as far as I currently know, I continue to not think that

Speaker 2 it is time to declare myself to have been wrong about this.

Speaker 1 Heck yeah. All right, Eliado.
Well, if the internet is still alive in a little bit of time in the future, we can check back in and see just how right you are.

Speaker 2 Well, every year that we're still alive

Speaker 2 is another chance for, you know, something else to happen.

Speaker 1 What a wonderful way to finish. Dude, I appreciate you.
Thank you so much for your work.

Speaker 2 Thank you for

Speaker 2 having me over to

Speaker 2 deliver the bad news.

Speaker 3 This episode is brought to you by LifeLock. It's Cybersecurity Awareness Month, and Lifelock has tips to protect your identity.

Speaker 4 Use strong passwords, set up multi-factor authentication, report phishing, and update the software on your devices.

Speaker 2 And for comprehensive identity protection, let Lifelock alert you to suspicious uses of your personal information.

Speaker 3 LifeLock also fixes identity theft, guaranteed or your money back. Stay smart, safe, and protected with a 30-day free trial at lifelock.com/slash podcast.

Speaker 1 Terms apply.