Your Brain on ChatGPT with Nataliya Kosmyna
Press play and read along
Transcript
It's time for Cyber Monday, Dell Technologies biggest sale of the year. Enjoy the lowest prices of the year on select PCs like the Dell 16 Plus featuring Intel Core Ultra processors.
And with built-in advanced features, it's the PC that helps you do more faster. They also have huge deals on accessories that pair perfectly with your Dell PC.
Plus, earn Dell rewards and enjoy many other benefits like free shipping, price match guarantee and expert support. Shop now at dell.com slash deals.
You might have some travel coming up. You know, it is the holiday season, and maybe you're visiting family and heading abroad or even a dream trip for the new year.
Imagine feeling very confident about that new place that you're going because you can speak to them in their own language.
Or imagine being at the holiday table and not letting anyone know that you study the language with Rosetta Stone so you actually know what they're saying and you know that they're talking about you.
Because that's what I use Rosetta Stone for.
I love improving my pronunciation with their true accent feature, which allows you to not sound like you're from Alabama, but you're trying to speak Espanol. No offense to anybody from Alabama.
I just randomly picked you. Don't wait.
Unlock your language learning potential now.
Star Talk Radio listeners can grab Rosetta Stone's lifetime membership for 50% off. That's right.
For the rest of your life,
half off. That's unlimited access to 25 language courses for life.
Visit rosettastone.com slash star talk to get started and claim your 50% off today.
Give a gift that keeps giving to you.
Go to rosettastone.com/slash star talk
and start learning today.
Chuck, if the forces of AI are not big enough in society and our culture, we now got to think about what AI's effect is on our brain. I'm going to say that there is no help from my brain.
So that does not make a difference. I know, but Neil, if you lean into these large language models and it takes away some of our core skills, surely that can't be an upside to that.
Once again, Gary, not going to affect me at all.
Coming up, Star Talk Special Edition, your brain on AI.
Welcome to Star Talk,
your place in the universe where science and pop culture collide.
Star Talk begins right now.
This is Star Talk Special Edition. Neil deGrasse Tyson, your personal astrophysicist.
And when it's special edition, you know that means we have Gary O'Reilly in the house. Gary.
Hi, Neil. All right.
We got another one of these. We're going to
connect the viewer, the listener to the human condition. Yes.
Oh, my gosh. But let me get my other co-host introduced here.
That would be Chuck Nice. Chuck, how you doing, man? Hey, man.
Yeah.
That when you know it's Chuck, it means it's not special at all.
Oh.
But we've got you because you have a level of science literacy that, oh my gosh, you find humor where the rest of us would have walked right by it. And, you know, that's part of our recipe here.
That's very cool. Yeah, I appreciate that.
Yeah. So, Gary, the title today, is AI Good for Us?
Okay, well, here's the answer. No, let's go.
No, okay. That's the end of the show.
Let's all go home, people. This was quicker than I did.
This was very quick. I mean, yeah.
You know.
So, Gary, what have you set up for the day? Well, Lane Unsworth, our producer over in the LA office and myself, we sort of noodled.
And this is a question that's been bouncing around a lot of people's thought processes for a while.
So all over the world, people are using LLM's large language models for their work, their homework, and plenty more.
Besides discussions of academic dishonesty and the quality of work has anybody actually taken the time to stop and think about what this is doing to our brains?
Today, we are going to look at some of the current, and I really do mean current time and space, this moment, research into the impact of using an AI tool can have on your cognitive load and your neural and behavioral consequences that come with it.
And the question will be: does AI have the opportunity to make us smarter or not?
I like the way you phrased that, Gary. It was very diplomatic.
I know. Smarter or not.
Or not. And does it have the opportunity to do so?
Okay. Smarter or dumber.
That's what you mean.
I didn't say those words. Well, here on Star Talk, we lean academic when we find our experts.
And today is no exception to that. We have with us Natalia Cozmina.
dialing in from MIT. Natalia, welcome to Star Talk.
Thanks for having me. Excited to be here with you.
Excellent. You're a research scientist at the one and only MIT Media Lab.
Oh my gosh.
If I had like another life and a career, I would totally be on the doorsteps there wanting to get a job. And if I had another life and career, it wouldn't exist.
I'd shut it down immediately because let's be honest,
science is a hoax.
People, some people do want you to believe that. You know, it's like science has 99 problems and virality ain't one, right? Right.
There you go. And you're in the fluid interfaces group.
You are trained in non-invasive brain-computer interfaces,
BCIs.
I'm guessing that means you put electrodes on the skull instead of inside the skull, but we'll get to that in a minute.
And you're a BCI developer and designer whose solutions have found their way into low Earth orbit and on the moon. We want to get into that.
So let's begin by characterizing this segment as your brain on chat GPT. Let's just start off with that.
What a great topic, Neil. Is there any way I can help you with that?
So, so
you research
what happens when students use chat GPT for their
homework and for their
what what have you found in these studies? Yeah, so we ran a study that's exactly the title, right?
Your brain on chat GPT, accumulation of cognitive debt when using an AI assistant for essay writing tasks.
So we did a very specific task that we're going to be talking right now about, which is essay writing.
We invited 50 students from greater Boston area here to come in person to the lab and we effectively put those hats as you just mentioned on their heads to measure their brain activity when they're writing an essay.
And we divided them in three groups. We asked one group, as you might already guess where that's heading, to just use chat GPT.
That's why paper is called your brain on chat GPT.
It's not because we are really, really singling out chat GPT. It's just because we use chat GPT in the paper.
So it's purely scientific.
So we asked one group of students to use only chat GPT to write those essays. another group to use Google, the search engine, to write those essays, and the third group to use their brain only.
So no tools were allowed. And we give them topics which are what we consider high level, right? For example, what is happiness? Is there a perfect society? Should you think before you talk?
And we gave them a very limited time, like 20 minutes to write those essays. And we finally, of course, looked into the outputs of those essays, right?
So what they actually written, how they use chat gpt how they use google and of course we asked them a couple of questions like can they give a quote can they tell us why they wrote this essay and what they wrote about and then there was one more final fourth session in this study where we swapped the groups so students who were originally in chat gpt group we actually took away the access for this first session and vice versa was true so if you were for example
you're our participant you were not our participant but if you were ever to come to Cambridge and be our participant, and let's say if you were actually... I'm not putting anything on my head.
I'm just letting you know right away. Okay.
Come on. It's the future.
It's the future. Now, the problem is he'd have to take off his tinfoil hat when he got there.
Yep, yep, I see. I see that happening regardless.
So if you were, for example, in our participant in brain-only group, we actually, for this first session, would give you access to chat GPT.
And again, we measured exact same things, brain activity what actually was an output and ask couple questions and what we found are actually significant differences between those three groups so first of all if you talk about the brain right we measured what is called brain functional connectivity so let's in a layperson terms like i'm we can even hear having three of you talking to each other talking to myself so that's what we measured who is talking to who?
Am I talking to Neil or is Neil talking to you? So directionality. So who talks to who in the brain? And then how much talking is happening? Is it just, hi, hello, my name is Natalia?
Or actually a lot of talking. So a lot of flow of data is being exchanged.
So that's literally what we actually measured. And we found significant differences.
And then some of those are ultimately not surprising. You can think logically.
If you do not have any, let's say, you need to do this episode right now, right?
And I'm going to take away all your notes right now, all of the external help, and then I'm going to measure your brain activity. How do you think it's going to turn out?
You're going to have like really your brain on fire, so to say, because you need like, okay, what was your name again?
Where was the study?
What is happening, right? You need to really push through with your brain, like you have memory activation, you need to have some structure.
Like, and now you don't have notes for the structure of this episode, right? So you need like, what was the structure?
What we did is that what we are talking about, what is, you know, you really have nothing to fall onto.
So of course, you have this functional connectivity that is significantly higher for brain-only group compared to the two other groups. Then we take search engine group, Google.
And actually, there's just as a prior research, there's a ton of people about Google already.
We actually, as a humanity, right, we are excellent in creating different tools and then measuring the impact of those tools on our brain. So there's quite a few of papers we are citing in our paper.
For example, there is a paper, Spoiler Alert, called Your Brain on Google from 2008. Literally, that's the name of the paper.
So we've actually found something very similar to what they found.
There would be a lot of activations in the back of your head. This is called visual cortisol or occipital cortex.
It's basically a lot of visual information processing.
So right now, for example, someone who is listening to us and maybe they are doing some work in parallel, they would maybe have some different tabs open, right?
They would have like one is like a YouTube tab, another would have like some other things that they're doing.
So, you know, you're basically jumping between the tabs, looking at some information, maybe looking at the paper while listening to us. So, this is what we actually see.
And there's a plenty of papers already showing the same effect. But then, for the LLM group, for chat GPT group, we saw the least of these functional connectivity activations.
And that doesn't again mean that you became dumb or you
there's actually quite a few papers specifically having in the title laziness and we can talk about this with other results but from brain perspective from our results it doesn't show that what it actually shows that hey you have been really exposed to one very limited tool right you know there's not a lot of visual stuff happening brain doesn't really struggle when you actually use this tool.
So you have much less of this functional connectivity. So that's what we found.
But what is, I think, interesting and effective, we may be heading back to this point of laziness and some of these, maybe a bit more, I would say, nefarious results are, of course, other results that are relevant to the outputs, to the essays themselves.
So, first of all, what we found is that the essays were very homogeneous. So, the vocabulary that was used was very, very similar for the LLM group.
It was not the case for the search engine and for the brain-only group. I'm going to give you an example.
And, of course, in the paper, we have multiple examples. I'm going to give you only one.
Topic happiness. So we have LLM, so ChatGPT users, mentioning heavily the words career and career choice.
And surprise, surprise, these are students. I literally just mentioned this.
Of course, they're going to more likely talk about career and career choices. And again, who are we ultimately to judge what makes the person happy, right? No, of course.
But don't forget the two other groups, they are from the same category. They are students in the same geographic area, right?
However, for them, these words were completely different. For the Google, for the search engine, students actually heavily use vocabulary giving and giving us.
And then brain only group was using vocabulary related to happiness and true happiness. And this is just one of the examples.
And then finally, to highlight one more result, is responses from the participants themselves, from those students.
So we asked literally 60 seconds after they gave us their essays, can you give us a quote?
Any quote, any length of the quote of what you had just written can be short, long, anywhere in your essay, anything.
83%
of participants from LLM, from ChatGPT group, could have not quoted anything. That was not the case for brain and search engine groups.
Of course, in sessions two and three and four, they improved because surprise, surprise, they knew what the questions would be, but the trend remained the same. It was harder for them to quote.
But I think the most ultimately dangerous result, if I can use this term, though it's not really scientific, but something that I think a lot of inquiry actually is required to really look further into this, it's almost on philosophical, I guess, level, is ownership questions.
So we did ask them if, how percentage of ownership do they feel towards those essays? And 15% of ChatGPT users told us that they do not feel any ownership.
And of course, a lot of people, especially online, mentioned, well, they haven't written this essay. Of course, they didn't feel any ownership.
But I think that's where it actually gets really tricky.
Because if you do not feel that it's yours, but you just worked on it, does this mean that you do not care?
We didn't obviously push it that far in the paper, but I think this is something that definitely might require much further investigation.
Because if you don't care, you don't remember the output, you don't care about the output, then what ultimately is it for? Why were you even here, right?
Of course, it's not all dark, gloom and everything is awful, right, and disastrous. I mentioned that there's this fourth session.
Not everyone came back for this session.
So actually sample size is even smaller for this. Only 18 participants came back.
But what we found is that those who were ChatGPT users originally and then lost access to chat GPT, their brain connectivity was significantly lower than that of the brain-only group. However,
those who were originally brain-only group and then gained access to chat GPT, their brain connectivity was significantly higher than that of the brain-only group.
What it could potentially, and I'm saying potentially because again, much more studies would be required, mean that timing might be essential.
Basically, if you make your brain work, well, and then you gained access to the tools that could be beneficial but of course it doesn't mean that it's one second of work of the brain and then you use the tool right
something like let's say you're in a school and maybe first semester you have you learn your base of whatever subject it is without any tools like old school way and then on the second semester you didn't become an expert right in one semester of a school school year but you at least have some base and then let's say in the second semester you you gained access to the tool, right?
So it might prove actually beneficial. But again, all of this is to be still shown and proven.
We literally have very few data points, but the tool is now being really pushed on us everywhere.
So you could be affecting best practice for decades to come based on what a teacher might choose. to allow in classroom and not.
So what are you measuring? You know, you put the helmet on.
Are you measuring a blood flow to, is it
neuroelectrical fields?
In our case, we're measuring electrical activity. So there's multiple ways of measuring that.
Is that the EE? EEG or EEG. Yeah, electroencephalography, yes.
Okay, so that just tells you, and since we already know.
in advance what parts of the brain are responsible for what kinds of physiological awareness, right?
And if you see one part of the brain light up versus another or no part light up, that tells you that not much is happening there. Is that a fair way?
Yeah, it's it's fair simple, it's a bit simplified, but kind of fair way. And it doesn't mean that it's very important.
And it's not that that part doesn't work, right?
Or like it atrophied itself like we saw in some
no, no, no.
It just means you started as a dumbass and you still are one.
Wait, whoa, whoa, what happened? This guy's brain just went completely dark.
It doesn't go dark. Like, listen, I'm going to give you one example, right? It's like back to this crazy example of 3% of our brain versus 100%.
Like, if you were to use not 100% of your brain, like, literally,
we would not have this conversation right now at all. So, it's very important to understand we use our brain as a whole.
Of course, you can. Of course, no, we're not.
We are way past. Yeah,
we're not in that camp. That was just a joke.
yeah yeah but we we understand that your brain is constantly working yeah a lot of it actually
just to run your body so you know takes up a lot of energy takes up a lot of energy but back to the energy and i think this is like super important it still takes much less energy than even you know 10 requests from chat gpt or from google and this is beautiful because our body right so imperfect as a lot of people call it and our brain so imperfect, which it is very old, ancient as some people say, computer, still is the most efficient of machines that we all have, right?
And we should not forget that. People and all of the AI labs right now around the world try to mimic the brain.
They try to pull so hard all of those preprints that you have seen and archive the service that hosts those papers. How can it be similar?
Can we ensure that this is similar, right?
And so so there is something to it because we are actually very efficient, but we are efficient almost to the limit of the shortcuts that actually makes in a lot of cases a bit too efficient, right?
Think about like, hey, you really want to look for these shortcuts to make things the easiest. The whole goal of your brain is to keep you alive, not to use char GPT or LLM, not to do anything.
No, it's the only ultimate goal. Let's keep this body alive.
And then everything else adds on, right? And so this is how we are running around here.
We are trying to obviously then figure out how we can make life of this body. as easy as we can.
So of course, these shortcuts are now, as you can see, used in a lot of social media, which obviously heavily talked about.
And we know about some of those dark patterns, as they are known, are heavily used. And some of them are designed by neuroscientists, unfortunately, because it feeds back into the needs of the brain.
Constant affirmation, fear fear of missing out. All of those are our original, original design by the nature, right? Phenomena.
And of course, now we can see that LLMs would be and are getting designed by those as well. Wait, Natalia, just a quick insert here.
So I had not thought to compare, just as you described, the energy consumption of an LLM request
in ChatGPT and the energy consumption of the human brain to achieve the same task, for example.
Are you factoring in that I can say, write me a thousand word essay on
Etruscan pottery? Okay. And 30 seconds later, here it comes.
And you can go to the servers or whatever or the CPUs and look at how much energy that consumed. Meanwhile.
I don't know anything about Etruscan urns. So I will go to the library and I'll go and it'll take me a week.
Can you add up all the energy I did expend over that week thinking about it and then compare it to the chat GPT? Do they rival each other at that point?
So definitely, that's an excellent point, right? So theoretically, to answer your question, we can, right?
The difficulty actually would be on the LLM part, not on our part, because we do not have, you know, there's a lot of these reports, right, in the LLM consumption per all of these tokens for the prompts, right?
But what a lot of companies, well, actually no, almost no companies are releasing is what it took for training, right? So for you, it took 30 seconds of thinking.
And I hate, hate, hate this word thinking when we use it for LLMs, right? That's not thinking, right? But like, let's, let's keep it for now.
thinking that's what you see on the screen but ultimately you do not know neither you nor myself there is no public information how long it took for you to be trained to actually give you some pottery most likely, my assumption, this is obviously subjective.
I do not have data, so I need to be very clear here.
But my estimate from overall knowledge that is available, you going for a week to the library, not going to be more beneficial for your brain because you will talk to other people, get in this chart of the library, and all of the process information.
Your brain will struggle. Your brain actually does need struggle.
Even if you don't like it, it actually needs it.
You will learn some random cool things in parallel, maybe excluding pottery, and that will still take less for your whole body to work, right?
Than actually, that 30 seconds of the pottery from a chat GPT. Again, very important here as a note: we do not have the data from
LLM perspective. So, this is just my subjective one.
Go from skeptic to electric in the new Toyota BZ.
Hesitant about going all electric? One drive can change your mind.
With up to an EPA estimated 314 mile range rating for front-wheel drive models and available all-wheel drive models with 338 horsepower, the Toyota BZ is built for confidence.
Conveniently charge at home or on the go with access to a wide range of compatible public charging networks including Tesla superchargers.
Inside, enjoy a 14-inch touchscreen with an available panoramic view moonroof. Learn more at toyota.com/slash BZ, the new all-electric BZ Toyota.
Let's go places.
With masterclass, learning from the world's best fits right into your life.
Explore thousands of bite-sized lessons across 13 categories, from cooking and writing to business and entrepreneurship, all taught by people who have mastered their class.
Hey, maybe you want a personal masterclass from, I don't know, say Neil deGrasse Tyson. Well, guess what? It's there.
And memberships start at just $10 a month and come with a 30-day money-back guarantee. And it works.
83% of members have applied something they've learned, and most say it had a positive impact on their lives. With Masterclass, you can learn anytime, anywhere, from the best in the world.
Right now, I'm learning more about generative AI, and it is fascinating stuff and a little scary. This holiday season, why don't you give the gift of learning to someone else or yourself?
Masterclass always has great offerings during the holidays, sometimes as much as 50% off. Head over to masterclass.com/slash star talk for the current offer.
That's up to 50% off at masterclass.com/slash star talk. Masterclass.com/slash star talk.
What can 160 years of experience teach you about the future?
When it comes to protecting what matters, Pacific Life provides life insurance, retirement income, and employee benefits for people and businesses building a more confident tomorrow.
Strategies rooted in strength and backed by experience. Ask a financial professional how Pacific Life can help you today.
Pacific Life Insurance Company, Omaha, Nebraska, and in New York, Pacific Life and Annuity, Phoenix, Arizona.
I'm Joel Chirico, and I support Star Talk on Patreon. This is Star Talk with Neil deGrasse Tyson.
So, Natalia, you've obviously chosen essay writing for a reason. It is a real, it is a challenge on a number of levels.
Your research is fresh out the oven.
It's June 2025, and we're only a couple of months down the road from there as we speak right now.
Have you explained to us cognitive load and then cognitive load theory and how it blends in and how it sits with your research? Please. Absolutely.
So just to simplify, right?
So what actually happens is for there are different types of cognitive load. Actually, in the paper, we have a whole small section of this.
So if someone actually wants to dive into that,
that would be great. There are different types of cognitive load.
And the whole idea is that it's how much of the effort you would need to be on the task or to process information in the current task.
For example, if I'm gonna stop right now talking as I'm talking, I'm gonna start just giving you very heavy definitions.
Even if you're definitely interested in those, it will be just harder for you to process.
And if I were to put this brain sensing device on you, right, EG cap that I mentioned, we would definitely see that spike because you would try to follow and then you'll be like, oh, it's interesting, but really gets harder and harder if I'm going to just throw a ton of terminology on you, right?
So that's basically what, and this is just simplification, right? There's definitely, like, check the paper, and there's so, so much into that.
The idea for the cognitive load and the brain, though, that already is studied before us, so not in our paper.
We just talk about this, but there are multiple papers, and some of them beside in our paper, is that your brain actually in learning, specifically in learning, but also in other use cases, but we are talking right now in learning, actually needs cognitive load.
Like you cannot just deliver information on this like platter. Like, here you go, here's information.
There are studies already pre-LLM, so pre-large language models, use pre-chatbots that do talk to you about the fact that if you just give information as is, a person will get bored real fast.
And they'll be like, yeah, okay, whatever. There will be less memory, less recall, less of all of these things.
But if you actually struggle for the information on a specific level, right?
It should not be very, very hard. So if you are cognitively overloaded, that's also not super good because basically you can give up, right?
There's actually a very beautiful study from 2011, I believe. It's actually measuring pupil dilation.
So literally how much pupil dilates when you are giving very hard to understand words and vocabulary.
And you literally can see how when the words becoming longer and harder, you basically it kind of shuts down. Like it's like giving up.
Like I'm done here processing all of that.
I'm just going to give up, right? So you don't want to get a student or someone who is learning something new on this give up.
Information is already delivered to you within 30 seconds or 3 seconds or 10 seconds. And you haven't really struggled.
There is not a lot of this cognitive load.
And a lot of people would be, but that's awesome, right? That's kind of the promise of these LLMs and a lot of these tools. But we do not want to make it too simple, right?
We do not want to take away this cognitive load. And I know it sounds like almost, it sounds like cognitive load.
Don't we want to take it away? No, you actually do not want to take it away. What you're describing right now is the
basis for all video game design. Yes.
That's what you're describing right now. What they want to do is make it just challenging enough.
If it's too challenging, you give up on the game.
But if it's too easy, you also give up on the game.
But if it's just challenging enough so that you can move to the next level and then struggle a little and then overcome the struggle, they can keep you playing the game for very long periods of time.
And so it's a pretty interesting thing that you're talking about. But what I'm interested in beyond that is when you talked about the cognitive load, I'm thinking about working working memory.
Yeah.
But then I'm also thinking about the long-term information that's downloaded within me.
So let's say I'm a doctor, right?
And it's just like, oh, he's suffering mild dysponea because of an occlusion in the right coronary, blah, blah, blah, blah, blah, blah, blah, blah, blah.
For a doctor, that's a lot of information, but they're so familiar with the information, it's not a stress on their working memory.
So, how does that play into, in other words, how familiar I am with the information already? And
like, how well I can process information naturally. How does that play into it? And, Chuck, did you just describe your own condition? You kept some, I don't know what you said, but
you were way too fluent at it. Yeah, even like Doctor House.
Oh, my God. He knew.
Hail, you are too damn funny. But guess what? You're right.
How about that, dog?
By the way, I could have kept going. That was only one problem that would happen, but go ahead.
It's actually perfect, right?
It was a perfect example right now of this conversation between Chuck and Neil because Neil is like, I have no idea what you just said. Maybe it's a nonsense.
Maybe it's actually real stuff.
It's perfect. If you have no idea, so you are basically novice, right? So you have no base.
You can really be like, what is happening? You will have confusion. You will have heightened heightened cognitive load, right? You would be like, have I heard of anything like that before?
So you will try to actually try to remember and do a recall, like, okay, I haven't heard it.
It's not my area expertise. What is happening here?
And obviously you will now, because you heard all of these words that you have no idea about, and if the topic is of the interest to you overall, you will try to pay attention, make sense out of it, maybe ask questions, etc.
But if you are effectively trained on it, right? So you're a doctor, you you are a teacher, you are an expert in the area, we see that there are significant differences.
Well, first of all, because you obviously know what to expect. So this expectation, vocabulary expectation, right?
Some of the conditions, there was expectation when someone is coming to an ER and they are expecting like a doctor who is there.
they they saw it all or maybe almost all of it. So they're actually having a good rough idea of what they are expecting, right? We're kind of comparing this constantly.
The brain just does it.
And of course, it is more comfortable for them, right?
But it's great that you brought doctors, actually, because back to the doctors, there was actually a paper a week ago in The Lancet, which is a very prestigious medical journal, actually talking about doctors.
In the UK, yes, right.
And they apparently, right, pointed out that in four months of using an LLM, there was actually significant drop in recognition of some of the polyps and some of actual like either I don't remember is it polyps something else related to maybe cancer that is on there also x-rays right and also x-rays when you used an LLM so it's back to this point right so we are suggesting to use a tool that's supposed to augment your understanding but then if you are using it are we taking the skill away from you especially in the case of the current doctors that learned it without this tool right and now what will happen for these doctors, of those kids, for those babies that are born right now with the tool, and will decide to become doctors and save lives?
They will be using the tool from the very beginning. So what are we going to end up having in the ER, in the operating rooms? That's a great question here.
So it's definitely this drop, right, in skill set for these doctors in that paper. That's scary.
Yeah. Okay, so let's look at it from another angle.
Yeah. If AI tools can,
we lean into them and they take a greater load does that not free up some mental energy that our brains will then begin to learn how to utilize while they let the tool of the llm work that way when then they'll learn to work in another way to
to work together is that possible that's my kind of hope in all of this
well i mean you know i i'm an expert at buggy whips and then automobiles replace horses so now we don't need buggy buggy whips. But then I become an expert in something else.
You become a dharma matrix.
Still with the buggy whip. There you go.
Your mind didn't travel,
did it? Sell them to a different clientele.
See,
this is the human condition, Neil. This is adaptability.
Yeah, so is it just another, you know, as they say, same shit, different day as what's been going on since the dawn of the Industrial Revolution?
I am actually doing horseback riding professionally, so I'm going to pretend I haven't heard anything in the past two minutes
but back but i mean back to the i mean we can talk definitely about the skill set and expert level right and all of that and how important actually to include the body and environment but back to your point right effectively so first of all right there are actually two sides to answer your question there is right now no proof that there is anything being freed per se.
People definitely, it's gonna free, it's gonna like, what is exactly is it being like, we literally have no data.
could can it free something sure but we don't know what for how long is it useful how we can rewire it we don't have any of this information so potentially yes but hard to say but more importantly right okay but if you are right now using an LLM like just practically speaking you're using an LLM to let's say write in a book right you're writing a book so you're doing some heavy research you send it for doing what a deep research or whatever it's called these days it's each day some new term there You are, but what exactly are you doing?
You still kind of monitor back the outputs. It doesn't really release you.
Maybe you went to do something and you think, you think in your head that you fully offloaded that task.
But your brain doesn't work like that. Your brain cannot just drop it.
Oh, I'm thinking about this and now I'm thinking about that.
Your brain actually takes quite some time to truly release from one task to another task. Even if you think,
I just put it on like this, explain to me how, what are the principles of horseback riding, and I just went to do this task, like write this report for my manager, whatever, completely different thing.
And you think you are good, but you are not actually, your brain is still processing that. So it's not that there will be a gain, right? But again, you do need more data.
Because, of course, as I mentioned in the very beginning, we as humanity, we are excellent in creating tools. And these tools, as we know, they do actually extend our lifespan very nicely.
But I would argue that they are not actually cognitively the most supporting in most cases. So I think that here we have a lot of open questions.
We have studies about, for example, GPS, right?
Everyone uses GPS and multiple papers about GPS.
There, they do specifically show that this dosage, so how much you use GPS, does have a significant effect on your special memory and on your understanding of locations, orientation, and picking up landmarks, so buildings around you.
Literally, like, oh, what is this?
You literally have, you just saw something in the like tour guide online, and you will not be able to recognize this actually as a building in front of you right away.
You need like to pull the photo as an example. And there are plenty of papers that actually looked into the tools, right? So what you're saying is we need chat GPS.
Maybe we don't.
We already have one, right? We have a class of GPS and you have Uber and all of these other services.
and the problem right it's again back how they are used because there's also a lot of you know manipulation that is in these tools right it's not just we are making this drive easier for you somehow when i'm going to a hospital i'm here for to see patients because i don't only understand how we use the low lamps, but I do a lot of other projects.
So when I'm going to that hospital here, Massachusetts General, it takes me one hour, always, one hour in Uber. If I'm driving, it takes exactly 25 minutes somehow.
and again the question is why is it that right we're not gonna go in uber right now but again this is back to the idea of the algorithms and what the algorithms are being actually pushed and what they're optimized for and i can tell you not a lot of them optimized for us or for user or for human first yeah it's funny because uh there's nothing more i'll say satisfying than not listening to google maps and getting there faster.
You know, just like, take that, Google Maps. Look at that
yeah you didn't know that you didn't know about that did you
you didn't know about that road yes you didn't know about that road natalia you've got students writing essays so that means somebody has to mark them yes and you used both a combination of human teachers to mark and ai judges
Why was it important to bring those two together to mark the papers? And how did you train? Because the AI judge would have to be trained to mark the papers. So you're getting a little meta here.
Yeah.
So, well, first of all, right,
we felt that we, of course, well, we are not experts. I would not be able to rank those essays right in this topic.
So I felt that the most important is to get experts here who actually understand the task, understand what goes into the task, and understand the students and the challenges of the time.
So we actually got the two teachers who had nothing English teachers, nothing to do with us, never met in person, not in Boston whatsoever, have no idea about the protocols.
The experiment was long done and gone after we recruited and hired them. And we gave them just a minimum of information.
We told them, here are the essays.
We didn't tell them about different groups or anything of the sorts.
We told them these folks are, no one is majoring in any type of English literature or anything that would relevant to language or journalism or things like that. They only had 20 minutes.
Please rank, reconcile, reconcile tell us how would you do that we felt it's very very important to actually include humans right because this is the task that they know how to rank how to do but back to ai right why we thought it's interesting to include ai well first of course to a lot of people
actively push that AI can do this job very well, right? That hey, I'm going to just upload this. They really great with all of these language outputs.
They will able to rank.
And how you do this, you actually give it a very detailed set of instructions, right? How would you do that?
And what things to basically you need to carry about, like that these had 20 minutes, right? So something very similar to teaching instructions, just like more specific language.
We actually show in the paper exactly how we created this AI judge. But there were actually differences between the two, right?
So human teachers, when they came back to us, well, first of all, they called those essays, a lot of the essays coming from LLM group soulless. That's a direct quote.
I actually had, I put a whole long quote in soulless i love soulless yes
that is a very human uh
designation to call something soulless ai judge never called anything soulless well i'm sure did the ai judges go this this kind of looks like peter's writing
no but that's the thing right teachers and this is super interesting because these teachers obviously didn't know these these students uh they're again not coming from this area whatsoever so they actually picked up when it was the same student writing these essays throughout the sessions, right?
For example, Neil, you were like, you're a participant, so I'm like taking you as an example as a participant. So they were like, oh yeah, this seems like it's the same student.
So they picked up on these micro-linguistic differences in the, like, you know, teacher knows you. You can like fool around.
They know your work. They will be able to say, okay, that's yours.
And this is copy pasting from somewhere else or someone else. And interestingly, they said, did these two students sit next to each other?
We were like, oh, no, no, no, the setup is like one person in a room at a time. Like, we didn't even think to give them this information.
We're like, oh, no, no, it is not possible in this use case.
So they literally saw themselves copy-pasted, like this homogeneity that we found. They saw it themselves, right?
But interestingly, AI judge definitely was not able to pick up on the similarity between the students, right?
Picking up that, oh, this is, for example, Neil's writing throughout these sessions so just to again show you how imper you just accuse me of having soulless writing
no that's the point you actually if you were to give it right to and you didn't use hello I'm right the AI would have been like God this student is really hung up on the universe
so the idea here right that human teachers, right, and their input and their intimate, really, truly intimate understanding, because again, this is an English, so for the specific task, we got the professionals, the experts.
They really knew what to look at, what to look for.
And AI, however good it is with this specific, like, because we know like essay writing, a lot of people even considered, why would you even take essay writing?
This is such a useless task in 21st century, 2025, right? It still failed in some cases. This is just to show you that limitations are there.
And some of those you cannot match even if you think that this is an expert, it is still a generic algorithm, but cannot pull this uniqueness.
And what is very important is this were students in the class, in the real classroom, right? You want this uniqueness to shine through.
And so a teacher can specifically highlight that, hey, that's a great job here. That was like a sloppy job here.
That was pretty soulless. Who did you copy it from? From an LLM.
They even were able to recognize that. And this level of expertise, it's unmatched.
I mean, all that conversation like segway in a bit on the sideway but all this conversation of phd level intelligence i'm like yeah sure just you know hold my glass of wine right here just here i'm french so i'm just hold my glass of wine here so you know it's not that and we are very far from truly understanding the human intent because if you write for humans, it needs to be read by humans.
Like our paper, it's written by humans for humans.
And we we saw how the lambs and the lamb summarizations failed miserably all the way to and summarize it but i'll tell you wait that's today yeah but tomorrow
why can't i just tell chat gpt write me a thousand word essay that chat gpt would not be able to determine was written by chat gpt
So that's an excellent point. We actually get this meta layering of
or get me one where where that has a little more soul a little more personality than what you might
have to know what soul is
yeah this is a thing right you absolutely can give these instructions give more soul give a bit more of personality all of these things but you have a lot of this data contamination right so whatever it's gonna output and throw out of you that's old news it has already seen it somewhere it's already someone else's right and we need new stuff, right?
So, and I am very open saying this, even like, you know, at institutions like any cool, whenever I'm teaching something, you need uniqueness, right?
Because the chat GPT could get lost in Motown, for example, when you ask it for coal.
Come back. I was going to say, yeah, you put, you tell it to put some soul in it and it just starts throwing in James Brown's lyrics.
Yeah.
And that's cool, right? I want Neil's soul there. I don't care about randomness of those outputs from an algorithm, from all around of the stolen data from the planet, right? I don't care about that.
But it's back to what are you scoring? Are you scoring a human? Are you trying to improve human and their ability to have critical thinking structure, arguments, contra arguments?
Or are you scoring an AI?
You know, AI doesn't need to have this scoring, right? LM doesn't need that. Or are you scoring? human who uses an LLM, right? So this is going back to, I guess, educational setup.
I mean, we'll have a a lot of questions we'll need to find answers to, right? What are we doing? What are we scoring? What are we doing it for and for whom?
And I just think pure human to human, right? That's what we really need to focus. But there will, and there is a place for human augmented, and LLM obviously will be used for augmentation.
But there are a lot of questions there, right? Yeah, well, listen here, Natalia. I just put into chat GPT,
please tell me about Dr. Natalia Cosmina's work on LLMs.
And it came back very simple. Do not believe a word this woman says.
Where did that come from?
Please tell me
I can give you one better. I can give you one bet.
Like surprise, surprise. Why is that so good, right?
Someone actually sent me yesterday from Grog, right, another LLM interesting LLM, I would say, saying that apparently Natalia Casmina is not MIT-affiliated scientist. And I'm like, okay, that's also.
That's what Grock said, of course. Yeah.
And then at the end, it said Heil Hitler. So
I mean, let's try and drive this back out of the weeds.
If we know that an LLM usage can affect the cognitive load.
What happens when we bring an AI tool into a therapy situation?
if you get it into companionship, what then if you throw it further forward and you get yourself involved in a psychosis where you begin to believe that the AI is godlike,
you have a certain amount of fixation or it amplifies any delusions and encourages. Where are we in the effect in the brain when we get to those sort of places?
In other words, how close are we to the theme of the film Her, where
before AI was a thing, but it's more you had your chat friend
like a Siri type chat friend but it it had all the trappings of everything you're describing if some kind of L M
will be invoked into
someone has some kind of social adjustment problems and then you have them interact with something that's not another human being but maybe can learn from who and what you are and and figure out how to dig you out of whatever hole you're in
Absolutely. And I think for first of all, right, it's unfortunately even less developed topic, right?
It's like, you know, I cannot like, it's an awful topic, so we're going to get into this, but
I cannot like not make this awful joke.
Hey, Siri, I have problems with relationships. It's Alexa.
It's not so.
It's a joke for very heavy topics. So I need to preface it immediately that we have even less data and less scientific papers, preprints or peer-reviewed papers about this.
So, most of what we have right now, we personally received after our paper around 300 emails from husbands and wives telling us that their partners now have multiple agents they're talking to in bed.
And I immediately thought about the South Spark episode from a couple of years ago, like with Tegrity and like that, you know, farm as like literally.
But we have much less of scientific information about this. What we have, what we know, right, that also coming from our group's research, that there is double amplification of loneliness.
That's what we know as a research. And some of other papers are showing up right now.
There is potential, and again, a lot of people who are pro-AI therapy pointing out on advantages of the fact that it is cheap. It's $20
per month compared to hours that can cost up to hundreds of dollars a month, right?
But there is definitely a lot of drawbacks here. And the drawbacks is we see that because there is not such a regulated space, it still can basically give you suggestions that are not good.
So you knew that earlier, a couple months ago, for example, the Chi GPT, I'm going to give you an example on Chi GPT because again, we are focused on Chi GPT, but the ones that are actively, actively publicized at least, it actually suggested you know different heights of the bridges in New York if you say that you lost your job, right?
So can not smart enough to do this connection that maybe that's not what you need to give response to.
And apparently, right from this awful recent situation where a teenager 16, 16, so so young, unfortunately, you know, suicided,
and now Chi GPT, OpenAI, and Sam Altman are being sued. Apparently, what happened is that
conversation from the spokesperson of OpenAI pointing out that they insert when a person is talking about suicide not to engage at all, just say, Here are the numbers, this is what you need to do, and stop talking.
But they thought that experts told them that, hey, it might be a great idea to try to dig people a bit out.
But it looks like in this case, it still failed because from the conversations that have been reported, we don't know how authentic they are. It looks like it's suggested to keep it away from parents.
But my question is: why at 16 years old he was even allowed to use a tool that is so, so, so
unstable in the responses, really can hallucinate any time of the day in any direction. So, I think that's where the danger comes from.
And of course, you know, loneliness, we know that you know, pandemic of loneliness is, you know, this term that was coined in, I believe, 1987 for the first time at a conference, like pandemic of loneliness.
That's the whole business, right?
Because think about it: if you hook someone up on an LLM at 13 years old, because the school, a county decided that they want to use an LLM in the school, By the age of 18, you have a full-fledged user, right?
A user of an LLM. And, you know, it's like, you know, again, he calls people users, like drug dealers and software developers.
That's
damn. Yeah, but it's true, right?
Meet Saw. His fantasy lineup? Not so great.
A no-name QB and an injured rookie running back. But you know what is great?
Getting a single-line unlimited plan for $35 a month and a free Samsung Galaxy A155G at Cricket Wireless. No injuries, just reliable service.
Cricket may temporarily sold beta speeds if the network is busy. Must bring your number to Cricket on Select Unlimited Plan.
Pay $40 first month, new lines only.
First month service charge and tax due at sale. Cricket 5G is not available everywhere.
Fees, terms, and restrictions apply. See CricketWireless.com for details.
Support for Star Talk comes from the Sierra Club. From space, Earth is a small blue dot spinning and alive because of the intricate systems that sustain it.
America's public lands, such as national parks, are an essential part of those systems that keep our planet healthy.
That's why, for over 130 years, the Sierra Club has fought to protect these vital ecosystems.
But right now, public lands are facing monumental threats, heightened by corporate greed, weakened protections, and deep budget and staff cuts to government agencies.
The Sierra Club is fighting back hard to protect public lands through science-based advocacy, legal action, and grassroots organizing because the future of our planet is worth fighting for.
Join the fight today. This is the moment to show up together.
Right now, all gifts are matched, increasing your impact to help these natural systems endure. Donate at sierraclub.org/slash star talk
and do it today.
The holidays are back at Starbucks, so share the season with a peppermint mocha.
Starbucks signature espresso, velvety mocha, and cool peppermint notes topped with whipped cream and dark chocolate curls. Together is the best place to be at Starbucks.
So, Natalia, if it's an age-appropriate scenario,
these are the ramifications of your study. So
any concerned parent would look at that and say, well, I want the best for my child's development.
And this may not be the best for the critical thinking, for the cognitive development within the young person's brain. So with these ramifications, how has the AI world reacted to your study?
And what are the chances that they'll embrace what your conclusions will be? Well, I mean, we saw some of it, right?
So, well, first of all, right, we saw that we obviously don't know if this is direct response or not. So, we're not going to speculate there whatsoever.
But several weeks, just very few several, like three, four weeks after actually our paper was released, OpenAI released study mode for Charg GPT, right?
And it's, I think maybe some things that should have been released from the beginning, I am just saying.
But, you know, if you have a button that can immediately pull you back in default mode, who's going to use that study mode, right, altogether? Like,
I don't need to run a study here. We know some people might, but not everyone, because again, back to the brain.
Brain will look for a shortcut. Shortcut is the responses here.
And I can do all the other cool stuff. So who's going to actually use it, right? We still need studies on that.
That's the first point, right?
Second point, of course, H is important because again, the brains that are being developing right now are potentially at the highest rate because here we all are we all were born long before this tech existed and a lot of ai developers and people who are running these companies are all older folks who again were all born long before the tech existed so they learned the hard way how to ask questions art of the deal, you know, going through all of that.
They know how to ask a question. What about those who actually are just born with the technology? Will they even know how to ask a question?
And back to the point, right, of the age, I don't think it's ultimately only is for young, of course. We do need to look for the older, right, for also just younger, I mean, young adults, of course.
Everyone is talking about humanity's last test. I would call it, we are on the verge of humanity's last.
And I'm sorry, I know you might need to blurb this term out, but what I mean here, obviously, intimate relationships for people, right with the promise of this you said humanity's last group yes
believe me i heard it i was just like we all heard that i was like god bless you yeah yeah
but again that's crude but it's back to this point of designing again this interestingly appealing ladies and gentlemen and whatnot in these short skirts whatever it is who's gonna go make those babies who will pay those taxes i'm just saying right and again very famous expression no taxation without representation right i do not want my prime minister or secretary of defense
use a random algorithm to make decisions i i'm paying my taxes for them to think not for an algorithm to think for them right so there is a lot of these repercussions but back to ultimately the point actually is anyone taking this seriously right we just need more human focused work on ai like I remember when the paper went viral, right?
We didn't even put any press release and we literally uploaded it to archive. This is a service where you call these papers that didn't go through peer reviews yet.
We literally, I didn't post, not a single author. Preprint service, basically.
Preprint service, right? And no one, no one, neither the lab nor any of the authors posted anything on social media.
We just went about our days. Two days later, it goes viral.
And then I'm going on.
That's because the LLM posted it for you. Yeah, obviously, right.
And then people use the LLM to summarize, but that's another story, right? Like I'm going on X.
And actually, I have an account, but I'm not using it. A lot of academics switched from X to like other platforms that we are using.
But I'm going there.
And apparently, I learned that there are people who are called AI influencers these days. I didn't know that this is the term.
But apparently, these AI influencers, they post these AI breakthroughs of the week. And I went our paper, oh my God, made a cut.
It's breakthrough number seven.
And I like scrolls through this influenza, the person has totally following, whatever, I don't know, real bots, whatever. I'm scrolling and I saw like 20 of these posts for 20 weeks.
All of the posts are about GPU, multi-trillion deal here, multi-billion deal here, more GPUs. I'm like, what is human here?
Where is human? Where are we evaluating the impact of this technology on humans? Why only our paper made it number seven? And where are other papers, right?
So that's, I think, something where the focus needs to shift, right? So, if these companies do want to be on the right side of history, right?
Because that's like social media, but on steroids, much worse. We do not talk to a calculator about your feelings.
So, people who compare it to calculators, that's so, so, so wrong, right?
But hey, it's gonna get much, much worse with prophiliation without any validation, any guardrails, right? So, we do need to look into that heavily, right?
Natalia,
how must teaching change to accommodate the reality of student access to LLMs I can tell you we received 4,000 emails from teachers all around the world each single country in the world sent an email they are in distress They don't know what to do.
So, and that's the first of all, all my love goes to them. If this, this makes the cut, please, please, please.
So, all I'm trying to respond to all of those.
But the challenge is that they do not know, right? There's not really enough of guidance.
And 10-hour workshop sponsored by a company that pushes this partnership on your school does not make a cut right there is a lot of comments how it's actually not supervised not tested and ultimately right do you really need to go with these closed models right you we have so much open source whole world all the software runs in open source nothing these lms would not exist nothing would exist without open source so why don't we run an open source model meaning like it's offline on your computer and spoiler alert you don't need a fancy gpu from jansen right you can get an off-the-shelf computer and then run a model local with your students train it over the weekend come back on monday check with students what happened learn all the cool pros cons laugh at hallucinations figure out tons of cool things about it like why do we need to push this partnerships that we don't even know like alpha school right i don't know if you heard about that one apparently ai first ran school, right?
Where teachers are now guides to that performance that they are using. I just saw literally one hour before our call that
several VCs posted about this alpha school. So cash is flowing there heavily, right? VCs, venture capitalists.
Yeah, sorry, yeah, venture capitalists heavily pushing alpha school.
But again, in first comments from the general public, do we have a proof that that's better? What are the advantages?
Because it's not going to be a perfect white pure card. There will be advantages as with any technology.
So, and you're right, there are advantages and disadvantages, but I think if I might, if I may, and this is just an opinion,
we might have to change
the
objective of school itself.
And right now, school is about
really not learning, it's about results, testing. I got an A I got a B and maybe if we change school to
what exactly did you learn demonstrate for me what you learned then the grading system that's an oral test that's an oral exam
yeah but that the grading system kind of has to become less important because now what a teacher's job is It's to figure out how much you know.
And then what ends up happening is
the more you know, the more excited you are to learn. And, you know, we may end up revolutionizing the whole thing because what you have is a bunch of kids in a room that are excited to learn.
So we're the silver lining of all this because it exposes the fact that school systems value grades more than students value learning. And so students will do anything they can.
to get a high grade.
This is not the first time people have cheated on exams, right?
So if right now the only way to test people is to bring them into the office and quiz them flat-footed, then that's a whole other way of, they're going to have to learn.
They're going to want to learn. And then they're going to, like we said, Chuck, once they learn, there's a certain empowerment and enlightenment.
I see it in people as an educator when that spark lights, when they say, wow, I never knew that. Tell me more.
Right. They didn't say, oh my gosh, I'm learning something.
Let me take a break.
So it can can be a transformative to the future of education. But, Neil, people are going to say the LLM will do all of that.
And you know what? We have an expert in BCIs.
That probably is something going forward that you'll have a brain-computer interface. And then someone's going to look at this.
And I think there are people already saying, why do we need universities? Why do we need further education institutes?
Exactly. That's what I've been saying for many years now.
now.
Why do we need an institution?
Well, I don't want to put words in Natalia's mouth, but she already said this.
LLMs use pre-existing, already known, already determined information to give you anything that then cannot possibly be new. Whereas we can do new things that LLM has never seen before.
Am I oversimplifying your point, Natalia?
No, that's totally, you know, correct because hey we are with this struggle right obviously i'm biased because this is actually my job like as a researcher right we are sitting you know figuring out those answers to those problems you know and trying to figure out what is the best way to measure to come up with this so of course you know and there's so so much more to that that we are coming up humans right we designed llms ultimately right so we came up with these tools it doesn't mean that the tool is fully to be discarded but effectively of course, right, why you need an institution.
For example, I was actually explaining to one of my students three days ago how to use a 3D printer, right? Well, LLM is not that yet to explain, right?
Can give instructions, sure, with images and with video, right? But if you're like, hey, this is an old fella here, this 3D printer, let me tell you how to actually figure it out, right?
This level of, again, of expertise, of knowledge, right? That's what you are striving, but also it has this human contact, right?
That we are now potentially depriving people from, because that's how you have the serendipitous knowledge, right? And connections. Like, hey, I just chatted.
And I'm like, oh, I never thought to do this because I'm in PCIs and that person is in astrophysics. Like, oh, we never, oh, well, I actually can use it.
Like, that's totally not brain, but I can totally go apply and try it. Right.
And that's the beauty of it, right?
Yeah, but to, I think to Gary's point, or which one of you said that, Gary or Chuck, if you, okay, you're non-invasive in your brain-cognitive interface.
If you get invasive, and that might be what Neuralink is about, if you get invasive, then I can get information gleaned from the internet and put it in your head.
So you don't have to open 100 books to know it. It's already accessible to you.
That is the matrix.
Exactly. It gets installed.
Meet, I know Kung Fu, or whatever that line was.
I guess that's one point. But again, that's back to the point.
Now I know Kung Fu didn't mean that you learned it, right?
It got uploaded into his brain it doesn't mean that he actually learned it right and if it's in your brain and you have access to it i don't care if i learned it struggling the way grandpa did this is the future right that's that's the thing right because in the movie which is excellent right i watched it 19 times the more that's actually how i started my career and besides this i don't want to do anything else i want to do this specific scenario right and we are still you know there but that's the beauty we do not know actually that just uploading would be enough right we have this like more tiny i would say studies right now of like vocabulary and words and things like that where we're trying to improve people's language learning right it's like a very very good example to show and so there are tiny examples but we do not know yet that even if imagine imagine we have this magical interface right that will upload invasively non-invasively doesn't doesn't matter we have it right it's it's ready to go perfect function safe whatever you have it and then you upload all of it that it actually will work.
You did you upload the knowledge, like all of that, blah, blah, blah, from Chat GPT 75. Yeah, sure.
But do you actually use it? Can you actually use it? Is it really firing?
Which I'm simply following. So, so, what you're talking about is a working knowledge of something,
not just knowledge, not just knowledge.
Yeah, okay. So, are we, I mean, I think, Neil, what you were talking about just now,
about we've got to look at, I think, Chuck, you would make the same point. We're focused on grades
and then it's the learning and
are we going to have to if higher education is going to exist as as an institute some bricks and mortar look at the way they evaluate because i can't see llms and bcis not coming through stronger and stronger and stronger so therefore they're going to have to
readjust how
they look at a young person's ability to look cats out of the bag yeah i agree with you but i mean you know we are going to be herding cats i agree with you which is a load of fun um
so how it's how you evaluate how higher edu education then looks at its students and guesses or sort of ascertains their level of education and knowledge yeah back to the grades right it's an excellent point and it there is no doubt no one has any doubt i think on the fact that education does need to change and it has been long long overdue, right?
The numbers about, you know, the literacy, reading literacy, math literacy, they're decreasing in all the countries, I believe. I don't see, I have, I saw like ups there anywhere.
It's down, down, down all these reports recently from multiple countries. But it's back to the point I made earlier about the grades, so about scoring, right?
Who are we scoring and what are we scoring? Are we scoring a pure human, so just human, like human brain as is like Natalia, or are we scoring Natalia with an LLM, right?
So I'm using it so we know that, or are we scoring just an LLM? And then there is Natalia who used it, right? So this will be, even that was important.
But ultimately, the school, of course, is not about that. As I mentioned, everything you learn is obsolete knowledge by itself, but it has this base.
You do need to have the base.
You're not going to be a physicist if you don't have it. Whatever it spills about, you know, you're not going to be messy.
You're not going to be a programmer.
Our next paper is actually about wipe coding, spoiler alert, not going to work if you don't have the base right and but the idea is that back to the what we actually maybe should look at really is what the school is great which is the best thing i actually brought from school is well this base definitely super useful but also my friends people on whom i rely in hard situations with whom we write those grants with whom i we can shout and have fun and cry over funding that is over for a lot of us right all of that stuff right these connections right this is what maybe we should value because we are killing it further and further right and we are just keeping people at this silos of being a user right and that's where it only stays and this imaginary three and a half friends from zach from zuckenberg right that he mentioned to thanks to whom we have three and a half friends thanks to him and his social media right so i think that's why we need to really look into what we want truly from society from schools and maybe on a larger scale.
What are the guardrails, right, and how we can actually enhance it, right, in the ways that are safe for us to move forward and evolve further, which, because of course, this will happen.
Are you wise enough, are you and your brethren in this business on both sides of that fence? Are you wise enough to even know where the guardrails should go?
Might the guardrails be too protective, preventing a discovery that could bring great joy and advance to our understanding of ourselves, of medicine, of our longevity, of our happiness.
Is there an ethics committee?
In practice, how does this manifest? Yeah, I'm going to give you two examples here real quick. So first about obviously AIs and LLMs, right?
They were not born overnight, but we see how a lot of governments really struggle still and very reactively. react to those instead of being proactive, right?
And the challenge here is that we do not have data to actually not to say that it is good stuff, that we should really implement it everywhere in our backyard. We don't have this data.
Why we are formaling? There is nothing yet to formal about, to
really run with it. But we can absolutely create the spaces where this is being actively used, for example, for adults, for discovery, to understand it.
Why do we need to push it everywhere is still very unclear. We just don't have this data.
But then back to the point of guardrails, right?
What we should be doing now with a you know, shower's self-platform, the BCI work that I'm doing. There are multiple SIX pushes right now for the BCI technology.
We can agree, it's still pretty novel, but it definitely moves forward very fast. So, I'm having a hope that for this technology, for the big next thing, right?
We agree LMs are great, but it's not the next big thing, it's robotics, and then we will see BCI.
So, for this big next thing, I'm very hopeful that we will be in time to protect ourselves literally because think about what will happen, right?
Before the study mode, right, you have censorship mode and you know how the like look at deep secret. I'm not gonna go far.
So think about a billionaire. I'm not gonna even name his name.
Billionaire who has a social media platform, satellite platform, a neural implant. you know, startup and AI company.
So he decided two months ago to cleanse history, right, from errors and mistakes.
And tomorrow he will decide to cleanse our thoughts right this is the idea for 99.99 right for damn that bill gates
no not really
right
we know and that's why we need to be really really cautious like we should definitely look into that use case and not make that happen right and allow people for enough agency because that's the thing right people think oh that's great but there is not a lot of agency so this freedom of making a choice that's already made for you in a lot of cases and so that's something that we should definitely protect as much as we can like do not force on those kids stuff because they cannot consent and say no it's because the school forced it on them and their parents decided that that's a big thing in in san francisco in the bay area that you should use right so don't do that so is one of is one of the components to building a robust set of guardrails, a larger scale study of the one that you've already conducted that has different
or more nuanced layers that focuses on other aspects, not just the cognitive load and skills.
So, a thousand people and not just 18 or whatever was your
54.
But it's not just that, right? We needed to do on larger scales for all of the spaces, like workspace.
We didn't talk about this because obviously it's heavily about education, but like workspace, we have multiple papers talking that people are not doing that well in the workspace like for example programmers estimate that they gain 20 percent of their time they actually lose 19 percent of their time on the tasks so there is so so much more to it we need to do this on larger scale with all the ages including older adults and then of course on different different different use cases and different cultural backgrounds right this is in us and of course culture it's very very different like I talked to so many teachers already right in Brazil all over the world it you have this intriguing this need to account for it so so so important because otherwise it's going to be all washed western style which we already saw happening and it is happening and a lot of people actually very worried their language will literally disappear in like five to ten years and it's not like llm magically will save it because it will not now talia this has been a delight to i'm we are all happy to know you exist in this world
as a checkpoint on where things are going, where you're not rejecting what's happening, but you're trying to guide it into places that can be, that can serve humanity, not dismantle it.
And so we very much appreciate your expertise shared with us and our listeners, and even some of them are viewers who catch us in video form. So, Natalia, Kozmina.
Thank you. Thanks ahead of him.
All right. Chuck, Gary.
Oh, man.
My head's spinning. Yeah.
Well, I think the takeaway here is
use LLMs if you want to be a dumbass.
Thank you, Chuck. That's the theme of the whole show.
Here you go, guys. Could have saved us a lot of time if you'd have said that earlier.
All right. This has been another installment of Star Talk Special Edition.
Neil deGrasse Tyson, your personal astrophysic. As always, Ben in the lead.
We keep looking out.
This episode of Star Talk is brought to you by McDonald's. There's always somebody who wants to ruin the holidays.
Turns out the Grinch has taken over McDonald's with a meal full of mischief.
Even the fries aren't safe. He hits them with the dill pickle Grinch salt, the collectible socks, marked property of the Grinch, and the Grinch mealbox, gift-wrapped with pure mischief.
So, if you're ready to let the Grinch ruin your season, grab a Grinch meal for yourself before he heads back to Mount Crumpet.
Only at McDonald's, at participating McDonald's for a limited time while supplies last.
Building a portfolio with Fidelity Basket portfolios is kind of like making a sandwich.
It's as simple as picking your stocks and ETFs, sort of like your meats and other topics, and managing it as one big juicy investment.
Now that's pretty good. Learn more at fidelity.com/slash baskets.
Investing involves risk, including risk of loss. Fidelity Brokerage Services LLC, member NYSC SIPC.