Back to Index

Ep. 244: Cal Newport’s Thoughts on ChatGPT


Chapters

0:0 Cal's intro
12:15 Ho does ChatGPT work? (And should we worry about it?)
54:5 Cal talks about ZocDoc and Better Help
57:15 Is there anything AI won’t do better than humans?
62:40 How will AI end up disrupting knowledge work?
67:42 Should I quit web development before AI eliminates the industry?
72:2 Will AI create mass job loss in the next five years?
78:5 Cal talks about Blinkist and Ladder Life
81:23 NPR leaves Twitter

Transcript

That is the deep question I want to address today. How does chat GPT work? And how worried should we be about it? I'm Cal Newport, and this is Deep Questions, the show about living and working deeply in an increasingly distracted world. I'm here in my Deep Work HQ, joined once again by my producer, Jesse.

So Jesse, you may have noticed that we have been receiving a lot of emails in the last few months about chat GPT. Yeah. And related AI technologies. And our listeners want to know my thoughts on this, right? I'm a computer scientist. I've thought about the intersection of technology and society, and I've been silent about it.

Well, I can reveal the reason why I've been silent about it is that I've been working on a big article for the New Yorker about exactly this technology, how it works and its implications for the world. And my general rule is when I'm writing an article, I don't talk about that subject publicly until the article is done.

I mean, that's basic journalistic practice, but actually, Jesse, I've never told this story, but that rule was really ingrained in me when I was in college. So when I was coming up as a young writer, you know, I got started pretty early, wrote my first book in college. I was commissioned to write something for the New York Times.

I don't remember exactly what it was, maybe an op-ed, something to do with college students or something like this. And I had an early blog at that time. And I wrote something on the blog like, "Hey, isn't this exciting? I'm going to write an article for the New York Times." And maybe like I put the short email on there and it was like, "Yeah, we'd love to have you write the piece or something." And that editor went ballistic.

Really? Yeah. Cancelled the piece. Cancelled the piece? Chewed me out. Now, this was early internet, right? I mean, this was 2004 probably. So I don't know. Maybe it was more, it felt more like a breach then, but ever since then, if I'm writing an article. Did you ever talk to that editor again?

No. I ended up writing a lot for the Times, but not really until 2012. Was that, you're going to be your first big splash? That would have been my first big splash. Were you like depressed for a couple of days? A little shook up. And then starting with So Good They Can't Ignore You and going forward, I had a really good relationship with the Times, especially through digital minimalism.

I've written tons of articles for them, but there is a lesson learned. So I've written now, the day that we're recording this podcast, April 13th, my new article, The New Yorker has been published. So I am free. The gag order has been lifted and we can get into it when it comes to chat GPT.

In fact, I'll even load the article up here on the screen. For those who are watching, you will see this on the screen. If you're not watching, you can watch at youtube.com/calnewportmedia. Look for episode 244. You can also find that at the deeplife.com episode 244. Here Jesse is the long awaited article.

The title is What Kind of Mind Does Chat GPT Have? The subhead is large language model seems startlingly intelligent, but what's really happening under the hood? It's a big long article. So it's good. I'm excited to read it. Yes, we can talk chat GPT. I mean, you probably haven't been following it too closely just based on our conversation.

Some people really in the weeds and some people don't want to know. I'm guessing you're not in the weeds on chat GPT, but I could be wrong. No, I'm not in the weeds at all. I listened to like a few of the hard fork episodes on it. That was about it.

And what was the tone of those episodes? They were given some examples of what it was when it first came out. I probably listened to them like six weeks ago. Yeah. And then, yeah, that was kind of it. Well, so I'll give a quick primer then before we get into the guts of what I want to talk about today.

So chat GPT is a chat bot. You can sign up for an account at open AI and it's a web interface. You type in questions and chat GPT or prompts or requests and chat GPT response types text back like you're talking to someone over Slack or instant messenger. So this was released in November, late November of last year, and almost immediately people began circulating online screenshots of particularly impressive interactions or particularly funny interactions that they had with chat GPT.

Here's one of the first ones to go viral. I talk about this one in my article. So here's a tweet of a screenshot that went along. This was from a software developer named Thomas Pacek and he asked chat GPT the following, write a biblical verse in the style of the King James Bible explaining how to remove a peanut butter sandwich from a VCR.

Chat GPT rose to the challenge and wrote a response that begins, and it came to pass that a man was troubled by a peanut butter sandwich for had been placed within his VCR and he knew not how to remove it. And he cried out to the Lord saying, Oh Lord, how can I remove this sandwich from my VCR for it is stuck fast and will not budge.

And the response goes on. Here's another early viral example of chat GPT's prowess. This was a someone named Riley Goodside who asked chat GPT to write a sign failed scene in which Jerry needs to learn the bubble sort algorithm. And chat GPT once again rose to the occasion. A not a properly formatted script, but has some of the aspects of it.

It opens in a monk's cafe. It says Jerry is sitting at the counter with George. Jerry sighs and says, I can't believe I have to learn to bubble sort algorithm for my computer science class. George laughs, bubble sort. That's the most basic sorting algorithm there is. Even a monkey could do it.

Audience laughs. Jerry. Yeah, well, I'm not a monkey. I'm a comedian. And then the screen, the scene goes on from there. All right. So this is the type of thing chat GPT can do. These impressively perceptive answers to pretty esoteric requests. Now, if you go back and actually watch the media cycle around chat GPT, which I have to say is driven very strongly by Twitter.

I think the fact that anyone can sign up for an account and that screenshots of your interactions can be easily embedded in the Twitter really helped get the hype cycle around this technology spinning much more furiously than it has for past artificial intelligence innovations. Anyways, if you go back and look at this media cycle, it took a week or two before the tone shifted from exuberance and humor.

Like if you look at this example I just gave about Seinfeld, the tweet says, actually not that one I meant of the VCR. The tweet says, I'm sorry, I simply cannot be cynical about technology that can accomplish this. So it went from this sort of exuberance and happiness to something that was a little bit more distressing.

Here's a couple of examples I want to bring up here. Here is an article from NBC news. The headline is chat GPT passes MBA exam given by a Wharton professor. Oh, that got people worried. Here is a another article from around this period from time magazine headline. He used AI to publish a children's book in a weekend.

Artists are not happy about it. It details a product design manager who used chat GPT to write all the texts of a book, which he then self published on Amazon and started selling a bit of a stunt, but it implied certain types of future scenarios in which this technology was taking away creative work that really made people unsettled.

As we get closer to the current period, I would say the tone shifted since the new year and in particular coming into March and April, the tone shifted towards one of alarm, not just about the focused economic impacts that are possible with this type of technology, but some of the bigger societal, if not civilization level impacts of these type of technologies.

I would say one article that really helped set this tone was this now somewhat infamous Kevin Roos piece from the New York Times that is titled a conversation with Bing's chat bot left me deeply unsettled. Bing released a chat bot after chat GPT based on a very similar underlying technology.

Kevin Roos was, I guess, beta testing or using this new tool and fell into this really sort of dark conversation with the chat bot where among other things, the chat bot tried to convince Kevin to divorce his wife. The chat bot revealed that she had a sort of hidden double identity.

I think that identity was called venom, which was a very sort of dark personality. So Kevin set a tone of, Ooh, I'm a little bit worried. And it escalated from there. In late March, we get this op ed in the New York Times. This is March 24th, written by some prominent authors, Yuval Harari, Tristan Harris, and Azza Raskin.

And they really in this article are starting to point out potential existential threats of these AIs. They are arguing strongly for, we need to take a break and step back from developing these AIs before it becomes too late. Here's their last paragraph. We have summoned an alien intelligence. We don't know much about it, except that it is extremely powerful and offers us bedazzling gifts but could also hack the foundations of our civilization.

We call upon world leaders to respond to this moment at the level of challenge it presents. The first step is to buy us time to upgrade our 19th century institutions for an AI world and to learn to master AI before it masters us. A few days after this op ed, an open letter circulated signed by many prominent individuals demanding exactly this type of pause on AI research.

Okay, so this is the setup. Chat GPT is released. Everyone's using it. Everyone's posting stuff on Twitter. Everyone's having fun. Then people start to get worried about, wait a second, what if we use it for X, what if we use it for Y? And then people got downright unsettled.

Wait a second, what if we've unleashed an alien intelligence and we have to worry about it mastering us? We have to stop this before it's too late. So it really is a phenomenal arc and this all unfolded in about five months. So what I want to do is try to shed some clarity on the situation.

The theme of my New Yorker piece, and I'm going to load it on the screen and actually read to you the main opening paragraph here. The theme of my New Yorker piece is we need to understand this technology. We cannot just keep treating it like a black box and then just imagining what these black boxes might do and then freak ourselves out about these stories we tell ourselves about things that maybe these black boxes could do.

This is too important for us to just trust or imagine or make up or guess at how these things function. So here's my, the, the nut graph of my, my New Yorker piece. What kinds of new minds are being released into our world? The response to chat GPT and to the other chat bots that have followed in its wake has often suggested that they are powerful, sophisticated, imaginative, and possibly even dangerous.

But is that really true? If we treat these new artificial intelligence tools as mysterious black boxes, it's impossible to say. Only by taking the time to investigate how this technology actually works from its high level concepts down to its basic digital wiring, can we understand what we're dealing with.

We send messages into the electronic void and receive surprising replies, but what exactly is writing back? That is the deep question I want to address today. How does chat GPT work? And how worried should we be about it? And I don't think we can answer that second question until we answer the first.

So that's what we're gonna do. We're gonna take a deep dive on the basic ideas behind how a chat bot like chat GPT does what it does. We'll then use that to draw some more confident conclusions about how worried we should be. I then have a group of questions from you about AI that I've been holding on to as I've been working on this article.

So we'll do some AI questions and then the end of the show, we'll shift gears and focus on something interesting. So an unrelated interesting story that was arrived in my inbox. All right, so let's get into it. I want to get into how this actually works. I drew some pictures, Jesse, be warned.

I am not a talented graphic designer. That's not true. I'm not much of an artist. So Jesse watched me hand drawing some of these earlier on the tablet. I got to say this ain't exactly Chiat Gay, the famous ad agency level work here, but you know what? It's going to get the job done.

So I have five ideas here I'm going to go through. And my goal is to implant into the high level ideas that explain how a computer program can possibly answer with such sophisticated nuance. These weird questions we're asking it how it can do the Bible verse about the VCR, how it can do a Seinfeld scene with bubble sort.

And we're going to do this at the high level. We're going to essentially create a hypothetical program from scratch that is able to solve this. And then at the very end, I'll talk about how these big ideas I'm going to these five ideas I'm going to present. We'll talk about how that's actually implemented on real computers, but we'll do that real fast.

That's kind of a red herring. The neural networks and transformer blocks and multi headed attention. We'll get there, but we'll do that very fast. That's the big conceptual ideas that I care about. All right. I need idea number one about how these type of programs work is word guessing.

Now, I got to warn you, this is very visual. Everything I'm talking about now is on the screen. So if you're a listener, I really would recommend going to YouTube dot com slash Cal Newport media and going to episode two forty four. And if you don't like YouTube, go to the deep life dot com and go to episode two forty four because it's very visual.

What I'm going to do here. All right. So what we see on the screen here for idea number one is word guessing. And I have a green box on the screen that represents the LLM or large language model that would underpin a chat bot like chat GPT. So what happens is.

If you put a incomplete bit of text into this box, an example here is I have the partial sentence fragment, the quick brown fox jumped. The whole goal of this large language model is to spit out what single next word should follow. So in this case, if we give it the quick brown fox jumped as input in our example, the language model has spit out over.

This is the word I'm guessing should come next. All right. So then what we would do is add over to add the word to our sentence. So now our sentence reads the quick brown fox jumped over. So we've added the output. We've expanded our sentence by a single word.

We run that into the large language model and it spits out a guess for what the next word should be. So in this case is the. And then we would now expand our sentence. The quick brown fox jumped over the. We put that as input into the model. It would spit out the next word.

This approach, which is known as auto regressive text generation, is what the models underneath all of these new generation chatbots like chat GPT actually use. They guess one word at a time. That word is added to the text and the newly expanded text is put through the model to get the next word.

So it just generates one word at a time. So if you type in a request to something like chat GPT, that request plus a special symbol that means, OK, this is the end of the request and where the answer begins, that is input. And it'll spit out the first word of its response.

It'll then pass the request plus the first word of its response into the model to get the second word of the response. It'll then add that on and pass the request plus the first two words of its response into the model to get the third word. So this is generating text one word at a time.

It just slowly grows what's generating. There's no recurrency in here. It doesn't remember anything about the last word it generated. Its definition doesn't change. The green box, my diagram never changes once it's trained. Text goes in, it spits out a guess for the next word to add to what's being generated.

All right. Idea number two, relevant word matching. So how does it figure out how do these large language models figure out what word to spit out next? Well, at its core, what's really happening here, and I'm simplifying, but at its core, what's really happening here is the model is just looking at the most relevant words from the input.

It is then going to match those relevant words to actual text that has been given. We call these source text in my article. So examples of real text, it can match the relevant words to where they show up in real text and say, what follows these relevant words in real text?

And that's how it figures out what it wants to output. So in this example, the most relevant words are just the most recent words. So if the input into our box is the quick brown fox jumped over, perhaps the model is only going to look at the last three words, fox jumped over.

Then it has this big collection over here to the side of real text that real humans wrote, all these examples. And it's going to look in there and it's going to say, okay, have we seen something like fox jumped over show up in one of our input texts? And okay, here's an input text that says, as the old saying goes, the quick brown fox jumped over the lazy brown dog.

So it looks, fox jumped over. Here it is. We found it in one of the source texts. What came after the words fox jumped over? The great, let's make the what we guess. Now, of course, in a, in a real industrial strength, large language model, the relevant words aren't just necessarily the most recent words.

There's a whole complicated system called self-attention in which this, the, the model will actually learn what type, which words to emphasize as the most relevant words. But that's too complicated for this discussion. The key thing is, is just looking at some words from the text, effectively finding similar words in real texts that it was provided and saying what happened in those real texts.

And that's what it figures out, how it figures out what to produce next. All right. This brings us to idea number three, which is voting. So the way I just presented it before it was, you know, Hey, just start looking through your source text till you find the relevant words, see what follows output it.

That's not actually what happens. We want to be a little bit more probabilistic. So what I would say a closer way of describing what happens is we can imagine that our large language model is going to look for every instance, every instance of the relevant words that we're looking for.

And it's going to see what follows those instances and keep track of it. What are all the different words that follow in this example, fox jumped over. And every time it finds an example of fox jumped over, it says, what word follows next? Let's give a vote for that word.

And so if the same word follows in most of the examples, it's going to get most of the votes. Now I'm using votes here sort of as a metaphor. What we're really doing here is trying to build a normalized probability distribution. So in the end, what we're going to get, what the, what the large language model is going to produce is for every possible next word, it is going to produce a probability.

What is the probability that this should be the next word that follows? But again, you can just think about this as votes, which word received the most votes, which word received the second most votes, how many votes this word received compared to that word. And we're just going to normalize those is really what's happening.

But you just think about it as votes. So in this example, we see the phrase, the quick brown fox jumped over the lazy brown dogs. I mean, that shows up in a bunch of different sources. So the word the gets a lot of votes. So it has sort of a high percentage here.

But maybe there's similar phrases. Like look at this example here, the cat jumped over a surprised owner. Cat jumped over is not the same as fox jumped over because cat is different than fox. But in this voting scheme, we can say, you know what, cat jumped over is similar to what we're looking for, which is fox jumped over.

So what word follows that? The word a, well, we'll give that a small vote. And so now what we're able to do is not only find every instance of the word, the relevant words that we're looking for and generate votes for what follows, we can also start generating weaker votes for similar phrases.

And in the end, we just get this giant collection of every possible word and a pile of votes for each. And what the system will do is now randomly select a word. So that's why I have a picture. It's just, you would admit expertly drawn picture of a three dimensional dice.

That's pretty good. Honest question. Did you know that was a dice before? Oh, yeah. Look at that guy's 3d rendering. Yeah. So for those who are listening, I have a picture of a dice to indicate randomness. It'll then randomly select which word to come next and it'll weigh that selection based on the votes.

So if in this case, the has most of the votes, it almost certainly that's the, the word it's going to choose to output next, but look, a has some votes. So it's possible that it'll select a, it's just not as likely the word apple has zero votes, 0% probability because it never shows up after the phrase, anything similar to fox jumped over.

No phrase similar to that is ever followed by apple. So there's no chance it'll select it and, and so on. And actually in these systems, the output is a vote or a percentage like this for every possible, they call them tokens, but word that could follow next. Here's a quiz, Jesse in the model on which chat GPT is based, how many different words do you think it knows?

So in other words, when it, when it has to generate, okay, here's a pile of votes for every possible next word, how many words or punctuations are a billion? No, it's 50,000, 50,000. So it has a vocabulary of 50,000. It's not all words, but basically it knows like tens of thousands of words.

And how big is like the biggest dictionary? That's a good question. Yeah. I don't think it's as, it probably, there's probably a lot of esoteric words. Yeah. Because the thing is it has some like vectors describing all these words, follow it along throughout the system. So it really does affect the size, how big, so it's like you want to have a big enough vocabulary to talk about a lot of things, but not so big that it really inflates the system.

Right. So voting is just my euphemism for probability, but this is what's, what's happening. So now we have a bunch of source text and we imagine that for our relevant words, we're just finding all the places in these source texts where relevant words show up or similar relevant words show up and see what follows it.

And in all cases, generate votes for what follows it, use those votes to select what comes next. All right. So this brings us to idea four, idea one through three can generate very believable text. This is well-known in natural language processing systems that do more or less what I just described.

If you give it enough source text and have it look at a big enough window of relevant words and then just have it spit out word by word in the way we just described that auto regressive approach, this will spit out very believable text. It's actually not even that hard to implement.

In my New Yorker article, I point towards a simple Python program I found online. It was a couple hundred lines of code that used Mary Shelley's Frankenstein as its input text. It looked at the last four words in the sentence being generated. That's what it uses to relevant words.

And I showed in the article, this thing generated very good Gothic text. Right. So that's how you generate believable text with a program. And notice nothing we've done so far has anything to do with understanding the concepts the program is talking about. All of the intelligence, the grammar, the subtleties, all of that that we see so far is just being extracted from the human text that were pushed as input and then remixed and matched and copied and manipulated into the output.

But the program is just looking for words, gathering votes, selecting, outputting blindly again and again and again. The program is actually simple. The intelligence you see in an answer is all coming at this point from the input text themselves. All right. But we've only solved half the problem. If we want a chat bot, we can't just have our program generate believable text.

The text have to actually answer the question being asked by the user. So how do we aim this automatic text generation mechanism towards specific types of answers that match what the user is asking? Well, this brings in the notion of feature detection, which is the fourth out of the five total ideas I want to go over today.

So what happens with feature detection is a response. We have a request and perhaps the answer that follows the request is being input into our large language model. So I've shown here a request that says write instructions for removing a peanut butter sandwich from a VCR. Then I have a bunch of colons and I have the beginning of a response.

The first step is to write because everything gets pushed into the model. You get the whole original question and you get everything the model has said so far in its answer. Right. Word by word, we're going to grow the answer. But as we grow this answer, we want the full input, including the original question input into our models.

That's what I'm showing here. Feature detection is going to look at this text and pattern match out features that it thinks are relevant for what text the model should be producing. So these yellow underlines here, instructions in VCR. So maybe that's one feature points out. It extracts from this text.

These are supposed to be instructions about a VCR. And maybe this orange underline, another feature says the peanut butter sandwich is involved. And so now the model has extracted two features. These are VCR instructions we're supposed to be producing and they're supposed to involve a peanut butter sandwich. The way we then take advantage and by we, I mean the model, the way we take advantage of those features is that we have what I call in my article rules.

I have to say, AI people don't like me using the word rules because it has another meaning in the context of expert decision systems. But just for our own colloquial purposes, we can call them rules that extract the each rule, think of it as an instructions for extracting features like a pattern matching instruction, and then a set of guidelines for how to change the voting strategy based on those particular features.

So here's what I mean. Maybe there's a rule that looks for things like instructions in VCR and it figures out, okay, we're supposed to be doing instructions about a VCR. And its guidelines are then when looking to match the relevant words. And in this example, I have the, I'm saying the relevant words are step is to, so like just the end of the answer here.

When looking to match step is to, when we find those relevant words, step is to showing up in a source text that is about VCR instructions, give extra strength to those votes. So here I have on the screen, maybe one of the input texts was VCR repair instructions. And it says when removing a jam tape, the first step is to open the tape slot.

So we have step is to open. So open is a candidate for the next word to output here. Because this source document matches the feature of VCR instructions, our rule here that's triggered might say, hey, let's make our vote for open really strong. We know it's grammatically correct because it follows step is to.

But we think it's also has a good chance of being semantically correct because it comes from a source that matches the type of things we're supposed to be writing about. So let's make ourselves more likely to do this. Now think about having now a huge number of these rules for many, many different types of things that people could ask about it.

And for all of these different things, someone might ask your chat program about peanut butter sandwiches, VCR repair, Seinfeld scripts, the bubble sort algorithm for anything that someone might ask your chat program about. You have some rule that talks about what to look at in the source text to figure out it's relevant and very specific guidelines about how should we then change our votes for words that match source text, that match these properties, these complicated rules.

If we have enough of these rules, then we can start to generate text that's not only natural sounding, but actually seems to reply to or match what is being requested by the user. Now I think the reason why people have a hard time grasping this step is they imagine how many rules them or them and a team of people could come up with.

And they say, I could come up with a couple dozen. Maybe if I worked with a team for a couple of years, we could come up with like a thousand good rules. But these rules are complicated. Even a rule as simple as how do we know they're asking about VCR instructions and how do we figure out if a given text we're given is a VCR instruction text?

I don't know. But I think about that and look at a lot of examples. And maybe if we worked really hard, we could produce a few hundred, maybe a thousand of these rules. And that's not going to be nearly enough. That's not going to cover nearly enough scenarios for all of the topics that the more than 1 million users who've signed up for chat GPT, for example, all the topics they could ask about.

It turns out that the number of rules you really need to be as adept as chat GPT just blows out of proportion, any scale, any human scale we can think of. I did a little bit of back of envelope math for my New Yorker article. If you took all of the parameters that define GPT-3, which is the large language model that chat GPT then refined and is based on.

So the parameters we can think of as the things they actually change, actually train. So this is really like the description of all of its rules. If we just wrote out all of the numbers that define the GPT-3, we would fill over 1.5 million average length books. So the number of rules you would have to have if we were writing them out would fill a large university library full of rules.

That scale is so big, we have a really hard time imagining it. And that's why when we start to see, oh my goodness, this thing can answer almost anything I send to it. It can answer almost any question I ask of it. We think there must be some adaptable intelligence in there that's just learning about things, trying to understand and interact with us because we couldn't imagine just having enough rote rules to handle every topic that we could ask.

But there is a lot of rules. There's 1.5 million books full of rules inside this chat GPT. So you have to wrap your mind around that scale. And then you have to imagine that not only is that many rules, but we can apply them in all sorts of combinations.

VCR instructions, but also about a peanut butter sandwich, also in the style of King James Bible, stack those three rules and we get that first example that we saw earlier on. All right. So then the final idea is how in the world are we going to come up with all those rules?

1.5 million books full of rules. How are we going to do that? And this is where self-training enters the picture. These language models train themselves. Here's the very basic way this works. Imagine we have this 1.5 million books full of rules and we start by just putting nonsense in every book.

Nonsense rules, whatever they are. Right. So they don't do it. The system doesn't do anything useful right now, but at least we have a starting point. And now we tell the system, go train yourself and to help you train yourself, we're going to give you a lot of real text, text written by real humans.

So when I say a lot, I mean a lot. The model on which chat GPT is based, for example, was given the results of crawling the public web for over 12 years. So a large percentage of anything ever written on the web over a decade was just part of the data that was given the chat GPT to train itself.

And what the program does is it takes real text, little passages of real text out of this massive preposterously large data set. And it will use these passages one by one to make its rules better. So here's the example I have on the screen here. Let's say one of these many, many, many, many sample texts we gave chat GPT was Hamlet.

And the program says, let's just grab some text from Hamlet. So let's say we're in Act 3 where we have the famous monologue to be or not to be, that is the question. What the program will do is just grab some of that text. So let's say it grabs to be or not to be.

And then it's going to lop off the last word. So in this case, it lops off the word be. And it feeds what remains into the model. So when you lop off be here, you're left with to be or not to. It says, great, let's feed that into our model.

We have 1.5 million books full of rules. They're all nonsense, because we're early in the training. But we'll go through each of those books and see which rules apply and let them modify our voting strategy. And we'll get this big vector of votes, then we'll randomly choose a word.

And let's say in this case, the word is dog because it's not going to be a good word because the rules are really bad at first, but it'll spit out some words. Let's say it spits out dog. Now the good news is for the program, because it took this phrase from a real source, it knows what the next word is supposed to be.

So on the screen here in orange, I'm showing it knows that be is what is supposed to follow to be or not to. So it can compare be to what it actually spit out. So the program spit out dog, it compares it to the right answer, the right answer is be.

And here's the magic, it goes back and says, let me nudge my rules. There's a formal mathematical process it does to do this. But let me just go in there and just kind of tweak these rules. Not so the program accurately spits out be, but so it spits out something that is minutely more appropriate than dog, something that is just slightly better than the output it gave.

So based on this one example, we've changed the rules a little bit, so that our output was just a teeny bit better. And it just repeats this again, and again, and again, hundreds of thousands of passages from Hamlet, and then from all the different Shakespeare works, and then on everything ever written in Wikipedia, and then on almost everything ever published on the web, bulletin board entries, sports websites, archived articles from old magazine websites, what just sentences of sentences, sentences, lop off a word, see what it spits out, compare it to the right answer, nudge the rules, take a new sentence, lop off the last word, stick in your model, see what it spits out, compare it to the real one, nudge the rules.

And it does that again, and again, and again, hundreds of billions of times. There's one estimate I found online that said training chat GPT on a single processor would take over 350 years of compute time. And the only way that they could actually train on so much data so long was to have many, many processors working in parallel, spending well over a million dollars, I'm sure worth of compute time just to get this training done.

And it still probably took weeks, if not months to actually complete that process. But here's the leap of faith I want you to make after this final idea. If you do this training, the simple training process on enough passages drawn from enough source text covering enough different types of topics from VCR instructions to Seinfeld scripts, these rules through all of these nudging, these 1.5 million books worth of rules will eventually become really, really smart.

And it will eventually be way more comprehensive and nuanced than any one team of humans could ever produce. And they're going to recognize that this is a Bible verse. You want VCR instructions here and bubble sort is an algorithm. And this chapter from this textbook talks about bubble sort.

And these are scripts and this is a script from Seinfeld. And actually, this part of the script for Seinfeld is a joke. So if we're in the middle of writing a joke in our output, then we want to really upvote words that are from jokes within Seinfeld scripts. All of these things we can imagine will be covered in these rule books.

And I think the reason why we have a hard time imagining it being true is just because the scale is so preposterously large. We think about us filling up a book. We think about us coming up with two dozen rules. We have a hard time wrapping our mind around just the immensity of 1.5 million books worth of rules trained on 350 years worth of compute time.

We just can't easily comprehend that scale. But it is so large that when you send what you think is this very clever request to chat GPT, it's like, oh, this rule, that rule, this rule, this rule. Boom, they apply. Modifier votes. Let's go. I think that amazes you. So those are the big ideas behind how chat GPT works.

Now, I know all of the my fellow computer scientists out there with a background in artificial intelligence are probably yelling at your podcast headphones right now saying, well, that's not quite how it works, though. It's not it doesn't search for every word from the source text, and it doesn't have rules individually like that.

It's instead in a much more complicated architecture. And this is all true. It's all true. I mean, the way that these models are actually architected or in something called a transformer block architecture, GPT three, for example, has ninety six transformer blocks arranged in layers one after another. Within each of these transformer blocks is a multi headed self attention layer that identifies what are the relevant words that this transformer block should care about.

It then passes that on into a feed forward neural network. Is this neural networks that actually encode inside their weights, connecting their artificial neurons that actually encode in a sort of condensed, jumbled, mixed up manner, more or less the strategy I just described. So the feature detection that's built into the weights of these neural networks, the connection between certain features being identified, combined with certain relevant words, combined with vote strengths for what were should come next.

All of that is trained into these networks during the training. So all the statistics and everything is trained into these as well. But in the end, what you get is a basically a jumbled, mixed up version of what I just explained. I sat down with some large language model experts when I was working on this article and said, let me just make sure I have this right.

These high level five ideas. That's more or less what's being implemented in the artificial neural networks within the transformer block architecture. And they said, yeah, that's that's what's happening. It's again, it's mixed up, but that's what's happening. And so when you train the actual language model, you're not only training it to identify these features, you're baking in the statistics from the books and what happens with these for all that's getting baked into the big model itself.

That's why these things are so large. That's why it takes 175 billion numbers to define all the rules for, let's say, GPT-3. But those five ideas I just gave you, that's more or less what's happening. And so this is what you have to believe is that with enough rules trained enough, what I just defined is going to generate really believable, impressive text.

That's what's actually happening. Word guessing one word at a time. With enough rules to modify these votes and enough source text to draw from, you produce really believable text. All right, so if we know this, let us now briefly return to the second part of our deep question. How worried should we be?

My opinion is once we have identified how these things actually work, our fear and concern is greatly tempered. So let's start with summarizing based on what I just said. What is it that these models like Chats GPT can actually do? Here's what they can actually do. They can respond to a question in arbitrary combination of known styles, talking about arbitrary combination of known subjects.

So they can write about arbitrary numbers of known styles, talking about arbitrary combinations of known subjects. Known means it has seen enough of those things, enough of the style or enough writing about the topic in its training. That's what it can do. So say, write about this and this in this style.

Bubble sort Seinfeld in a script. And it can do that and it can produce passable text if it's seen enough of those examples. And that's also all it can do. So let's start with the pragmatic question of is this going to take over our economy? And then we'll end with the bigger existential question.

Is this an alien intelligence that's going to convert us into matrix batteries? So start with that. Is that capability I just described going to severely undermine the economy? And I don't think it is. I think where people get concerned about these Chats GPT type bots in the economy is they mistake the fluency with which it can combine styles and subjects with a adaptable fluid intelligence.

Well, if it can do that, why can't it do other parts of my job? Why can't it handle my inbox for me? Why can't it build the computer program I need? You imagine that you need a human-like flexible intelligence to produce those type of texts that you see. And flexible human-like intelligence can do lots of things that are in our job.

But it's not the case. There is no flexible human-like intelligence in there. There's just the ability to produce passable text with arbitrary combination of known styles on arbitrary combinations of known subjects. If we look at what most knowledge workers, for example, do in their job, that capability is not that useful.

A lot of what knowledge workers do is not writing text. It is, for example, interacting with people or reading and synthesizing information. When knowledge workers do write, more often than not, the writing is incredibly narrow and bespoke. It is specific to the particular circumstances of who they work for, their job, and their history with their job, their history with the people they work for.

I mentioned in my New Yorker piece that as I was writing the conclusion, that earlier that same day, I had to co-author an email to exactly the right person in our dean's office about a subtle request about how the hiring, faculty hiring process occurs at Georgetown, carefully couched, because I wasn't sure if this was the right person, carefully couched in language about, "I'm not sure if you're the right person for this, but here's why I think, and we talked about, this is why I'm asking about this.

We had this conversation before." Writing in GPT, ChatGPT's broad training, could have helped it accomplish that narrow task on my behalf. And that's most of the writing that knowledge workers actually do. And even when we have relatively generic writing or coding or production of text that we need a system to do, we run into the problem that ChatGPT and similar chatbots are often wrong.

Because again, they're just trying to make good guesses for words based on the styles and subjects you asked it about. These models have no actual model of the thing it's writing about. So they have no way of checking, does this make sense? Is this actually right? It just produces stuff in the style, like what is an answer supposed to more or less sound like if it's about this subject in this style?

This is so pervasive that the developer bulletin board Stack Overflow had to put out a new rule that says no answers from ChatGPT can be used on this bulletin board. Because what was happening is ChatGPT would be happy to generate answers to your programmer questions. It sounded perfectly convincing.

But as the moderator of the Stack Overflow board clarified, more often than not, they were also incorrect. Because ChatGPT doesn't know what a correct program is. It just knows I'm spitting out code. And based on other code I've seen and the features, this next command makes sense. And most of the commands make sense, but it doesn't know what sorting actually means.

Or that there's a one off issue here. Or that equality isn't quite, you need equality, not just less than or whatever, right? Because it doesn't know sorting. It just says given the stuff I've spit out so far and the features I detected from what you asked me, and all this code I've looked at, here's a believable next thing to spit out.

So it would spit out really believable programs that often didn't work. So we would assume most employers are not going to outsource jobs to an unrepentant fabulist. All right, so is it going to be not useful at all in the workplace? No, it will be useful. There'll be very bespoke things I think language models can do.

It's particularly useful, what we've found in the last few months, where these technologies seems to be particularly useful is when you can give it text. They can do this too. You can give it text and say, rewrite this in this style or elaborate this. It's good at that. And that's useful.

So if you're a doctor and you're typing in notes for electronic medical records, it might be nice that you can type them in sloppily. And a model like GPT-4, Chats GPT, might be able to take that and then transform those ideas into better English. It's the type of thing it can do.

It can do other things like it can gather information for us and collate it in a way like a smart Google search. That's what Microsoft is doing when it's integrating this technology into its Bing search engine. It's a Google plus. So I mean, Google is already pretty smart, but you could have it do a little bit more actions.

I mean, so there's going to be uses for this, but it is not going to come in and sweep away a whole swath of the economy. All right, let's get to the final deeper question here. Is this some sort of alien intelligence? Absolutely not. Once you understand the architecture, as I just defined it, there is no possible way that these large language model based programs can ever do anything that even approximates self-awareness consciousness or something we would have to be concerned about.

There is a completely static definition for these programs once they're trained. The underlying parameters of GPT-3, once you train it up, for example, do not change as you start running requests through it. There is no malleable memory. It's the exact same rules. The only thing that changes is the input you give it.

It goes all the way through these layers in a simple feed forward architecture and spits out a next word. And when you run it through again with a slightly longer request, it's the exact same layers, spits out another word. You cannot have anything that approaches consciousness or self-awareness without malleable memory.

To be alive, by definition, you have to be able to have a ongoing updated model of yourself in the world around you. There's no such thing as a static entity where nothing can change. There's no memory that changes, nothing in it changes that you would consider to be alive.

So no, this model is not the right type of AI technology that could ever become self-aware. There's other models in the AI universe that could be where you actually have notions of maintaining and updating models of learning, of thinking about yourself, interacting with the world, having incentives, having multiple actions you can take.

You can build systems that in theory down the line could be self-aware. Large language model won't be it. Architecturally, it's impossible. All right, so that's where we are. We've created this really cool large language model. It's better than the ones that came before. It's really good at talking to people, so it's easy to use and you can share all these fun tweets about it.

This general technology, one way or the other, will be integrated more and more into our working lives, but it's going to have the impact, in my opinion, more like Google had once that got really good, which was a big impact. You can ask Google all these questions, how to define words.

It was very useful. It really helped people, but it didn't make whole industries disappear. I think that's where we're going to be with these large language models. They can produce text on arbitrary combinations of known subjects using arbitrary combination of known styles where known means they've seen it a sufficient number of times in their training.

This is not a cow from 2001. This is not an alien intelligence that is going to, as was warned in that New York Times op-ed, deploy sophisticated propaganda to take over our political elections and create a one-world government. This is not going to get rid of programming as a profession and writing as a profession.

It is cool, but it is not, in my opinion, an existential threat. What's transformative in the world of AI probably will not be in the immediate future transformative in your day-to-day life. All right? So, Jesse, there's my professor sermon for the day. >> JESSE: Solid. >> COREY: You don't want to get me started on computer science lectures because I could fall into my old habits, my professorial habits, and really bore you.

>> JESSE: So how many rules will there be in five years? Will it double? >> COREY: I don't know how much bigger it can get. Yeah, it's a good question. So the jump from GPT-2 to GPT-3. So GPT-2 had some of the largest number of parameters before GPT-3 came out.

It had 17 billion or something like this, and GPT-3 has 170 billion. I was talking to an expert at MIT about this. The issue about making this too much larger is they're already sort of giving it all of the text that exists. And so at some point, you're not going to get back bigger returns.

So he said there's two issues with this. If you make your networks too small, they're not complicated enough to learn enough rules to be useful. But if you make them too large, you're wasting a lot of space. You're just going to have a lot of redundancy. I mean, it can only learn what it sees in its dataset.

So if 175 billion parameters is well fit to this massive training data that we use for these chatbots, then just increasing the size of the network is not going to change much. You would have to have a correspondingly larger and richer training dataset to give it. And I don't know how much more, at least for this very particular problem of producing text, I don't know how much more richer or larger of a dataset we could give it.

I actually think what the direction happening now is how do we make these things smaller again? GPT-3 is too big to be practical. 175 billion parameters can't fit in the memory of a single GPU. You probably need five different specialized pieces of hardware just to generate a single word.

That's not practical. That means I can't do that on my computer. That means if everyone at my office is constantly making requests to GPT-3 as part of their work, we're going to have this huge computing bill. So actually a lot of the effort is in how do we make these things smaller?

Just focus on the examples that are relevant to what these people actually need. We want them to be small. We want it eventually to have models that can fit in a phone and still do useful things. So GPT-3 I think was, and that's what all these other ones are based off of, that was open AI saying what happens if we make these things much bigger?

And now we're going to go back to make them smaller. And if you actually read the original GPT-3 paper, their goal with making it 10 times bigger was not that it was going to have in a particular domain 10 times better answers. They wanted to have one model that could do well in many unrelated tasks.

And if you read the paper, they say, look, here's a bunch of different tasks for which we already have these large language models that do well, but they're customized to these tasks. They can only do that one task. And what they were proud about with GPT-3, if you read the original paper, is this one model can do well on all 10 of these tasks.

It's not that it was actually doing much better than the state of the art in any one of these things. You don't need necessarily to hand train a model for each task. If you make it big enough, it can handle all the different tasks. So it wasn't getting 10 times larger did not make GPT-3 10 times better at any particular task.

In fact, in most tasks, it's as good as the best, but not much better. It was the flexibility and the broadness, but that's good to see. It's cool for these web demos. But going forward, the name of the game, I think is going to go back to, actually, we need to make these things smaller so that we can not have to use an absurd amount of computational power just to figure out that dog should follow the quick brown fox jumped over the lazy brown.

We need to maybe be a little bit more efficient. But anyways, I'm not particularly... It's a cool technology, but I don't know. I think once you open this, it's just not as worrisome. When it's a black box, you can imagine anything. I definitely like that Yuval Harari op-ed was definitely influenced by Nick Bostrom's super intelligence, which we talked about on the show a few months ago, where he just starts speculating.

He's a philosopher, not a computer scientist. Bostrom just starts speculating. What if it got this smart? What could it do? Well, what if it got this smart? What could it do? Just thinking through scenarios about... And he was like, "Well, if it got smart, it could make itself smarter, and then it can make itself even smarter, and it become a super intelligence." And then they have all these scenarios about, "Well, if we had a super intelligent thing, it could take over all world politics.

Because it'd be so smart and understand us so well that it could have the perfect propaganda. And now the bot could get us all to do whatever it wanted us to do." It's all just philosophical speculation. You open up these boxes, and you see 175 billion numbers being multiplied by GPUs doing 1.5 million books worth of pattern detection vote rules to generate a probability vector so it can select a word.

All right. Well, there's my computer science sermon. I have a few questions I want to get to from you, my listeners, that are about artificial intelligence. First, I want to mention one of the sponsors that makes this nonsense possible. And we're talking about our friends at ZocDoc. ZocDoc is the only free app that lets you find and book doctors who are patient reviewed, take your insurance, and are available when you need them to treat almost every condition under the sun.

So if you need a doctor, instead of just saying, "I'll ask a friend," or, "I'll look it up in the yellow pages," you can instead go to ZocDoc, your ZocDoc app, and it'll show you doctors who are nearby that take your insurance. You can read reviews. And then once you sign up with these doctors, they'll often use ZocDoc to make it easier for you to set up appointments, get reminders about appointments, send in paperwork.

Both my dentist and my primary care physician use ZocDoc, and I find it really useful because all of my interactions with them happen to that interface. So go to ZocDoc.com/deep and download the ZocDoc app for free. Then find a book, a top rated doctor today, many who are available within 24 hours.

That's Z-O-C-D-O-C.com/deep, ZocDoc.com/deep. The show is also sponsored by BetterHelp. As we often talk about when it comes to the different buckets relevant to cultivating a deep life, the bucket of contemplation is in there. Having an active and healthy life of the mind is critical to a life that is deep in many different ways.

It is easy, however, due to various events in life to fall into an area where your relationship to your mind gets strained. Maybe you find yourself overwhelmed with anxious thoughts or ruminations or depressive moments where you feel lack of affect. Whatever it is, it is easy for your mind to get out of kilter.

Just like if your knee started hurting, you would go to an orthopedist. If your mind started hurting, and by the way, there's really loud construction sound going on. This is the restaurant below us is being made. It's better be worth it. But returning to BetterHelp, let me add this example.

If you find yourself becoming increasingly enraged because of restaurant construction noise that occurs during your podcast, among the other things that could happen to affect your mental health, you need a therapist. Orthopedists will fix your bum knee. Therapists helps make sure that your cognitive life gets healthy, gets resilient, gets back on track.

The problem is it's hard to find therapists. If you live in a big city, all the ones near you might be booked or they might be really expensive. This is where BetterHelp enters the scene. If you're thinking about starting therapy, BetterHelp is a great way to give it a try because it's entirely online.

It's designed to be convenient, flexible, and suited to your schedule. You just fill out a brief questionnaire to get matched with a licensed therapist, and you can switch therapists anytime for no additional charge. So discover your potential with BetterHelp. Visit BetterHelp.com/deepquestions, one word, today to get 10% off your first month.

That's BetterHelp, H-E-L-P.com/deepquestions. I hope this restaurant's good, Jesse, after all the disruption. They put up the signage. I don't know if you saw that. I didn't see the signage. I've heard the music. Motocat. That's the name? Yeah. It's a Tacoma Park reference. Okay. Yeah. It's an old character from Tacoma Park history.

Anyways, soon it will be open. I heard I was talking to the guy, he thought May. Yeah. So soon. Nice. All right, let's do some questions. What do we got? All right. First question is from Manav, a student at Yale. Looking at tools like ChatGPT makes me feel like there's nothing AI won't eventually do better than humans.

This fear makes it hard to concentrate on learning since it makes me feel that there isn't certainty in my future. Are my fears unfounded? Well, Manav, hopefully my deep dive is helping dispel those fears. I want to include this question in part to emphasize the degree to which the hype cycle around these tools has really been unhelpful.

Because you can so easily embed screenshots of interactions of ChatGPT, a lot of people started trying it because the attraction of virality is very strong for lots of people in our online age. So it brought ChatGPT to the awareness of a lot of people and generated a lot of attention.

Now once we had a lot of attention, how do you one up that attention? Well, then you start thinking about worries about it. You start thinking about what if it could do this, what if it could do that? From what I understand, I'm not as plugged into these online worlds as others, but there's a whole tech bro push during the last few months that was increasingly trying to push.

It can do this, it can do that, it can do this. Exactly the same tonality with which the same group talked about crypto two years ago. Nothing is going to be done by ChatGPT, just like currency will be gone in three years because of crypto. They turned all their attention onto that.

And that got really furious. And everyone's trying to one up each other and do YouTube videos and podcasts about, "No, it could do this, no, it can do that." And then this created this weird counter reaction from the mainstream media, because the mainstream media has a, right now, an adversarial relationship with the Silicon Valley tech bro crowd.

They don't like them. So then they started pushing back about, "No, it's going to be bad. No, no, these tech bros are leading us to a world where we're going to be able to cheat on tests. No, forget cheat on tests. It's going to take all of our jobs.

No, forget take all of our jobs. It's going to take over the government and become super intelligent." So they started the counteraction to the overblown enthusiasm of the tech bros became an overblown grimness from the anti-tech bro mainstream media. All of it fed into Twitter, which like a blender was mixing this all together and swirling this spiral of craziness higher and higher until finally just the average person like Manav here at Yale is thinking, "How can I even study knowing that there will be no jobs and we'll be enslaved by computers within the next couple of years?" All right, so Manav, hopefully my deep dive helped you feel better about this.

Chats GPT can write about combinations of known subjects and combinations of known styles. It does not have models of these objects. It has no state or understanding or incentives. You ask it the right about removing a peanut butter sandwich from a VCR. It does not have an internal model of a VCR in a sandwich on which it's experimenting with different strategies to figure out which strategy works best and then turns to its language facility to explain that to you.

It just sees peanut butter as a possible next word to spit out and you ask about peanut butter and your response, so it puts more votes on it. Mixing, matching, copying, manipulating existing human text, the humor and the jokes it spits out, the accuracy and the styles it uses are all intelligence borrowed from the input that it was given.

It does not have a broad adaptable intelligence that can in any significant sense impact the knowledge work sector. It's important to emphasize, Manav, it's not like we're one small step from making these models more flexible, more adaptable, able to do more things. The key to chat CPT being so good at the specific thing it does, which is producing text in known styles on known subjects, is that it had a truly massive amount of training data on which it could train itself.

We could give it everything anyone had ever written on the internet for a decade and it could use all of that to train itself. This is the problem when you try to adapt these models to other types of activities that are not just producing text. You say, "What I really want is a model that can work with my databases.

What I really want is a model that can send emails and attach files on my behalf." The problem is you don't have enough training data. You need training data where you have billions of examples of, "Here's the situation, here's the right answer." And then like most things that we do, we learn after a small number of examples.

A model to do other activities other than produce text needs a ton of data. And in most other types of genres or activities, there's just not that much data. So one of the few examples where there is, is art production. This is how Dali works. You can give it a huge corpus of pictures that are annotated with text and it can learn these different styles and subject matters that show up in pictures then produce original artwork.

But that's one of the few other areas where you have enough uniform data that it can actually train itself to be super adaptable. So I'm not worried that Manav, like all of our jobs will be gone. So your fears are unfounded. You can rest easy, study harder for your classes.

All right, let's keep it rolling. What do we got, Jesse? - All right, next question's from Aiden. It seems almost inevitable that in 10 years, AI will be able to perform many knowledge workers jobs as well as a human. Should we be worried about the pace of automation and knowledge work and how can we prepare our careers now for increased power AI in the coming decades?

- So as I just explained in the last question, this particular trajectory of AI technology is not about to take all of your jobs. There is however, and this is why I included this question. There is however, another potential intersection of artificial intelligence and knowledge work that I've been talking about for years that I think we should be more concerned about or at least keep a closer eye on.

The place where I think AI is gonna have the big impact is less sexy than this notion of I just have this blinking chat cursor and I can ask this thing to do whatever I want. Now where it's really gonna intersect is shallow task automation. So the shallow work, the stuff we do, the overhead we do to help collaborate, organize, and gather the information need for the main deep work that we execute in our knowledge work jobs.

More and more of that is gonna be taken over by less sexy, more bespoke, but increasingly more effective AI tools. And as these tools get better, I don't have to send 126 emails a day anymore because I can actually have a bespoke AI agent handle a lot of that work for me, not in a general it's intelligent sense, but in a much more specific like talking to Alexa type sense.

Can you gather the data I need for writing this report? Can you set up a meeting next week for me with these three principles? And then that AI agent talks to the AI agents of the three people you need to set the meeting up with and they figure out together and put that meeting onto the calendar so that none of us, three of us have to ever exchange an email.

The information it gathers from the people who have it by talking to their AI agents, and I never have to bother them. We never have to set up a meeting. It's able to do these rote tasks for us, right? This was actually a future that I was exposed to a decade earlier.

I spoke at an event with the CEO of a automated meeting scheduling company called X.AI. I remember him telling me, this is the future. When you have a AI tool that can talk to another person's AI tool to figure out logistical things on your behalf so that you're never interrupted.

I think that's where the big AI impact is going to come. Now, this does not automate your main work. What it does is it automates away the stuff that gets in the way of your main work. Why is that significant? Because it will immensely increase the amount of your main work you're able to get done.

If you're not context switching once every five minutes, which is the average time the average knowledge worker spends between email or instant messenger chat checks, if you're not doing that anymore, you know how much you're going to get done? You know how much if you can just do something until you're done, and then the AI agents on your computer says, "Okay, we got the stuff for you for the next thing you need to work on.

Here you go." And you have to have no overhead communicating or collaborating, trying to figure out what to do next. You can just execute. You know how much you're going to get done? I would say probably three to four X more of the meaningful output that you produce in your job will be produced three to four X more if these unsexy bespoke AI logistical automator tools get better.

So this has this huge potential benefit and this huge potential downside. The benefit of course is your work is less exhausting. You can get a lot more done. Companies are going to generate a lot more value. The downside is that might greatly reduce the number of knowledge workers required to meet certain production outputs.

If three knowledge workers now produce what it used to take 10, I could grow my company or I could fire seven of those knowledge workers. So I think this is going to create a disruption. We underestimate the degree to which shallow work and context shifting is completely hampering our ability to do work with our minds.

But because it's like the pot that keeps getting hotter until the lobster is boiled, because it's inflecting everybody, we don't realize how much we're being held back. When computer tools aided by AI remove that, it's going to be a huge disruption. And I think ultimately the economy is going to adapt to it.

The knowledge industry is going to explode in size and scope as we can unlock all this cognitive potential on new types of challenges or problems that we weren't thinking about before. Ultimately, it'll probably be good and lead to a huge economic growth. But I think there's going to be a disruption period because we really are at such, again, we just don't emphasize the degree to how inefficient we are and how much if we could remove that inefficiency, we don't need most of the people sitting here in this office to get the same work done.

Getting over that, that's going to be the real disruption. And there's no scary how from 2001 type tool involved here. These things are going to be boring. Meeting, information scheduling, basically whatever you type in an email, it could do that for you. That's going to be the real disruption.

I don't know when it's coming, but that's coming soon. A lot of money at stake. All right, this is good. I'm like debunking people's fears. I got a therapist today. All right, next question is from Ben, a Silicon Valley engineer. I've decided that web development freelancing will be the best possible career path to achieve my family's lifestyle vision.

And I plan to freelance in addition to my full-time job until freelancing can support our life on its own. Over the last few weeks, however, I've been hearing about the breakthroughs of chat, GPT and other AI tools. Do you think I should stay on the path of learning the ropes of freelancing web development or should I focus more on the future of technology and try to stay ahead of the curve?

Well, Ben, I'm using the inclusion of AI in your question to secretly get in an example of lifestyle-centric career planning, which you know I like to talk about. So I love your approach. You're working backwards from a vision of what you want you and your family's life to be like, a tangible lifestyle, not specific in what particular job or city, but the attributes of the lifestyle.

So you're working backwards to say, what is a tractable path from where I am to accomplish that. And you're seeing here, web development could be there, freelance web development. And I don't know all the details of your plan, but I'm assuming you probably have a plan where you're living somewhere that is cheaper to live.

Maybe it's more outside or country oriented where your expenses are lower. And because web development is a relatively high reward per hour spent type activity, that strategic freelancing could support your lifestyle there while giving you huge amounts of autonomy. And satisfying the various properties that I assume you figured out about what you want in your life, these sort of non-professional properties.

So I really applaud this thinking. I also really applaud the fact that you're applying the principle of letting money be a neutral indicator of value. There's a strategy I talked about in my book, So Good They Can't Ignore You. This is a strategy in which instead of just jumping into something new, you try it on the side and say, can I actually make money at this?

The idea here is that people paying you money is the most unbiased feedback you will get about how valuable or viable the thing you're doing actually is. And so I really like this idea. Let me try freelancing on the side. I mean, I want to see that people are hiring me and this money is coming in before I quit my job.

It's a great way of actually assessing. Don't just ask people, hey, is this valuable? Do you think you would hire me? Look at dollars coming in. When the dollars coming in are enough to more or less support your new lifestyle, you then make that transition with confidence. So I appreciate that as well.

Two Cal Newport ideas being deployed here. So let's get to your AI question. Should you stop this plan so you can focus more on the future of technology and try to stay ahead of the curve? I mean, I don't even really know what that means. My main advice would be whatever skill it is you're learning, make sure you're learning a course at the cutting edge of it.

Get real feedback from real people in this field about what is actually valuable and how good you have to be to unlock that value. So I would say that don't invent your own story about how you want this field to work. Don't assume that if you know HTML and a little CSS, you're going to be living easy.

What are the actual skills people care about? What web development technologies sell? How hard are they? How good you have to be at that to actually be desirable to the marketplace? Get hard, verified answers to those questions. That's what I would say when it comes to staying ahead of the curve.

That's it. And as for some sort of fear that you becoming a web developer is quixotic because chat GPT is going to take that job away soon. Don't worry about that. So yes, make sure you're learning the right skills, not the skills you want to be valuable. But there's nothing wrong with this particular plan you have.

I love the way you're executing it. I love the way you're thinking about it. I also appreciate, I cut it, Jesse, but Ben had a joke in the beginning of his response. He said, "I've been doing lifestyle-centric career planning. I've been thinking about it. I don't like my job.

So what we're going to do is quit and I'm going to start a cattle ranch." And he was like, "Ah, just joking." I appreciate that. All right, let's do one more question before I kind of run out of the ability to talk about AI anymore. I'm getting, just, we're purging this.

It's been months that people have been asking us about AI. I'm just purging it all out there. Next week, we're going to talk all about, I don't know, living in the country or minimalism, not using social media, but we're getting all the AI out of our system today. All right, here we go.

Last question is from Anakin. "AI can solve high school level math word problems. AI can explain why a joke that has never seen before is funny. This one blows my mind. All this points to mass job loss within five years. What do you think?" Well, again, big thumbs down.

The current trajectory of AI is not going to create mass job loss in the next five years. Chat GPT doesn't know why a joke is funny. It writes jokes that are funny because it knows what part of a script you can identify as a joke that's a pattern-mashing problem.

And then it upvotes words from those parts of scripts when it's doing the word guessing game. And as a result, you get jokes that pull from existing structures of human and existing text. You don't actually know what humor is. You can see that in part if you look at that Seinfeld scene with bubble sort I talked about at the beginning of the program.

There's non-sequitur jokes in there, things that are described as the audience laughing that aren't actually funny. And that's because it's not actually looking at its script and saying, "Is this funny?" It's guessing words, guessing words that things are accurate. But let's talk about... I want to use this as an excuse to talk about another trend in AI that I think is probably more important than any of these large language models that also is not getting talked about enough.

So we talked about an earlier question, AI shallow work automation has been critical. The other major development that we're so used to now we forget, but I think is actually going to be the key to this AI shallow work automation, but also all sorts of other ways that AI interests our life is not these large language models, but it's what Google has in mind with Google Home.

It's what Amazon has in mind with Alexa. These at-home appliances that you can talk to and ask to do things, they're ubiquitous and they're ubiquitous in part because these companies made them very cheap. They wanted people to use them. They're not trying to make a lot of money off them.

Why? Why is Google or Amazon trying to get as many people as possible to use these agents at home that you could just talk to and it tries to understand you? It's data. The game they are playing is we want millions and millions of different people with different accents and different types of voices asking about all sorts of different things that we then try to understand.

And we could then use this data set to train increasingly better interfaces that can understand natural language. That's the whole ballgame. Now, chat GPT is pretty good at this. They figured out a new model. I don't want to get into the weeds. They have a human semi-supervised, semi-human supervised reinforcement learning model that they inserted during the GPT-3 training to try to align its responses better with what's being asked.

But this is the whole ballgame is just natural language understanding. And Google is working on this and Amazon is working on this and Apple is working on this with their Siri devices. And this is what matters, understanding people. The background activities, I think this is what we often get wrong.

The actual activities that the disruptive AI in the future is going to do on our behalf are not that interesting. It's not, we're going to go write an ARIA. It's we're going to pull some data from an Excel table and email it to someone. It's we're going to turn off the lights in your house because you said you were gone.

It's really kind of boring stuff. All of the interesting stuff is understanding what you're saying. And that's why Google and Amazon and Apple invested so much money into getting these devices everywhere is they want to train and generate the biggest possible data set of actually understanding what real people are saying and figuring out, did we get it right or did we get it wrong?

And can we look at examples and let's hand annotate these examples and figure out how our models work. And I think this is really going to be the slow creep of AI disruption. It's not going to be this one entity that suddenly takes over everyone's job. It's going to be that more and more devices in our world are increasingly better at understanding natural language questions, whether it be typed or spoken, and can then act accordingly, even if what we're asking to do is simple.

We don't really need these agents to do really complicated things. We just need them to understand what we mean. Most of what we do, that's a drag on our time, it's a drag on our energy, it's pretty simple. And it's something a machine could do if they just knew what it was we wanted.

And so I think that's where we should be focusing is interfacing is what matters. These 175 billion parameter models that can generate all this text is really not that interesting. I don't need a Seinfeld script about bubble sort. I need you to understand me when I say, give me the email address of all of the students from my first section of my class.

I need you just to understand what that means and be able to interface with the student database and get those emails out and format it properly so I can send the message to the class. That's what I want you to do. I don't want you to write an original poem in the style of Mary Oliver about a hockey game for my students.

I need you to just go through when I say, look at the assignment pages for the problem sets I assigned this semester, pull out the grading statistics and put them all into one document just to kind of, okay, I know what that means. And now I'm doing a pretty rote automatic computer type thing.

And I don't care if you come back and say, okay, I have three follow-up questions. So I understand what you mean. You mean this section, that section, this section, that's fine. They'll just take us another 10 or 15 seconds. I don't need, in other words, how from 2001, I just need my basic computer to understand what I'm saying.

So I don't have to type it in or click on a lot of different buttons. So I mean, I think that's really where we're going to see the big innovations is the slow creep of better and better human understanding plugged into relatively non-interesting actions. That's really where the stuff's going to take a bigger and bigger picture.

The disruption is going to be more subtle. This idea that it's now this all at once large language model represents, we have this new self-sufficient being that all at once we'll do everything. It's sexy, but I don't think that's the way this is going to happen. All right. So let's change gears here, Jesse.

I want to do something interesting to wrap up the show. First, I want to mention one of the longtime sponsors that makes Deep Questions possible. That's our friends at Blinkist. When you subscribe to Blinkist, you get access to 15 minute summaries of over 5,500 nonfiction books and podcasts. These short summaries, which are called Blinks, you can either read them or you can listen to them while you do other activities.

Jesse and I both use Blinkist as one of our primary tools for triaging which books we read. If there is a book we are interested in, we will read or listen to the Blink to see if it's worth buying. The other week, Jesse and I went through his own Blinkist list and we calculated that basically what it was roughly what 30%, I think you said 30% of the books for which you read Blinkist if you go on the buy.

But there we go. That is a case in point of the value of Blinkist. It really helps you hone in on the right books to buy. The books you don't buy, you still know their main ideas. You can deploy them. Blinkist is really a tool that any serious reader needs to add to their tool kit.

Right now, Blinkist has a special offer just for our audience. Go to Blinkist.com/deep to start your seven-day free trial and you will get 45% off a Blinkist premium membership. That's Blinkist spelled B-L-I-N-K-I-S-T. Blinkist.com/deep to get 45% off any seven-day free trial. That's Blinkist.com/deep. This offer is good through April 30th only.

Even better news, for a limited time, you can take advantage of their Blinkist Connect program to share your premium account. You will, in other words, get two premium accounts for the price of one. So when you subscribe to Blinkist, you can give an account to a friend who you think would also appreciate it.

So Blinkist.com/deep to find out more. I also want to talk about Latter. It's tax season. Tax season is when I often get stressed about putting things off because I don't even know where to get started. This got me thinking about the other classic thing that people put off, which is getting life insurance.

This is one of those things that you know you need, but you don't know where to get started. Who do you talk to for life insurance? How much should life insurance cost? You're going to have to go to a doctor and get blood drawn and your eyeball scanned before you can actually get a policy.

And we get paralyzed by all this complexity and say, forget it. Well, this is where Latter enters the scene. Latter makes it easy to get life insurance. It's 100% digital, no doctors, no needles, and no paperwork. When you're applying for $3 million in coverage or less, you just answer a few questions about your health on an application.

You need just a few minutes in a phone or laptop to apply. Latter smart algorithms work in real time, so you'll find out instantly if you're approved. No hidden fees, cancel any time, get a full refund if you change your mind in the first 30 days. Life insurance is only going to cost more as you get older, so now is always the right time to get it.

So go to ladderlife.com/deep today to see if you're instantly approved. That's LADDERLIFE.COM/DEEP. Ladderlife.com/deep. All right, Jesse, let's do something interesting. For those who are new to the show, this is the segment where I take something that you sent me to my interesting@calnewport.com email address that you thought I might find interesting.

I take something that caught my attention and I share it. So here's something a lot of people sent me. This is actually from just a couple of days ago. I have it up on the screen now, so if you're listening, you can watch this at youtube.com/calnewportmedia episode 244. It is an article from NPR with the following music to my ears headline, NPR quits Twitter.

There's a whole backstory to this. NPR is having a basically a feud with Twitter because Twitter started labeling the network as state affiliated media. The same description they use for propaganda outlets in Russia, China, and other autocratic countries that did not sit well with NPR. So then they changed it and said, okay, well, we'll call you government funded media.

But NPR said only 1% of our annual budget comes from federal funding. That's not really accurate either. You know what? That's not enough. They're walking away. And they put NPR politics, for example, put out a tweet that said all of our 52 Twitter accounts, uh, we're not going to use them anymore.

You want news for NPR, subscribe to our email newsletter, come to our website, listen to our radio program. We'll keep you up to date. You don't have to use this other guy's program. I really like to see that not because of the inter-scene political battles between Musk and the political, the different media outlets.

I mean, I wish they would just make this move even without that, but whatever gets them there, I think is good. As I've talked about so many times on this show before, it is baked into the architecture of Twitter that it is going to generate outrage. It is going to manipulate your emotions.

It is going to create tribalism and is going to really color your understanding of the world, your understanding of current events in ways that are highly inaccurate and often highly inflammatory. It's just built into the collaborative curation mechanism of all these retweets combined with the power law network. We've talked about this before.

It's not a great way to consumer share information. Now more and more outlets are doing this. So a couple of weeks ago, we gave the example of the Washington post nationals baseball team coverage shifting away from live tweeting the games and instead having live updates on their Washington post.com website.

And at the end of all those updates, then they write a recap article and it's all packaged together and you can see how it unfolded and they have more different people writing and it unfolds in real time. And I think the whole thing is great. There's no reason to be on someone else's platform, mixing tweets about the baseball game with tweets about, you know, the Martian that's going to come and destroy the earth because you didn't give it hydroxychloroquine or whatever the hell else is going on on Twitter.

Twitter was a lot of fun. It's a lot of engaging. It's not very engaging. It's not the right way to consume news. It's not the right way to spread news. And I'm really happy about this trend. I think we would be in a much calmer place as individuals. I think we'd be in a much calmer place politically.

I think we'd be in a much calmer place just from a civic perspective. If more and more sources of news, if more and more sources of expression, if more and more sources of commentary move to their own private gardens, here's my website, here's my podcast, here's my newsletter, not this giant mixing bowl where everyone and everything comes together in a homogenized interface.

And we have this distributed curation mechanism rapidly amplifying some things versus others. That's not a healthy way for a society to learn about things and communicate. And I wrote about this in Wired Magazine early in the pandemic. I wrote an op-ed for Wired. I said one of the number one things we could do for public health right now at the beginning of the pandemic would be shut down Twitter.

And I gave this argument in that article that, look, if professionals, if we retreated to websites, we could have considered long form articles with rich links in the other types of articles and sources, where the websites themselves, you could indicate authority by the fact that this website is hosted at a hospital or a known academic institution or a known news organization.

We'd be able to curate on an individual sense of this website is at, the reference in old SNL skit, clownpenis.fart, and it has weird gifs of it of eagles that are flying away with Osama bin Laden. I'm not going to pay as much attention to that, to this long form article that's coming out of the Cleveland Clinic.

And I said, if we went back, humans will be much better at consuming information. The information will be much better presented if we went back to more bespoke distributed communication, as opposed to having everyone, the cranks and the experts and the weirdos and everybody all mixed together in the exact same interface with the exact same, their tweets look exactly the same, and a completely dispassionate distributed curation mechanism, rapidly amplifying things that are catching people's attention.

We need to get away from that automation. We need to get away from that distributed curation. And we get back to more bespoke things where we can learn from where you hosted, what does this look like? What are the texts? What are the things that have you written? We can really have context for information.

Anyways, I don't want to go too far into this lecture, but basically this is a good trend. I think individually owned digital distribution of information, podcasts, websites, blogs, and news is going to be a much healthier trend than saying, why don't we all just have one or two services that everyone uses?

So good job NPR. I hope more and more news sources follow your lead in the Washington Post's, Nationals, Reporters Leagues. I think this is the right direction. Do you know the clown penis dot fart reference? No. It was in the late nineties. It was a classic SNL skit and it was like an advertisement and they really had it.

You know how they used to have those advertisements for brokerage firms? Where it's like super, I welcome, it's like Wilford Brimley, like welcome to such and such financial partners where trust is our number one, whatever. And so it's like this real serious ad and they're like, you can find out more at our website clownpenis.fart.

And then they kind of reveal by the time we set up our website, it was the only address left. So it was like the premise of it, it was like this really serious financial brokerage firm, but in the late nineties, it felt like all the URLs were gone. And so it was like the only URL that was left.

And so it was just a very serious commercial that they kept having to say clown penis dot fart, classic SNL. All right. I'm tired. Too much AI. I'm happy now to talk about AI whenever you want. My gag order has been lifted, but maybe it'll be a while until we get too much more deeply into that.

But thank you all for listening. Thank you all for putting up with that. We'll be back next week with the next, hopefully much less computer science filled episode of the podcast. And until then, as always, stay deep. (techno music)