back to indexYour Brain Is Out Of Shape? - Simple Routine To Learn Faster, Achieve Focus & Enhance Memory

Chapters
0:0 Getting Smarter in a Dumb World
27:55 With remote work dwindling, should I change careers to achieve my ideal lifestyle?
33:3 Should a 24 year old ditch his college degree to pursue another field?
37:21 How should merit be rewarded in a revenue-constrained environment?
41:11 How can one find and engage with these more niche online communities?
43:29 How can partners support each other in building a deep life?
45:33 Using lifestyle planning to turn down a promotion
54:31 How do AI’s “reason”?
00:00:00.000 |
So last week we talked about new testing data that was revealing that humans may have reached their 00:00:05.920 |
peak intelligence levels right around the time smartphones became ubiquitous, 00:00:09.840 |
and ever since that point our intelligence levels as measured by test have been going down. So my 00:00:16.840 |
conclusion last time we talked about this is that in this new world of smartphones we might have to 00:00:23.300 |
start training our brain in a way that we didn't worry about 20 or 30 years ago. And my analogy was 00:00:29.700 |
physical fitness. That when you're in the first half of the 20th century we didn't think a lot about 00:00:34.780 |
physical fitness because you got a lot of exercise just in your day-to-day life whether it's on the 00:00:40.320 |
farm or you know you worked in a city. You know people walked a lot in cities Jesse if you think 00:00:45.880 |
about it. People would just walk like four miles to work or whatever. But then when we got to the 00:00:50.200 |
second half of the 20th century and people had more sedentary lifestyles and jobs, we had cars and 00:00:54.720 |
offices and air conditioner etc. We had to start exercising. We never had to exercise before. 00:00:59.500 |
Something changed about our world. We had to start exercising. I'm arguing that we probably have to 00:01:04.520 |
do the same thing with our brains. 30 years ago, we didn't have to worry about specifically exercising 00:01:09.580 |
our brains. Today we do. All right, so here is the follow-up I want to do in today's deep dive. 00:01:15.020 |
I gave some basic advice last time. The cognitive equivalent of telling someone, hey, make sure you 00:01:21.600 |
do some movement every day. Like get your go for a walk every day. But a lot of people ask me a follow-up 00:01:27.960 |
question. What would it look like not just to sort of be active somewhat cognitively, but to get in 00:01:34.980 |
really good shape? What is the brain training equivalent of those routines you see, you know, 00:01:41.800 |
online where the people are in incredible shape? And it's like here is how you're going to completely 00:01:46.540 |
change your fitness, your health, and your body. What is the equivalent of that for your brain? 00:01:49.580 |
That's what I want to get into today. I'm going to suggest some ideas for if you wanted to be 00:01:53.800 |
the equivalent of swole, but for your brain. I'm not going to say your IQ because this really isn't 00:01:59.580 |
about doing well on IQ tests, but really just about the ability to hold your attention and think 00:02:03.020 |
interesting thoughts. What would that look like? I am not recommending that everyone needs to do this. 00:02:07.420 |
I just think it's an interesting thought experiment. What would it take to not just stave off the 00:02:14.900 |
cognitive decline induced by smartphones, but actually become someone whose brain is a real 00:02:21.040 |
asset? All right. So that is our goal here. Let me start just by reminding us of the advice I gave 00:02:31.400 |
when I last talked about this, because I'm now going to take that advice, that basic advice, and say, 00:02:36.680 |
this is just how you lay a base. The equivalent in cognitive fitness and physical fitness, just sort 00:02:42.280 |
of getting your body moving and flexible and your muscles woken up so you're ready to actually do 00:02:46.120 |
more intense training. So just as a reminder, last time we talked about this, I recommended go on walks 00:02:52.500 |
and think during the walks. Don't have something in your ear. Avoid dopamine stacking. So just make a 00:02:59.440 |
rule. I don't dopamine stack. So that means if I'm watching something, I'm not also on my phone. 00:03:03.680 |
You know, if I'm reading something, I'm not also checking out videos. I just do one source of 00:03:07.880 |
stimuli at a time. Read more. Doesn't matter what it is. Find what you're excited enough about that 00:03:15.220 |
you're more likely to stick with it, but just grapple with the written word and maybe seek out a hobby 00:03:19.480 |
that requires more concentration. Like you learn to play the guitar, practicing the guitar requires 00:03:23.780 |
concentration, right? So those type of activities remain good activities for laying a base of cognitive 00:03:29.080 |
fitness. Now I have some ideas for how you can push that base into something more elite. 00:03:34.100 |
The first idea is interval training. Now this is the basic concept that I talked about, geez, I think 00:03:41.560 |
all the way back in deep work. And it was something that I had actually messed around with prior to deep 00:03:47.500 |
work when I was a graduate student, primarily writing books for students. And I was working with a lot of 00:03:52.720 |
undergraduates to help them do better in school. And I worked with a variant of interval training 00:03:59.240 |
because a lot of these undergraduates starting around this time, we're starting to struggle with 00:04:02.680 |
concentration. This sort of the distraction era was, was, was on its way in. And we would do this 00:04:07.940 |
training routine where you would focus hard on something. In their case would be academic work 00:04:13.400 |
for a limited amount of time and they would have a timer, but they had to focus hard for that amount 00:04:18.600 |
of time. If their attention wandered, it was like, um, no, no, you got to restart that. You need to 00:04:22.200 |
focus for this amount of time. If you, if you look at your, wouldn't have been a phone back then, but if 00:04:25.580 |
you jump on the internet or something like that, you have to restart. And then only once they got 00:04:29.780 |
comfortable with a given interval, I would increase it by 10 minutes. All right, now let's try getting 00:04:34.380 |
comfortable with this new length of time. So it's interval training. You get comfortable with an 00:04:37.740 |
interval that's a stretch, and then you stretch that interval even farther. And I found in like 00:04:42.720 |
a semester or two, you could get a relatively distracted undergraduate from this sort of 00:04:47.000 |
pre-smartphone era up to like 90 minutes of concentration pretty comfortably. I can lock 00:04:52.160 |
in for 90 minutes. So I'm going to recommend that as like the first elite training technique. 00:04:56.580 |
The key is to have something demanding of concentration to be focusing on during your 00:05:03.080 |
interval training sessions. And so this could be, you know, reading a hard book or taking an online 00:05:08.040 |
course or working on a musical instrument practice is a great one that requires intense 00:05:12.420 |
concentration or maybe a professional problem. This just like really challenging. You need a good 00:05:16.940 |
thing to focus on that is hard. And then you use this timer. I recommend maybe once every two weeks 00:05:24.280 |
or so. Like if you're comfortable with something, you want to give it at least two weeks before you 00:05:28.320 |
increase it. And it might take more if it just takes you more time. If you find a duration, like you 00:05:33.660 |
really can't get through it very well. It might take you longer until you're comfortable. That's okay. 00:05:38.360 |
that stretches where your brain gets stronger. All right. The second idea, dialectical reading. 00:05:45.800 |
I think this is, we're moving here from interval training is about just the generic ability to 00:05:51.840 |
concentrate. Dialectical reading is now about your specific grappling with specific subject matters and 00:05:59.120 |
your ability to actually work with complicated information, not just concentrate, but work with 00:06:03.620 |
complicated information. So at the core of dialectical reading is the following. You want to avoid 00:06:11.500 |
the engagement with content that the algorithmically driven internet promotes. So if you're looking at 00:06:19.620 |
content on Twitter or the recommendation algorithms on YouTube, it's tuned to press a button in the human 00:06:28.540 |
psyche. That's very satisfying, but also sort of almost anti-intellectual. The button it wants to 00:06:32.900 |
press is your team is great. The other team is terrible. And here's an example of someone from the 00:06:38.340 |
other team being terrible. And it's a slam dunk, how terrible they are. That feels great for, for a human 00:06:43.680 |
beings that are wired for sort of in-group out-group thinking. It's sort of a sugar high of out-group 00:06:48.420 |
isolation. So you watch like a Twitter thread is going to be, look at these, look what they're 00:06:54.240 |
doing over. It's so obviously wrong. And aren't they terrible? And here's an example of terrible. 00:06:57.820 |
They don't even realize how terrible it is. It wants to, there's this sort of righteousness of like, 00:07:01.100 |
I am just super right. Feels great. My tribe is killing it. That tribe is terrible. And you get this 00:07:06.120 |
on all sorts of divides of, you know, it could be from sports fandom, the political divides to, 00:07:12.300 |
to whatever else is going on. Okay. Dialectical reading says, no, no, we're going to teach you how to 00:07:16.060 |
actually engage with information. So what you do is you take a topic that's interesting to you, 00:07:20.480 |
probably a topic where you feel like I have an instinct towards this is the right side. 00:07:25.280 |
Instead of looking for outrage bait content, that's just going to pat you on the back and be like, 00:07:30.320 |
yeah, you're, you're, you're right. You're Martin Luther King. And everyone on the other side is 00:07:35.480 |
Bull Connors, right? Like, or it's just like an obvious setup where you're just the best and 00:07:38.760 |
everyone else is dumb. What you do is you find a really good book pushing one side and you find a, 00:07:45.920 |
the best possible defense of the other side of the issue. And you read both those books. That's the 00:07:52.200 |
dialectic. So I'm reading two opposing, but very well-constructed takes on the same issue. 00:07:58.720 |
The collision of two smart opposing takes creates a, an intellectual leap forward that is 10x more 00:08:09.780 |
sophisticated than what you get just, you know, going to your choir stalls and having the preacher 00:08:14.980 |
preached to you. It is, put the specifics of the content aside, it is a fantastic intellectual 00:08:21.580 |
experience. I mean, this goes back to Socrates, who was a big believer in this method of intellectual 00:08:26.600 |
engagement. Ironically, and not to do a big aside, ironically, Plato writing in the Phaedraeus 00:08:33.180 |
dialogue in the voice of Socrates, Socrates was a character in it, had the character Socrates express 00:08:39.100 |
skepticism about the written word. Like, Ooh, this is a problem that we're writing things down. 00:08:44.300 |
And a lot of tech optimists pointed to that and say, look, we were worried about even books, which we 00:08:48.920 |
know are great. So of course we worry about everything, but this is actually what Plato was 00:08:52.360 |
capturing in the Phaedraeus dialogue was not that books were bad, but that when it comes to like 00:08:56.620 |
developing ideas, you need this sort of dialectical interaction, right? Where you're going back and 00:09:02.560 |
forth. I'm believing this, you're believing that, let's go back and forth and see what comes out of that 00:09:07.020 |
collision. Well, you can do that now with books, ironically, because there are enough books. Here's a 00:09:11.760 |
really smart person on this side. Let me read their take. Here's a really smart person on my side. Let me 00:09:15.280 |
read their take. Smart people opposing makes you smarter. Now, one of the fears that people have about 00:09:21.040 |
dialectical reasoning, it's going to make you smarter. It's just the level of sophistication with which you 00:09:25.460 |
engage with content goes up. It's like, look, Jesse and I are here in Washington, D.C. You have a 00:09:31.200 |
conversation with politics with just, you know, the average person you meet here in Washington, D.C. 00:09:35.720 |
is going to be a completely different experience than in Kansas City because they're everyone here. 00:09:40.820 |
There's so many experts who understand things. There's just, they have way more nuance and it's 00:09:44.940 |
just the way they understand political information is so much more complicated than the average person. 00:09:49.280 |
This dialectical reading gives you that type of more nuanced understanding. You see that it's the 00:09:54.860 |
complexities and where people are coming from. It just opens up another layer of understanding of the 00:09:59.740 |
world. People worry about it because if there's an issue they feel really strongly about, 00:10:04.320 |
they're worried that this will somehow like trick them out of their belief. To which there's two 00:10:08.500 |
responses. If encountering an opposing point of view is going to change your belief on something, 00:10:13.300 |
then maybe it's actually not that good of a belief for you to have in the first place. Maybe it's more 00:10:16.820 |
complicated than you thought. Maybe you're wrong. But two, what tends to happen is for the things where 00:10:21.340 |
you really are right, like, you know, your instincts are right, dialectical reading makes that stronger. 00:10:26.480 |
When you've encountered the very best arguments from the other side and you still come out saying, 00:10:30.420 |
this is why I believe in this, that belief is stronger. Your ability to advocate for it is 00:10:34.660 |
stronger. Your understanding of why your thing is right and how to advance that is stronger. So you 00:10:38.360 |
should never fear reading different sides of things. So dialectical reading is just going to make your 00:10:42.400 |
grasp of information of the world not only more sophisticated, it's going to make X or like TikTok 00:10:49.100 |
content essentially like unreadable or watchable. You'd be like, oh my God, I've been watching, you know, 00:10:56.420 |
succession. And now I've just gone back to leave it to beaver. Like I just, I, this is stupid. So it 00:11:03.740 |
really will change your brain in an interesting way. Make you less mad too. Hey, it's Cal. I wanted to 00:11:08.520 |
interrupt briefly to say that if you're enjoying this video, then you need to check out my new book, 00:11:13.980 |
Slow Productivity, The Lost Art of Accomplishment Without Burnout. This is like the Bible for most of the 00:11:22.240 |
ideas we talk about here in these videos. You can get a free excerpt at calnewport.com 00:11:28.860 |
slash slow. I know you're going to like it. Check it out. Now let's get back to the video. 00:11:34.560 |
Third idea, create idea documents. I'm a big proponent of this. Is there something you care 00:11:42.520 |
about or you want to care about or you just find is interesting, right? So something you might want to 00:11:46.540 |
care about is like, here's like a contemporary issue going on that affects me. And I really want to 00:11:49.860 |
understand it well and have like a point of view and be able to support my point of view. And this 00:11:53.980 |
is a cause I care about, or it's just something of intellectual interest. Like, you know what? I 00:11:58.320 |
really want to engage more with the Socratic philosophers or Greek or the big ideas of like 00:12:04.060 |
the American Republic or whatever. Keep a document where you take notes, take notes on what you've 00:12:09.660 |
learned and how you think about it and keep updating and growing this document. In the writing down of 00:12:14.980 |
information, in the editing of information, and in the rearranging of written information, 00:12:18.860 |
your cognitive scaffolding around these ideas gets more sturdy, gets more sophisticated. 00:12:25.520 |
Writing is thinking. And so as you listen to a podcast, listen to a book, maybe do like a great 00:12:33.740 |
courses, you're doing a great courses course, audio, one lesson every day during your commute, 00:12:37.940 |
take notes on what you're learning, update your sort of summary of where you stand on something. 00:12:42.860 |
That act of writing will get you significantly more leaps in understanding than if you just take 00:12:49.660 |
in the information and then don't engage with it again. So organize your thoughts on the paper is 00:12:54.060 |
going to better organize your thoughts in your head. It's just going to make you a smarter person. 00:12:57.040 |
It's why, by the way, like when you encounter professors, they're very articulate about the 00:13:00.820 |
things they care about because they're always writing about it, you know? Or, you know, if you get 00:13:05.440 |
me going on one of my hobby horse issues, like get me going on email in the workplace or something like 00:13:11.840 |
this, right? Something I've written about extensively, I can just go, right? I can just cook and I can give 00:13:17.300 |
you 15 minutes of preacher-like content with this point, that point, this point, that point. It's because 00:13:22.380 |
I've written about this so much, it has organized and structured my thoughts and I can access that 00:13:27.520 |
better. So it'll just make you sound smarter if you're working with written forms of the things you 00:13:31.940 |
care about. Next, I'm going to recommend become a connoisseur in something. So become really good, 00:13:40.380 |
not good at an activity, but good at understanding quality in that activity. You can become a connoisseur 00:13:48.240 |
of NFL football without having to become good at playing NFL football. But you can really understand, 00:13:53.920 |
for example, the complexities of different offensive schemes or defensive schemes, like what's going on, 00:13:58.120 |
right? You can become a connoisseur of music without having to become a musician, 00:14:00.940 |
become a connoisseur of movies without having to actually make your own movies. 00:14:03.880 |
I've seen this again when you come to more elite level thinkers. When you learn how to appreciate 00:14:10.020 |
what really good means in one field, that translates to others. Even if you're not an expert on those 00:14:15.160 |
other fields, you have an appreciation for expertise. You have an appreciation for quality. It adds shades 00:14:21.360 |
of nuance and subtlety into the world. It's like the number of colors available in the world gets 00:14:27.440 |
increased. When it's not just, oh, music is on and I'm hearing it. But you're instead understanding, 00:14:33.860 |
like, oh, this is why this was brilliant. Like the ability to hear, see, or encounter something and 00:14:38.640 |
say, that's brilliant. That ability to do that just allows you to see so many more shades. The world 00:14:45.460 |
becomes richer. It becomes technicolor. It goes from, you know, the industry standard aspect ratios to 70 00:14:53.400 |
millimeter projection, the cinemascope or something like that. So it changes your world and the richness 00:15:00.900 |
of information that seems available. So for whatever reason, that just works really well. Again, if you're 00:15:05.340 |
trying to push your brain, final, final suggestion, digital diet, start to get careful and selective about what 00:15:15.640 |
information you consume digitally. Because again, a lot of the stuff that you're encountering is working 00:15:21.740 |
counter to all the benefits you're trying to have here. So maybe very wary about my Instagram scrolling 00:15:29.340 |
habit. That might not be serving your interest in making your brain smarter and having been more of 00:15:37.580 |
like an elite cognitive athlete. That's kind of the equivalent of like, I really like my McDonald's 00:15:41.020 |
milkshakes, but I'm also trying to train for the marathon. I think I have to not take the milkshakes for 00:15:44.840 |
a while. So become really careful. I'm, you know, it's like, it's books, it's TV shows. If it's online content, 00:15:51.720 |
maybe you're moving towards podcast, newsletters, maybe like YouTube videos that are, you're not 00:15:59.420 |
surfing recommendations, but you've bookmarked the pages where it's someone who is, you know, 00:16:05.380 |
more thoughtful, more expert, more interesting commentary, right? I mean, there is a world of 00:16:09.760 |
just to name someone who's killing it right now. Like think about Ezra Klein right now, who's actually 00:16:13.240 |
killing it, Jesse. I don't know. He's everywhere. Him and Derek Thompson with their, um, 00:16:17.400 |
with their new book. Right. Uh, and he's great. There's a world of difference, right? If it's, 00:16:23.240 |
I'm going to listen to like Ezra's podcast versus I'm going to go on X to hear people talk about the 00:16:29.380 |
same political issues, right? It's just a, it's just a world of difference. It's, it's, uh, still 00:16:33.340 |
digital content. You're going to learn a lot about important issues in politics if you care about that, 00:16:38.720 |
but it's thoughtful, it's thought through, it's going to, it's going to be presented with nuance, 00:16:44.020 |
for example, a completely different experience than, you know, the WWE battle that's going on. 00:16:50.580 |
If you're in something like, uh, an X thread or like an apologist show or something like that. 00:16:54.980 |
So digital diet matters. So start caring about that. Just like if I was giving you advice for 00:16:59.600 |
getting like in really good shape, we were going to have to end somewhere saying you have to care 00:17:03.780 |
about what you eat. And that's the case as well. All right. So here it is. Let's see if I can summarize 00:17:07.940 |
all my suggestions, lay the base before you try to, you want to become a cognitively elite, 00:17:12.900 |
lay the base with stuff like walking dopamine stack bands, reading and concentration requiring hobbies. 00:17:20.000 |
Once you're sort of like used to that somewhat more heightened life of the mind, interval, 00:17:25.100 |
train your concentration, engage in dialectical reading, keep idea documents, become a connoisseur 00:17:31.100 |
in at least one thing, and then put in place a digital diet. Do that for a season. Your cognitive 00:17:37.980 |
experience of the world is going to be different. It's just, it's going to be, you're going to seem 00:17:40.600 |
smarter to other people. You're going to seem smarter to you. You're going to see nuances you 00:17:43.560 |
lost before. Your stress levels will probably be lower. Your ideas, your creativity is going to be 00:17:48.000 |
higher. I think life would just be better. Not everyone has to do all, not everyone has time or 00:17:51.720 |
interest in becoming like a cognitive athlete, but if you do, that's the type of thing I believe 00:17:56.340 |
you need to do. You know, this is kind of what college used to be. I guess it kind of still is, 00:18:02.320 |
was like a four-year cognitive athlete training program. Yeah. I say used to because someone sent 00:18:09.180 |
me an article, and maybe it was on someone's newsletter, but someone, one of our listeners 00:18:14.220 |
sent me an article that was, it was a professor at like a mid-tier U.S. college, and he was complaining. 00:18:21.320 |
He's like, look, I know I'm Gen X. I've been a professor for 30 years. I know when you complain 00:18:26.860 |
about students, people are like, oh, there you go again, telling kids to get off your lawn. But he's 00:18:30.280 |
like, this is important, and I need to say it. And he kind of goes through, like, here's what I've seen 00:18:33.580 |
change in, you know, these students at the college where he teaches. And he starts by saying they're 00:18:39.860 |
functionally illiterate, right? And he's like, they don't read. They don't read. They don't do any of the 00:18:44.800 |
assignments. They lie about it. I don't think they could read. I think if I gave them just like a 00:18:48.640 |
Richard Powers book, like an award-winning book, but not like Finnegan's Wake, but like a book that 00:18:53.720 |
like it's out right now, or like Colson Whitehead, they couldn't read and understand it, right? Like he's 00:18:57.580 |
really worried about it. And when you get to the bottom, and he's like, they can't write, and the 00:19:01.380 |
punctuation and capitalization is all over the place, and, you know, there's no form. And when you get to 00:19:06.160 |
the bottom, you're like, okay, what's going on here? He's like, well, it's the phone, stupid. He's 00:19:09.300 |
like, it's the phone. They're completely addicted to the phones, and this is like a huge problem. 00:19:14.020 |
So I don't know if you're not at like an elite school like Georgetown, where I don't know. The 00:19:20.560 |
kids do read, and they're very smart, and they're very locked in, and they're like more professional 00:19:24.060 |
than I am. Like they're great students. But if you're not at like one of these top 20 type schools, 00:19:27.980 |
college is no longer necessarily going to be for sure like, yeah, you're going to get your four years 00:19:33.380 |
of cognitive training. That's what I think college used to be. Like all this reading and writing 00:19:36.920 |
and discussion, dialectical discussion in classrooms and with other students, you come 00:19:41.920 |
out smarter than when you started. Not just because you took in literal information, but because you 00:19:46.460 |
trained your brain. And then the knowledge economy is like, great, these people, we have smarted them 00:19:51.120 |
up. And now like they can handle like the complexity of what they're going to have and like the knowledge 00:19:55.340 |
economy. I don't know if it's necessarily the case at all schools anymore. So you might have to, 00:19:59.740 |
you know, we have to think more about training, I guess, on your own. 00:20:04.180 |
Are you a big fan of Succession? Uh, that's a great show. Yeah. Yeah. I watch Succession. 00:20:10.080 |
We're watching all the shows. There's like a lot of shows right now. Yeah. Right now because White 00:20:15.060 |
Lotus is just ending, but it's still going. Pit is great. The Pit. I've seen a couple of those. 00:20:20.840 |
Yeah. I'm a big ER guy. So Noah Wiley. Um, it's like, we've been liking The Pit and then there's this 00:20:26.500 |
new show on, oh, Severance. Like we, I realized like, I guess I have to watch that given what I write 00:20:31.580 |
about. Yeah. So that's going on and there's a new season of that. And then there's this new show. 00:20:36.560 |
Seth Rogen has a show on Apple TV called The Studio. I'll have to check that out. He's a studio. He gets 00:20:43.260 |
put in charge of a movie studio. It looks fantastic. So there's like all sorts of shows out here right 00:20:46.580 |
now. Did you read the article in the New Yorker a couple of weeks ago about the White Lotus director? 00:20:51.820 |
Yeah. Uh, Sean White, not Sean, Sean something. It's white. It is white. Yeah. Yeah. White Lotus, 00:20:57.740 |
white. Um, yeah. And there was also like a lot of video. He did a lot of press and there's like a lot 00:21:01.960 |
of videos going around about his creative technique. Yeah. He got really into that. Um, so yes, I do 00:21:09.540 |
watch those shows. Nice. Succession. That was a good show. Have you ever met? Occasionally I come 00:21:14.820 |
across people who are just very rich. Um, it's interesting. It's like, okay, someone that's spot 00:21:20.540 |
on. Like they spent a lot of time in that show. Only like really rich people would know this. Like 00:21:24.860 |
there's, it's this like norm core stuff. There's like a, like the nondescript baseball hat, right? 00:21:30.160 |
To like strong Jeremy Strong will wear it. It'd be like a specific, like $500 baseball hat. Yeah. 00:21:35.780 |
It's a lot of like, it's not supposed to be flashy. Um, but people who know, know that's like a super 00:21:42.440 |
expensive hat. And supposedly the other things, because they had consultants who were like, no, 00:21:46.380 |
this is how you should actually act. Right. If you're really that rich. And the other thing, 00:21:50.400 |
supposedly they told them early on is don't duck when you get off and on the helicopter. Because 00:21:56.100 |
if you were that rich, you'd be doing so much private helicopter flying. You would know that 00:22:00.280 |
like, it's too tall. It's not going to hit your head and you want it bought. Like ducking is what 00:22:03.840 |
you do if you've never really been on a helicopter before. Neither here nor there. Um, all right. 00:22:10.240 |
So we've got some questions coming up, but first let's hear about a sponsor. I want to 00:22:16.500 |
talk about our friends at notion because this is a product that, uh, I am a huge fan of notion 00:22:20.920 |
combines your notes, docs, and projects into one space that simple and beautifully designed. 00:22:25.180 |
It allows you to easily build basically custom information, software tools without having to 00:22:31.940 |
be the type of person who knows how to build, you know, a custom software product that you're 00:22:36.860 |
just, you're, you're setting things up very easily. This is whether you're talking about, 00:22:41.200 |
you know, an individual setting up a somewhat more advanced system for keeping track of all the 00:22:46.040 |
things you have to do. I mean, all, I get so many messages from leaders, readers who have taken my 00:22:50.440 |
ideas into custom notion systems, or you're like a really large company that's building like major 00:22:55.200 |
information tracking systems. Notion lets you do this really quickly. Uh, I mean, you, 00:22:59.740 |
you've integrated just, you've had your share of time working with it because our ad agency had that 00:23:04.000 |
really cool system they built out of notion that made stuff a lot easier. Yeah. And you could see 00:23:08.720 |
all the different pieces of information that were applicable to the podcast and it would give it in 00:23:13.700 |
different views. That was the cool part. So it was like, Hey, what ad reads are happening on this 00:23:17.120 |
day? Oh, let me click on that advertiser. Show me all of the ad reads we're doing in the next six 00:23:22.720 |
months from that advertiser. All I remember is we were doing something before that that was much worse. 00:23:27.300 |
and I don't remember what it was. Weren't they just like sending us spreadsheets? Yeah. It was all 00:23:31.040 |
tracking like a, it would send us like a, yeah, like a spreadsheet as an email. I don't know. There'd 00:23:35.420 |
be like a, a guy with an old timey newsboy cap would come up with a telegraph. Western union, 00:23:41.360 |
Western union, just then you got a Udacity, Udacity ad next episode. Stop. Make sure you mention 00:23:50.920 |
whatever. Stop. That's how we were doing it before. Uh, and it made it much easier. So I love notion 00:23:56.440 |
because I love smart information systems and they're, they're the best in the business they 00:24:00.620 |
have. Here's what's cool though. What they've done recently, which I just want to underscore quickly 00:24:04.660 |
is they have integrated AI to make your systems like immediately much smarter. The fully integrated 00:24:12.140 |
notion AI helps you work faster, write better, and think bigger doing tasks that would normally take 00:24:15.980 |
you hours in just seconds. So there's a lot of ways the AI now is seamless. Um, searching for 00:24:21.740 |
certain information using AI, you can do like a really intuitive, simple search. It just, it can go 00:24:26.660 |
through your information and find what you're looking for. When you're actually entering information into 00:24:30.720 |
the system, it's right there to help you write faster, get a first draft or brainstorm or polish up notes. 00:24:36.760 |
Right. So you've, you've, uh, this is a big one for me. Sometimes I take notes on the computer when I'm 00:24:40.960 |
doing an interview and it's very sloppy. And if you're throwing them in a notion system, the notion AI can 00:24:44.280 |
clean it up and fix the sort of obvious mistakes. Um, so it's cool. So they're, they're continuing 00:24:49.620 |
a product. I already love is using AI in a really cool way. Uh, there's a reason why notion is used 00:24:54.260 |
by over half of fortune 500 companies because teams that use notion, send less email, cancel more 00:25:00.160 |
meanings, save time searching for work and reduce spending on tools. Notion is my type of software. 00:25:06.580 |
So try notion for free. When you go to notion.com slash Cal, that's all lowercase letters, notion.com 00:25:12.680 |
slash Cal to try the powerful, easy to use notion AI today. When you use our link, you're supporting 00:25:18.020 |
our show. Go to notion.com slash Cal. I also want to talk about a new sponsor Kensta Insta with a K 00:25:26.820 |
something I've spent a lot of my adult life doing is dealing with WordPress. Every website that I ever 00:25:35.720 |
have going back to the original study hacks blog has been loaded on some sort of WordPress host. 00:25:41.080 |
And it really has been a major time suck. It is difficult to get good WordPress hosting. 00:25:47.220 |
It'll be unintuitive controls. It'll be security flaws that aren't updated. And above all else, 00:25:53.400 |
these installations tend to get really slow. This is where Kensta enters the scene. It does WordPress 00:26:00.340 |
hosting, right? They focus, focus, focus on speed, speed, speed. We're going to keep our installation 00:26:06.900 |
fast so that your site actually loads, which again, if you're doing WordPress hosting, this becomes a real 00:26:13.280 |
issue. They're great at it. They also have enterprise level security. You're not worried about, oh, we're 00:26:19.140 |
out of date on the latest security fix and hackers are going to get in there and mess with our site. You can 00:26:22.980 |
rest easy if you're using Kensta. Fantastic dashboard. This makes a big difference. A lot of dashboards for 00:26:30.080 |
these WordPress hosts. And again, Jesse, you don't have to see a lot of these because you don't do as 00:26:33.200 |
much of the web stuff, but a lot of them look like they were basically out of the movie war games 00:26:38.480 |
starring Matthew Brodwick, you know, where he's like typing into the, the, the green text, like goes 00:26:43.580 |
across and the Kensta interface is fantastic. Perhaps like that's my second favorite thing about 00:26:51.040 |
them. My first is the speed. Second favorite thing, real support. 24 hours a day, seven days a week, 00:26:55.920 |
365 days a year. You get a real human on the line when you have an issue, right? So if you've dealt 00:27:01.200 |
with WordPress, all of these pain points should sound familiar. And then I can point you with confidence 00:27:06.440 |
towards the solution you should use for your WordPress hosting, which is Kensta. I wish I had known 00:27:12.020 |
about them. God, it's been 20 years. I think I launched calnewport.com on an old WordPress 00:27:18.080 |
installation in 2006. 20 years. I wish I knew about Kensta 20 years ago. Sure. They didn't exist 20 00:27:24.500 |
years ago, but that's no excuse. It would have made my life a lot easier. So if you're tired of being 00:27:29.020 |
your own hosting support team, switch to Kensta and get your first month free and don't worry about the 00:27:34.260 |
move. They'll handle the migration for you. No tech expertise required. Just visit kensta.com 00:27:41.440 |
slash deep questions to get started. That's K-I-N-S-T-A.com slash deep questions. All right, 00:27:49.820 |
Jesse, let's move on to some questions. All right. First questions from Anna. I envision a future where 00:27:56.800 |
I live close to family, grow my own food, work from home, and have time for painting. However, 00:28:01.500 |
job opportunities in my animation industry are dwindling, especially remote work. Should I change 00:28:07.280 |
careers to achieve my ideal lifestyle? I feel like it's our second animation question in a few weeks. 00:28:13.120 |
It's interesting. I wonder if that's quite a coincidence. I'm reading a book on another Disney 00:28:18.160 |
book right now, and it's talking about the period where they get CalArts up and running. So this was 00:28:23.120 |
basically like the training camp that Disney helped fund to help generate new animators. So I know more 00:28:30.860 |
about animation than I probably should. All right, but Anna, this is not about animation when you're 00:28:34.880 |
asking a question about lifestyle-centric planning. So first of all, I love the approach you're taking 00:28:38.620 |
here. This is classic LCP. You have an ideal lifestyle vision in mind, and now you're working 00:28:44.180 |
backwards from that vision to saying, how can I actually make this possible? So when you're working 00:28:49.000 |
backwards from a lifestyle vision, the things that matter typically is going to fall into the 00:28:53.360 |
categories of obstacles and opportunities. What are the obstacles in the way for me getting where I am 00:28:58.580 |
right now to something closer to that ideal lifestyle? The opportunities are, what can I 00:29:03.400 |
leverage? What type of like unique setup? This could be like career capital or family connection or 00:29:08.400 |
whatever. What opportunities do I have to leverage as well? So you want to systematically work at reducing 00:29:13.680 |
or bypassing obstacles while taking advantage of everything that you have, all the opportunities you 00:29:18.760 |
have as well. So that's probably the right thing to do here. The key obstacle in this analysis is you're 00:29:24.600 |
seeing the particular types of jobs you have in the animation industry is not remote work friendly, 00:29:29.860 |
and your lifestyle vision, when you work backwards from it, remote work really makes that work. 00:29:34.600 |
Right? Because if you can be in a lower cost of living place, you could be at home and growing food 00:29:40.360 |
and be near family and maybe not need to work as many hours because where you're living is less 00:29:44.340 |
expensive. If instead of being, you know, near Burbank or Disney is, you're instead, you know, 00:29:49.300 |
where you grew up in Indiana, it's going to be cheaper to live. Like remote work seems to be at the core 00:29:53.480 |
of it. So there's a key obstacle here. If your current job trajectory isn't making that possible. 00:29:58.600 |
So now you know what you're doing. Oh, I am systematically exploring and looking for insight. 00:30:03.500 |
It might not be obvious, but I'm looking for a moment of insight about how do I reduce that 00:30:06.940 |
obstacle? That's probably going to come from leveraging the opportunity of the various skills 00:30:11.020 |
you have and looking for adjacent applications. So like, sure, maybe if I want to be in like feature 00:30:15.440 |
motion picture animation, there's like three options. They're all based in California and they're 00:30:19.140 |
cutting back and there's not remote work options, but maybe there's an adjacent step here. Actually, 00:30:23.980 |
this graphic skills I have could be used, maybe not in like feature film, whatever animation, but it 00:30:30.800 |
could be used in whatever game development, object design, or in certain types of advertising or 00:30:36.480 |
something. There's there. I find that adjacent field where it's very useful. And now I have taken 00:30:43.080 |
an opportunity and gone around an obstacle. So it's not easy, but it's specific. That's what's 00:30:48.380 |
key about LCP is that it's very specific what you're doing here. I have this obstacle. I'm enumerating 00:30:53.860 |
my opportunities. The only piece of like semi-advanced advice I would give is that if you're stuck, 00:30:59.700 |
A, keep searching for ideas and inspiration, talk to people, read things, listen to things. You're 00:31:04.660 |
looking for that path. It could be narrow, but you're looking for the path that gets you closer. 00:31:09.340 |
Increasing the opportunity side is often a low hanging fruit that can make a big difference. 00:31:13.680 |
And that might mean learning a new skill, a relevant skill. I know how to do this. What if I learned how 00:31:19.200 |
to do this, which is someone who already knows how to do this first thing would have an easier time 00:31:23.140 |
learning it, like looking for these adjacent skills that if you learned would open up new opportunities. 00:31:27.340 |
So you're kind of playing a game here, but you're thinking about it the right way. You have a lifestyle 00:31:32.940 |
in mind. You're trying to work backwards from it. So, so you're doing the right thing and don't be 00:31:36.420 |
frustrated if it takes a while to finally find that path. Also, I'll throw in one more coda. 00:31:41.800 |
You know, you know, you know what you want your, your ideal lifestyle. You can isolate some of the 00:31:47.640 |
general properties here, like a little bit more of a slowness, a connection back to nature when you're 00:31:52.080 |
growing your food connection to community. Keep in mind, even while you're looking for a larger 00:31:57.540 |
reconfiguration to get to your full ideal lifestyle, you could be making changes to your life as it exists 00:32:03.480 |
now to have more of those properties. So even if you're not getting the full right off the bat, 00:32:09.600 |
I'm on the, the, the land with the farm near my family's house in Indiana or whatever, if you're not 00:32:14.500 |
there yet, use this lifestyle to help isolate. These things are important to me and get that into your 00:32:20.460 |
life right away. And maybe now you have a plot in a community garden and you've adjusted your hours. 00:32:25.520 |
It's still the in-person job, but it's, it's a four day a week job. Now you have a little bit more 00:32:29.180 |
slowness going into it. Uh, you're more community involved where you live, even if you can't be 00:32:34.320 |
closer to family, play with these things you're identifying as important and start getting them 00:32:40.140 |
in your life. Now, don't let perfect be the enemy of better. And that's like a key thing with LCP is 00:32:47.720 |
that it's not just until I'm living on the houseboat, I'm in trouble. It's what is it I'm identifying as 00:32:53.080 |
important? Let's keep adding this to my life. And actually that might open up other ideas for how to 00:32:56.760 |
get there. So anyways, uh, good luck with that, but thanks for the good LCP example. 00:33:03.400 |
Next question is from Ruben. I'm currently learning how to code during my free time to get a job as a 00:33:08.740 |
software developer. I believe it would be an ideal career path towards achieving my life. I want, 00:33:13.880 |
is this a bad idea to leave part of my career capital with my BA in psychology? I'm 24 years old. 00:33:19.940 |
Well, I mean, the good news is you're 24 years old, so you can't really have much career capital 00:33:24.680 |
anyways. So like you're, you're at the beginning. You have a lot less opportunity, loss opportunity 00:33:30.620 |
cost threats here. So yeah, this is not a bad time. I'm right out of college and I'm building out my 00:33:36.880 |
career capital portfolio. You're starting from low stores in all of your career capital category. So 00:33:42.120 |
I think it's completely fine. You could probably leverage the BA a little bit. Like here's something 00:33:46.360 |
I would say, here's a, an interesting piece of data that's been touted in the last few months. 00:33:51.480 |
We're starting to see some of the first places in the labor statistics. So not just anecdotally, 00:33:58.760 |
but some of the first places in the labor statistics where we think we're seeing the impact of AI 00:34:03.660 |
is in a class of people, uh, technical jobs called programming, which is different than software 00:34:10.180 |
development. So programming jobs post chat GPT have been suddenly declining. Software development 00:34:18.860 |
jobs have not at all. They're, they're basically the same, if not like up a little bit, you have to 00:34:23.380 |
kind of go into the weeds. What's the difference? Well, programming jobs, I guess there's, there's a, 00:34:27.620 |
a very specific type of computer programming related technical position where it's, you're given 00:34:34.100 |
very specific things to implement often with like simpler programming languages, like, okay, go add 00:34:39.360 |
this feature now. Can you go onto the side and change that? So it's like very, you're doing very 00:34:43.480 |
little planning or creativity. You're not watching the lifestyle of a product. You're just sort of 00:34:46.880 |
implementing. Chat GPT can produce that type of simpler code pretty well. And so now, uh, the person 00:34:53.640 |
who was tasking the programmer with that work could just sort of use AI help and might as well do it 00:34:58.840 |
themselves. Software dev is more complicated. You're actually, uh, it's part of the lifestyle of the, 00:35:03.460 |
the life cycle of the product you're involved in. It's creative. It's figuring out what the program 00:35:07.280 |
needs just as much as what code it has to be in there. You might be able to leverage your BA 00:35:11.400 |
in psychology to say, look, I, I speak normal person speak. I can like think about the consumer 00:35:17.760 |
and what they want in their product. I understand, uh, I understand other people. I can manage other 00:35:24.760 |
people. There is, I think, potentially a career capital advantage to be someone who's technically 00:35:29.280 |
savvy, who also can communicate like a normal human being. And I'm sort of throwing 00:35:33.280 |
my fellow computer scientists under the bus here, but let's just say, you know, we're not 00:35:37.380 |
all, uh, expert communicators. Like this is not necessarily, we're not necessarily, uh, it's 00:35:44.080 |
not a bunch of like John F. Kennedy's when you're at a computer science conference, right? Uh, it's, 00:35:48.840 |
it's, we, uh, I, I do pretty well, but you know, it's a lot of leet speak, right? It's a lot 00:35:54.660 |
of like weird sort of, uh, technical terms and jargon. So you might have an advantage with your, 00:36:00.860 |
your BA in psychology. My, my poor wife used to come to some of these MIT events, like holiday 00:36:07.920 |
parties or whatever. And what's going on? Don't take me to these things anymore. Like why half the 00:36:14.360 |
people here can't communicate at the other half are very normal. It's just, you get a higher 00:36:17.960 |
percentage than normal of people who are, you know? Yeah. Like if you see early footage of like Bill 00:36:23.800 |
Gates, it's just like a lot of him, like rocking and writing a simpler code in his head. 00:36:28.960 |
Yeah. Yeah. It's not exactly, you wouldn't want a, a, a theoretical computer science student at MIT, 00:36:35.020 |
like on a, uh, like a radio talk show host or something. Yeah. Like it wouldn't be a popular 00:36:39.660 |
show. You wouldn't want to replace mad dog with like one of my, my commenters. And they'd be like 00:36:44.680 |
the, uh, performance, the Red Sox was highly illogical. His numbers against the lefties were, 00:36:51.360 |
the performance were obviously lower. And now I'm going to recap in great detail 00:36:57.340 |
information from the latest Star Wars or Star Trek. I don't know. We are going to go through 00:37:04.160 |
the 10 points that make it clear that Star Trek Deep Space Nine is superior to the next generation. 00:37:08.120 |
I believe this is obvious. All right. I'm sorry. Actually, most of the computer science I know 00:37:11.500 |
are cool, but programmers aren't always. So maybe you do have an advantage there. So take advantage of 00:37:17.080 |
that. All right. Who do we got next? Next question from Bill. Do you have examples of 00:37:21.360 |
truly fair performance evaluation systems? How should merit be rewarded in a revenue constrained 00:37:27.020 |
environment? I work at a national nonprofit. Um, I think what you should do, and I think the, 00:37:33.020 |
the lawyers will back me up is you want to, uh, sort people into different like demographic groups 00:37:39.160 |
and then just pay your favored groups more is probably like, so if you like the country they're 00:37:43.880 |
from, pay them more than no. Aren't we allowed to say that now? Is that, aren't we all just like, 00:37:47.840 |
whatever, anything goes, I don't want my funding cut. So yeah, just no DEI. Um, no, there is, 00:37:57.000 |
so it is tricky, uh, performance-based review there. I wrote a New Yorker piece back in the pandemic 00:38:02.540 |
from, there is a system called ROE, R-O-W-E, results-oriented workplace environment. And there's 00:38:09.880 |
a book about this that's called something like work, like work doesn't have to suck or something 00:38:14.940 |
like that. I mean, there's only so many business books that have like that in the title, so you'll 00:38:18.840 |
find it. It's fascinating. It is a attempt to rebuild an office. If you apply the ROE methodology, 00:38:25.840 |
you rebuild your office entirely around results is all that matters. And there's no expectations on 00:38:33.700 |
you. Like, Hey, when you want to work, you work. I don't care. There's no vacation. There's no hour. 00:38:36.820 |
You don't have to be here any particular hours. You, you essentially sort of like negotiate, okay, 00:38:41.260 |
I will work on this now. Here's like the measure of success. Here's how we're going to communicate 00:38:45.560 |
about it to make it done. And you kind of have like this portfolio of projects that you do and 00:38:48.860 |
you're held accountable to it. It really does work pretty well when it's implemented. But the point 00:38:54.100 |
of this article is it's very hard to implement it. And honestly, the big issue they have, and I 00:38:59.780 |
interviewed the, the, one of the creators of it and I read the book and talked to a CEO of a company 00:39:03.980 |
that uses this methodology. The, one of the biggest issues they have is not figured out how to measure 00:39:08.340 |
performance. It actually gets pretty clear. Like, okay, you're doing this. All right. We know what 00:39:12.940 |
success means here, right? Or like, are you, we can just look at like in a ROE environment, I could just 00:39:16.760 |
be like, these are the three things you did last quarter. Does this seem valuable or not? Like there's no 00:39:20.560 |
hiding. The hard part actually is trying to eliminate the last vestiges of pseudo productivity 00:39:27.300 |
thinking in the office, right? Pseudo productivity being my term for my book, slow productivity for 00:39:32.800 |
using just visible activity as a proxy for useful effort. There is a huge amount of the ROE, I guess 00:39:39.440 |
you would call it like a application effort, trying to conversion efforts, is trying to get rid of 00:39:44.480 |
pseudo productivity speak. And they call it sludge in the book. But it's just this idea of people are 00:39:50.460 |
very, because of the way we built knowledge work, are just very uncomfortable about like, why aren't 00:39:54.340 |
you here or I can see you? Why didn't you answer an email right away? It is a difficult mindset to get 00:40:00.480 |
away from perform, performing activity to generating measurable results. And it's a hard shift. They also 00:40:08.320 |
found that they won't naturally, the people who created this methodology say it all goes well. But 00:40:14.220 |
when you talk to people who have actually implemented it, they will tell you, you're going to lose a lot 00:40:19.300 |
of people. Not everyone can do this. The CEO I interviewed for this New Yorker piece lost like 00:40:24.040 |
20% of his non-managerial staff and about 20% of the managerial staff. The non-managerial staff, 00:40:28.960 |
they're like, I am actually not able to perform like this. I could be busy. My TPS reports are handed 00:40:36.340 |
in. I answer emails. I'm always jumping on calls or whatever. But if you just say, forget all that, 00:40:39.740 |
just tell me what you did this last month. Some people like 20% are like, I actually am not 00:40:44.260 |
comfortable just focusing and doing hard things. And they lost 20% of the managers because they're 00:40:48.460 |
like, I don't know how to do this if I'm not just like being able to demand quick responses and just 00:40:54.300 |
have people jump. So it's possible, but it's hard. Read my New Yorker piece. That would have been from 00:40:59.120 |
like 2021. And I forgot what it was called, but just search like Cal Newport, R-O-W-E, you'll find it. 00:41:05.340 |
And in that book, Work Doesn't Have to Suck or something like that. We'll talk about that 00:41:09.800 |
philosophy. All right. Who we got next? Next question is from Ben. In episode 345, you talked 00:41:16.380 |
about the value of more niche online communities. What do these small alternatives look like and how 00:41:21.600 |
do people find and engage in them if they're not hosted on these global platforms? There's a few 00:41:26.700 |
forms. So these more like niche online discussion communities, which I think by the way, is like 00:41:31.340 |
the right way to extract the community building value of the internet. A few places to find these. 00:41:36.020 |
Bulletin boards are a big one. Oftentimes now the technology is not going to be a straight up 00:41:42.540 |
bulletin board technology. It might be like at talknats.com, the example I often use in my writing 00:41:48.100 |
of a Washington Nationals-based community. What they actually do, and a lot of sports sites do this, 00:41:53.360 |
is they'll post like a blog post and the conversation is happening in the comments of the blog post. 00:41:59.260 |
Straight up bulletin boards are fine. I actually look at Reddit threads as in this, even though 00:42:03.460 |
Reddit is a large global platform, the individual Reddit threads are not algorithmically sorted and 00:42:09.860 |
they are managed by the community themselves. Substack or other newsletter-based comments has 00:42:16.440 |
become a big place. Like I get a substack from some writer I like. The comments underneath the 00:42:23.100 |
substack becomes like a community gathering place. A lot of people have Discord servers set up, 00:42:27.280 |
like you like a given show or a given type of whatever. Discord servers are big. You jump on 00:42:32.580 |
there and you can talk and it's like a relatively smaller group of people. So all of this is out 00:42:37.200 |
there. Podcast, so anything that's Patreon protected, there's discussion communities around it, whether 00:42:43.160 |
it's actually using Patreon or, you know, its own homegrown solution. So you'll see this with like, 00:42:49.280 |
I don't know, like if you're within like Sam Harris's walled garden, there's like places you can 00:42:53.240 |
chat with other people. So they're out there. They're out there. The key though is that it is 00:42:57.880 |
a self-selected group of people interested in the same topic. It is a manageable number of people who 00:43:03.300 |
are there. So you actually like know a lot about these people. Like you've interacted them with 00:43:07.620 |
enough. You recognize them and have a bit of a shared history with them, even if you don't know who 00:43:11.200 |
they are like really in the physical real world. And it's not algorithmically curated. It's just the 00:43:16.160 |
content is there, probably just chronological. And the standards are set by the community. 00:43:20.860 |
that's what you're looking for, for, I think like a richer online engagement. 00:43:26.520 |
Steven asks, how can partners support each other in building a deep life? 00:43:31.460 |
I think the key here is that the life is one that you are coming up with together. 00:43:37.280 |
So when you have your sort of vision of an ideal lifestyle that you're, you're sort of working 00:43:42.160 |
towards, you need a shared vision first. Like this is where our family wants to be. This is what our 00:43:49.520 |
family life should be like in five years and in 10 years. Like if you have kids, for example, 00:43:53.220 |
you should break this up by the sort of like pre-elementary school, elementary, elementary 00:43:58.340 |
period, the sort of middle school, high school period, and then sort of the post high school 00:44:02.340 |
college period. What is our family life? What do we want? Like where we live, the rhythm, what it's 00:44:06.700 |
like. Is there someone home after school? What's our weekends like? You want to have that figured 00:44:10.980 |
out together. Then underneath that, you can have your individual. Okay. Now here's the things that are 00:44:16.160 |
important to me. Like what I want to do with my physical fitness, for example, might be different 00:44:20.560 |
than like what my wife wants to do or my interest in my hobby interest are for sure, you know, 00:44:25.440 |
different. So start with the shared thing. That's priority one. We're on the same page. We're planning 00:44:30.660 |
with that. And then you can have your own sort of buckets you're working on. And there, I think it's 00:44:35.060 |
helpful to some degree to kind of like, Hey, here's what I'm working on. What are you working on? What's 00:44:38.120 |
important to you? What's important to me? How can I help you with what you're doing? How can you help me 00:44:42.120 |
with what I'm doing? Uh, and that, that works out as well, but make sure that there is a shared 00:44:46.400 |
component like this stuff. If you're in a partnership with someone, that partnership has to extend to your 00:44:52.160 |
vision of what the deep life is. All right. Let's see here. I'm going to suggest we, cause I want to, 00:45:01.400 |
I'm going to, we're going to skip either the call or the case study just so we can roll into the tech 00:45:04.460 |
corner with time to go today. Okay. Which one let's, uh, I'm going to skip the call. Okay. And I'm 00:45:09.020 |
going to jump right ahead to the case study. This is where we find out that the call we skipped was, 00:45:13.580 |
I was going to say from someone famous, but I couldn't, I don't know like what name I'm not 00:45:17.600 |
up with like the hip references. Like, you know how old fashioned I am. Mr. Beast. Yeah. Well, 00:45:22.700 |
let's see. My first instinct was to be like King Charles. So how to touch. Yeah. Mr. Beast called in 00:45:28.680 |
and we missed it. It's like, I need you on my channel stat. All right. Our case study today comes from 00:45:33.200 |
Chris. Chris says, I was recently approached by a firm looking to fill a position that I was uniquely 00:45:38.560 |
qualified for. It would have been a dramatic step up in pay with a profit sharing component that may 00:45:43.540 |
have doubled or even tripled my current income in a good year. I received an offer for the role and 00:45:48.280 |
part of me wanted to just go for it because the income increase was so dramatic, but I decided to 00:45:52.340 |
put it through a more holistic lifestyle centric evaluation. I VBLCCP'd it. Here's what I realized 00:45:59.500 |
after further reflecting. This role was something of a grand goal trap. I was telling myself that if I 00:46:05.460 |
could just get to this executive level, I would have quote, made it end quote, and everything else 00:46:09.360 |
would have fallen into place. But in reality, the position would have required to move to a higher 00:46:13.940 |
cost of living area and from a fully remote role to a three days per week in office plus travel 00:46:18.800 |
requirements that seem to grow exponentially. As I really pressed the hiring team to clarify, 00:46:23.320 |
this would have substantially reduced the number of hours I'm able to spend with my three-year-old who I'm 00:46:27.740 |
committed to being present parent for and would have required a big renegotiation of the parenting 00:46:33.020 |
burden with my partner. It was also a high risk, high reward job, meaning what I would have been 00:46:37.180 |
more fearful of termination if I wasn't performing with complete excellence. I also did the work to 00:46:44.500 |
process what enough meant for the income requirements, my lifestyle vision. And I realized that while I do 00:46:49.880 |
need a bit more coming in annually, I'm not that far off. Instead of taking this dramatic leap up in both 00:46:55.440 |
income and work hours, I'm looking at other options like a more modest move to a full-time role 00:46:59.820 |
that increases pay but maintains most of my current balance. 00:47:03.020 |
That there, Jesse, is VBLCCP in action, values-based lifestyle-centric career plan. There you go. When 00:47:10.160 |
you know what matters to you, what you want your life to look like, what's happening, like your job 00:47:15.120 |
becomes just a tool you have among other tools to craft this life. And that's a fantastic decision. 00:47:21.160 |
I talk about that in my book, So Good They Can't Ignore You. That is one of the control traps, 00:47:25.980 |
that just as you get good enough to have some leverage over your life is exactly when people 00:47:30.760 |
are going to come to you and give you shiny opportunities to take your autonomy away. 00:47:34.440 |
So just as you're like, I'm pretty good at this, I'm making good money and have flexibility, 00:47:38.920 |
that's exactly when they say, hey, we want to triple your salary. But then in the quiet voice, 00:47:43.560 |
but you're never going to see your family again, you're going to travel all the time and be stressed. 00:47:47.160 |
But hey, you were going to triple your salary, right? That's the second control trap in action. 00:47:51.220 |
Lifestyle-centric career planning helps protect you against it. 00:47:53.840 |
Like I think, as Chris, is going to probably be substantially more happy on average the next 10 00:48:02.780 |
years with the decision he made than if he had just gone for the bigger job. Because why are you 00:48:08.000 |
doing the bigger job? Like it's all about what it's serving or like the work itself. So the work 00:48:11.980 |
itself wasn't going to be more interesting to him and the stress was higher and it was going to get in 00:48:16.000 |
the way of multiple different goals, other things that were important. So now he's just tweaking what 00:48:21.140 |
he has to get a little bit more money and then they're good. So I think that's great. I just want 00:48:25.760 |
to flag real quick how he talked about the processing what enough meant. We've talked about that in a prior 00:48:31.540 |
show, but it's just worth underscoring. That's this idea of you make sure you know the numbers. 00:48:37.640 |
Like how much money do we need for like this type of lifestyle we're looking for? Having the specific 00:48:42.940 |
numbers matter as opposed to just more. If you just say I need more money, you might take the triple 00:48:48.040 |
income job and be miserable where he crunched the numbers and said, we're close. We need to just fill 00:48:53.560 |
in this. Now let me be really, really careful in figuring out how to fill in that in a way that's 00:48:57.280 |
not going to get rid of all these other goals. So anyway, it's fantastic. Lifestyle-centric career 00:49:00.440 |
planning in action. All right. We got a tech corner coming up, but first let's hear from another 00:49:07.920 |
sponsor. We have a new sponsor this week. This is a company I've used for. I know this company well, 00:49:14.720 |
so I'm happy to have them as a sponsor. That is the online course company Udacity. 00:49:26.120 |
Udacity. I think I should get a bell for pronouncing it. It's one of these things I've 00:49:30.420 |
used Udacity courses for a long time, but you don't have an occasion to say them out loud 00:49:35.140 |
until you're actually talking. So Udacity is the site you go to if you want to learn skills that 00:49:42.180 |
are going to command high salaries. It's the online learning platform that has courses on all sorts 00:49:46.260 |
of topics, but including AI, data, programming, and more. I have attended or purchased multiple 00:49:54.280 |
Udacity courses. The most recent one I was doing with my son, it was a computer game programming 00:49:58.760 |
course. I did another one with my other son on like Tinkercad based 3D design, 00:50:03.620 |
like intro to 3D design. It's fantastic. It's a really easy interface. You can purchase the course 00:50:09.280 |
easily. Some are free. They're video based, keeps track of where you are. Knowledge is power. We talked 00:50:15.180 |
about this in the very first segment of the show. Like how do you get smarter? Online courses are a big 00:50:19.580 |
way. I mean, you've been messing around with this. You just signed up for one, right? 00:50:23.100 |
Udacity course. Yeah. I signed up for a digital marketing class. Oh, excellent. Yeah. And I'm 00:50:27.400 |
teaching one now on Udacity about how to do a killer French accent. So it's, you know, it's been a real 00:50:32.940 |
hit. Have your pipe. On a related note, someone just handed me a note and I am no longer welcome 00:50:38.760 |
in France. Oh, okay. Well, that's good to know. They just sent me a note. So I guess that got me in 00:50:43.840 |
trouble. There's tons of options for learning tech skills, but only Udacity is consistently ranked as 00:50:49.560 |
a top skill development platform because unlike others, it actually works. I can attest to this 00:50:54.800 |
with real world projects and human experts that grade your work. You'll truly get the skills you 00:50:58.720 |
need. Yeah. When my son and I were doing the game programming, I mean, you're building the games. 00:51:02.480 |
Like you're doing, you're doing the work right there. When you have a certification from Udacity, 00:51:08.640 |
recruiters and employers take notice. So for a better job, better salary and better skills, 00:51:13.480 |
check out Udacity today. I saw a YouTuber, a DIY YouTuber. I was watching with my kids the other 00:51:19.900 |
day and he said his habit, he's a brilliant guy. He said his habit is to spend the first 30 minutes 00:51:26.300 |
every day on an online course. He's like just making myself smarter before I move on to like the rest of 00:51:31.280 |
my day. I love that idea. So maybe do that. Sign up for Udacity. Always have a course you're working on 00:51:35.720 |
first 30 minutes of your day, get smarter, get your brain working. So the tech field has always evolved 00:51:41.280 |
and you should be too. Always evolving and you should be too. You can try Udacity risk-free for 00:51:47.000 |
seven days. Head to udacity.com slash deep and use the code deep to get 40% off your order. Once again, 00:51:53.900 |
that's udacity.com backslash deep for 40% off and make sure you use my promo code deep so they know I 00:52:01.240 |
sent you. I also want to talk about our friends at ZocDoc. I went to the dentist this morning. 00:52:06.100 |
All right. This true story. I went to the dentist this morning that we are recording this. 00:52:10.520 |
How did I originally find this dentist? It was like years ago, I was canvassing friends and someone 00:52:17.420 |
said, Oh, there's this new dentist in town. And I, I got in and I liked him and it's been great, 00:52:23.880 |
but the word got out and he's very popular. He's now booking seven months out. So like, 00:52:31.080 |
if you want to make an appointment with them, it's like seven or eight months before there's an 00:52:33.920 |
appointment. And so I was thinking today, as they were booking my next appointment, 00:52:37.200 |
like seven or eight months out. Uh, imagine if I was looking for a new dentist now, 00:52:41.500 |
what would I do? Like if I asked my friends and they're like, Oh, here's this one dentist and he 00:52:45.700 |
doesn't have any openings for months. Like that is what it's like right now. Often trying to find the 00:52:50.300 |
medical care you're working for. You get like one or two recommendations. And especially in a city, 00:52:54.680 |
when you say, I would like to actually come see a practitioner, like maybe, maybe at like some point 00:52:59.980 |
during like the current presidential administration, they just laugh like, Oh, you fool. You think the 00:53:07.320 |
dentist actually can see you. And then there's like a lot of laughter and you hear them call someone else 00:53:11.740 |
over and then they start laughing. Uh, and then like the doorbell rings and someone runs in and just 00:53:15.980 |
punches you in the stomach because they're, they're making fun of you that much. This is where ZocDoc enters 00:53:20.360 |
the scene. ZocDoc is a free app and website where you can search and compare high quality in-network 00:53:25.280 |
doctors and click to instantly book an appointment. You can search for, here's what I'm looking for. 00:53:30.920 |
And then see which of these take my insurance, which of these are actually booking appointments, 00:53:37.040 |
right? This is how you would have found my dentist a few years ago. Oh, he's actually looking to take on 00:53:41.640 |
new patients. Uh, okay, great. When is their next appointment available? Let me book it right now. 00:53:47.220 |
Let me do some of the paperwork online using the ZocDoc interface. Actually, my dentist uses ZocDoc. So it's, 00:53:51.820 |
so it's interesting. It makes it so much easier and so much less stressful and honestly humiliating to try to 00:53:57.780 |
find because no one's laughing in your face for daring to ask if they have actual openings. Uh, so this is the 00:54:03.440 |
way that you find medical care. Appointments made through ZocDoc also happen fast, typically within 24 to 72 00:54:08.860 |
hours of booking. You can even often score same day appointments. So stop putting off those doctor's 00:54:16.220 |
appointments and go to ZocDoc.com/deep to find an instantly book a top rated doctor today. That's Z-O-C-D-O-C.com/deep. 00:54:24.460 |
All right, Jesse, let's do our final segment. All right. Because I can't help myself. We're a tech 00:54:34.140 |
corner segment. We're going back to AI. There's so much to keep up with, with AI right now. And I want you 00:54:39.180 |
to be up to speed. All right. Here's what I've been rabbit holing on recently. 00:54:45.100 |
This idea that these new AI large language models like DeepSeq and the new O family of models from 00:54:52.220 |
OpenAI can quote unquote reason. So you give them like a math problem or something and they can sort 00:54:58.140 |
of work through the steps to get to an answer. They're doing much better on like math tests or 00:55:03.180 |
crossword puzzle tests or something like this, right? Like they, they are reasoning. I'm really 00:55:08.140 |
interested in this. All right. How did these things reason? Well, let's ask the companies. They're giving 00:55:14.060 |
incredibly vague language and a lot of journalists are just repeating this language. So like OpenAI, 00:55:19.580 |
when they talk about their new models, they say, uh, the new models, they, they slow down to think 00:55:25.500 |
more and that's how they're reasoning. And this is the way it's being reported in like mainstream 00:55:31.180 |
journalism. We're talking about, there was like just an article out in the times a couple of days 00:55:34.700 |
ago about how do these new models reason. And it's just saying like, yes, the new models, what they do 00:55:39.580 |
is they slow down. They slow down and they think, they think more. And so they can, they can produce 00:55:44.460 |
more reason answers. What does that mean? Right? Like Jesse, when you think about like a model, 00:55:49.740 |
like slowing down and thinking like, what does that mean? It feels like it's like an actual person. 00:55:52.780 |
That's like, what, taking a deep breath and like, I'm going to like take my time. Like what's slowing 00:55:57.900 |
down? Right. So I, you know, I called the source and, uh, talked about it. Like, Hey, 00:56:03.100 |
walk me through this tech, point me to the right papers. Here's what's as best we can tell is actually 00:56:08.940 |
going on. Okay. Uh, there was this idea that actually precedes chat GPT, right? This is an idea 00:56:15.420 |
from the academic community, studying language models, 2022, you get this paper that says, 00:56:19.500 |
if you want to increase the reasoning performance of language models, 00:56:24.460 |
you should use something called chain of thought prompting, C O T prompting. And what they were 00:56:30.780 |
showing in this paper is, uh, in the question, you're asking the language model, you basically 00:56:37.340 |
tell it, I want you to show your work. Don't just give me the answer. Like show me your steps. This is 00:56:43.500 |
what I'm looking for is walk me. I'll give you an example of it. See how I'm solving this problem here. 00:56:48.700 |
Step-by-step. Give me an answer to this new problem with similar style of step-by-step explanation. 00:56:54.460 |
And right. And what happened is this would give you better answers because language models, of course, 00:56:59.900 |
are there, they're trained to, um, expand text in predictable ways. They're trained on taking real text, 00:57:05.820 |
taking words or sequence of words out of them and having the language models, try to fill those in. 00:57:09.900 |
And the closer they get to real words there, like the, the more they get rewarded in their training. 00:57:13.820 |
So like what was happening is you would give a language model, like a logic problem. 00:57:18.620 |
You know, Jesse has this many apples and then someone comes along and takes this many apples. 00:57:22.540 |
And then the person who's taller gives you back this many apples. How many apples does Jesse have? 00:57:26.540 |
And what the language models would sometimes do is give an answer that sounds like what an answer 00:57:31.100 |
to that problem would look like. Like, yeah, I'm saying Jesse has some number of apples. Like that's a, 00:57:35.900 |
for our language models perspective, that looks like, like a reasonable response to this prompt. 00:57:41.660 |
It's talking about Jesse. It's talking about apples. There's a number, but that number could be arbitrary. 00:57:46.140 |
Chain of thought prompting says, no, no, you, you have to really ask the language model, 00:57:50.220 |
walk me step-by-step through your reasoning for how you get to the answer. And I'll give you an 00:57:53.820 |
example of me doing that in my prompt. Now you do something similar. When it walks through its reasoning 00:57:58.460 |
in the answer, it's more likely to get the correct answer because, um, as it fills out its reasoning, 00:58:03.820 |
it's actually accessing relevant circuits that do different parts of the reasoning. And it's actually more 00:58:09.660 |
likely, you're more likely to activate like the relevant circuits that can do the different 00:58:13.180 |
calculations and put them together in the right way. And you're more likely to get the right answer. 00:58:16.940 |
That's chain of thought prompting. Okay. So the issue is now we get, uh, chat bots and they're 00:58:23.660 |
getting popular. We really can't rely on prompt engineering as something that like the average 00:58:28.620 |
user is going to do, right? This is like the big, the big push with consumer facing large language model 00:58:35.260 |
tools is like, we can't expect the users to write these incredibly complicated prompts to get their 00:58:40.860 |
results. Researchers were doing this, but we can't expect the users to do this. So what seems to have 00:58:47.420 |
happened is they said, can we basically use reinforcement learning to kind of force one of these models to 00:58:54.380 |
always act as if it's being prompted to show its work? Right? So this is the key. And I'm not, 00:59:00.460 |
I'll just be incredibly brief on this, but like the, the, the key that got like the original chat GPT so 00:59:05.820 |
much notice is they took the first really large, large language model with was GPT-3, which could do 00:59:11.660 |
amazing stuff, but you really had to write incredibly careful prompts to get it. And the big innovation, 00:59:16.940 |
second innovation open AI had the first being just make the model large was this thing called, uh, 00:59:21.740 |
we think of it as like re a reinforcement learning based fine tuning, where they took this model 00:59:26.300 |
that had all this knowledge in it. And if asked correctly, could produce all sorts of stuff. 00:59:30.860 |
And then they started running it with a bunch of prompts. And every time it gave like a good type of answer, 00:59:36.780 |
they would zap it with a happy zap. And it was sort of like, Ooh, whatever weights we were using here, 00:59:41.340 |
let's make those stronger. And whenever it gave an answer that was, they didn't like, they would give it a, 00:59:45.100 |
a, an unhappy bad person's app. And it would be like, Ooh, let's kind of reduce some of these weights 00:59:49.980 |
that gave us this answer. And so for like chat GPT, they were saying, we're going to run, 00:59:54.620 |
we train GPT-3 and we're going to run a bunch of prompts. And when you answer this, like, uh, an 00:59:59.260 |
agent that you're chatting with a human, we're going to give you a happy zap. And then, so it got good 01:00:04.460 |
at answering questions without you having to write really careful prompts. I think what is happening, 01:00:09.020 |
according to my source is that they're just doing this now for chain of thought reasoning. 01:00:12.860 |
So they take one of these models, they'll, they train it up the old fashioned, you know, 01:00:16.700 |
with all the, all the texts on the internet and just like any of these other models. 01:00:20.060 |
And then as they start running through answers to it, giving it questions, 01:00:24.380 |
when it shows more of its work, they give it a happy zap. And when it just gives an answer, 01:00:29.340 |
they give it a negative zap. So this reinforcement learning fine tunes this big model, 01:00:33.340 |
like, okay, okay, okay. I, I get it. You want answers where I explain my work. 01:00:38.300 |
You could do this for anything, right? This is how they, um, 01:00:41.660 |
stop certain types of content from coming out of these language models. It's completely trained. 01:00:45.580 |
And then they, they give it some questions that might elicit responses that are like violent or 01:00:51.100 |
something they don't like. And every time it gives a bad answer, they zap it. And then like over time, 01:00:54.940 |
it's like, I'm not going to give those type of answers. 01:00:56.620 |
So this is what I believe the reasoning models have done is they take a standard big train model, 01:01:02.700 |
and then they do a reinforcement fine tuning step where they're saying, show your work, 01:01:06.620 |
show your work, show your work, give us big, long answers. Um, and when we know this since 2022, 01:01:12.060 |
pre-chat GPT, that when the chat bots, when the language models rather have to explain their work, 01:01:17.340 |
they're more likely to give accurate answers. All right. Now, part of what they're doing here, 01:01:22.300 |
in addition to that is these answers are so long, they're now hiding a bunch of it. It's like, 01:01:28.140 |
what's really happening is the, the chat bot is giving this like incredibly long 01:01:31.500 |
explanation for its answer, which helps the accuracy of it. But the explanation is so long 01:01:36.620 |
that if you're a user who asked a question, you're like, oh my God, I don't want to see all of this. 01:01:40.380 |
So they're, they're hiding a lot of that. And we know this because, uh, if you actually look at the 01:01:45.820 |
user agreements, what's happening now, if you use these reasoning bots is they say, don't just count 01:01:50.780 |
the tokens that you see. There's something called hidden tokens that it generated part of its answer 01:01:57.580 |
that we're hiding from you and the user, because you don't need to see all that, 01:02:00.300 |
that, but we have to charge you for it. Like, I mean, it produced all of this. 01:02:03.180 |
So you can kind of quantify that, oh, they are generating these really long answers and then 01:02:07.340 |
they're hiding a lot of it. So that like what you see is just some of the steps and then you get the 01:02:11.900 |
final answer. So that's what reasoning is. It's not thinking slower. The reason why it takes longer 01:02:18.620 |
is that it's just generating a much longer answer, right? That's what the slower is. Like when you 01:02:23.420 |
see, you know, you see this when you use chat GPT, you know, it takes a lot, each token they produce, 01:02:28.460 |
chat GPT still produces token by token. Deep seek is a gains efficiency by doing a few words at a time, 01:02:34.540 |
but you notice how the words like come in one by one. It's because for every one of these things they're 01:02:39.100 |
generating, you have 500 billion parameter, uh, transformer-based model pushing through to generate 01:02:45.500 |
just that word. So it takes longer to generate these reasoning model answers because what it's really 01:02:50.700 |
doing is writing like pages of explanation all the way up until its final answer. It just takes a long 01:02:57.740 |
time to do all that. They don't show you all that work, but that's what the slowness is, 01:03:02.620 |
is it's just giving really long answers. So that's what reasoning is. Um, it is different than 01:03:08.780 |
when I talked around the New Yorker last year about planning and how language models can't plan. And they 01:03:13.340 |
said, ultimately what language models will need to be, uh, flexible planners is planning engines outside 01:03:19.340 |
of the language model. And it like you, and you need some sort of coordinator that is going to, uh, 01:03:25.500 |
talk to a language model and talk to a planning engine. The canonical example is like playing 01:03:29.980 |
chess against an AI bot. Like ultimately you can't just have a giant neural network that can, uh, 01:03:35.660 |
be hard coded to, to look ahead a hundred thousand possible moves because that's just too big of a 01:03:41.340 |
neural network to hard code. You, you, you need some sort of like systematic simulator to check those. 01:03:45.820 |
And then maybe like the neural network to look at its quality, et cetera. So I still think there's only 01:03:50.940 |
so much, this helps answers. It's, it's not the, the really sophisticated how from 2001 style 01:03:57.340 |
reasoning is going to require non language model components that can actually recurse and do loops 01:04:03.180 |
and have state and actually explore things more systematically. But that is what I think is going 01:04:07.020 |
on with reasoning. It's the same language models based on the same type of original training that we've 01:04:12.300 |
seen recently. These big 500 billion parameter models, plus reinforcement learning that gets it to show its 01:04:18.940 |
work. And when language models show their work for certain types of problems, they're more likely to 01:04:23.180 |
be accurate. So that is what is going on. I can't find this explanation in almost any common press 01:04:28.060 |
article about this, but I think that is what is really happening. All right. So I can't help myself, 01:04:32.780 |
Jesse, but there we go. Are you going to write an article about it? I mean, it's not really like a 01:04:37.660 |
New Yorker article. I don't know. This is more of like a science, like a tech explanation article. 01:04:42.140 |
You did a chat GPT article, didn't you? I've done some. Yeah. It's kind of boring. 01:04:47.980 |
Yeah. So I'm making you guys listen to this instead. 01:04:50.860 |
I think it's interesting. I'm just kind of frustrated that none of like the, 01:04:55.100 |
the, the people who write the tech explainer articles, all the big names, 01:04:58.220 |
they're all just falling back on this. It's just taking time to think more. And that's how the 01:05:02.940 |
reasoning works. It's like, well, let's like actually understand, like, is there something, 01:05:06.380 |
is there a new architecture? Is there a new technology? Like what's happening? 01:05:09.260 |
Yeah. I had to read academic papers to try to just get that explanation. I think that's what's going 01:05:13.580 |
on. All right. Well, speaking of slowing down, I think that's all the time we have for today, 01:05:17.900 |
but we'll be back next week with another episode. And until then, as always stay deep. 01:05:22.460 |
If you liked today's discussions about how to become a more elite thinker, 01:05:26.940 |
you should check out episode 345, where I first introduced the idea that we need to train our brain. 01:05:32.860 |
That was called, are we getting dumber? I think you'll like it. Check it out. 01:05:36.300 |
So several people recently sent me the same article. It was from the financial time. So it was written 01:05:41.660 |
by John Byrne Murdoch and it had a provocative headline, have humans past peak brain power.