fbpx

PODCAST

Why Don’t We Learn From History by B.H.Liddell Hart w/Tom Libby & Jesan Sorrells

Why Don’t We Learn From History by B.H. Liddell Hart w/Tom Libby & Jesan Sorrells

Delving into B.H. Liddell Hart’s Why Don’t We Learn From History, hosted by Jesan Sorrells and Tom Libby, examines why societies—and leaders in business, politics, and communities—struggle to absorb history’s lessons. They discuss the enduring influence of emotional bias, the pitfalls of democratic and aristocratic decision-making, and the critical role of passionate history teaching in shaping informed citizens and leaders. The episode draws connections from military history to modern business trends, showing how historical context and human nature impact leadership choices and organizational outcomes.

  • Book Title: Why Don’t We Learn From History
  • Author: B.H. Liddell Hart
  • Guest Names: Jesan Sorrells (host), Tom Libby (co-host)


Time Stamped Overview

00:00 Welcome and Introduction – Why Don’t We Learn From History by B.H. Liddell Hart.
01:00 “Cracks in the Human Condition.”

09:17 Informal Committees Shape Decisions.

11:07 “Psychology, Memory, and Bias”

18:29 “Majority Shapes Right and Wrong.”

25:14 “Negotiating Reality in America.”

30:30 “Systems, Democracy, and Global Chaos.”

35:17 Historical Bias and Emotional Impact.

39:29 AI Adoption: Reality vs Expectations.

44:20 “AI Essential for Business Success.”

48:27 “Pre-WWII Polish Guarantee Debate.”

56:42 Post-COVID Education System Concerns.

59:43 “Making Poetry and History Alive.”

01:06:26 Teaching Complexity in History.

01:11:23 “Drunk on Ideas, Not Rote.”

01:16:12 “Perspective and Counterfactual Insights.”

01:23:39 “The Perils of Intellectual Neglect.”

01:24:54 “Emotion, Truth, and Progress.”

01:31:28 “Misguided Solidarity and History.”

01:39:23 Passion for History and Context.

01:41:20 “Cycle of Pride and Conflict.”

01:46:57 “Wealth Fails to Ensure Legacy.”


Opening and closing themes composed by Brian Sanyshyn of Brian Sanyshyn Music.


★ Support this podcast on Patreon ★

1
00:00:01,360 –> 00:00:05,000
Hello, my name is Jesan Sorrells, and

2
00:00:05,000 –> 00:00:08,480
this is the Leadership Lessons from the Great Books podcast,

3
00:00:08,720 –> 00:00:12,560
episode number 167,

4
00:00:13,760 –> 00:00:17,600
which puts us at about. Let’s see, how

5
00:00:17,600 –> 00:00:20,560
many episodes is that? Eight.

6
00:00:21,040 –> 00:00:24,640
Yeah. Eight episodes from our 175th

7
00:00:25,120 –> 00:00:28,910
episode. So we should get there by the end of this year. And thank you

8
00:00:28,910 –> 00:00:31,430
all for joining us on this journey. All right,

9
00:00:32,710 –> 00:00:35,990
so on this show, my co host

10
00:00:36,390 –> 00:00:39,670
who’s joining me today, Tom and I have

11
00:00:40,070 –> 00:00:43,830
circled around and around and around, coming back

12
00:00:44,070 –> 00:00:47,750
similar to the Nietzschean Myth of Return, to

13
00:00:47,990 –> 00:00:51,670
a core idea that is reflected in all of the books

14
00:00:51,800 –> 00:00:55,480
that we talk about on this show. And it is

15
00:00:55,480 –> 00:00:58,880
the idea that, and Tom is probably going to say it again today, the more

16
00:00:58,880 –> 00:01:02,680
things change, the more they stay the same.

17
00:01:03,560 –> 00:01:06,920
And this is true. It’s. It’s what people in Washington,

18
00:01:06,920 –> 00:01:10,760
D.C. pre Covid used to call a true fact.

19
00:01:10,920 –> 00:01:13,480
I don’t know what they call it now, but it is definitely true,

20
00:01:14,600 –> 00:01:18,430
which is why we keep returning to this idea repeatedly, no matter the

21
00:01:18,430 –> 00:01:21,630
book, no matter the genre, and no matter the author.

22
00:01:22,990 –> 00:01:26,790
In one way, if you’re listening to this, you could conclude that

23
00:01:26,790 –> 00:01:30,150
this fact of return reflects something inherent in the human condition

24
00:01:30,150 –> 00:01:33,989
itself. I think Richard Messing, who

25
00:01:33,989 –> 00:01:37,590
came on and talked with us about man’s search for meaning, and Viktor Frankl

26
00:01:37,590 –> 00:01:41,070
might say that. Right. In another

27
00:01:41,070 –> 00:01:44,470
way, you could say that this fact of noticing and

28
00:01:44,470 –> 00:01:47,950
recognizing such a conclusion reflects the idea that we are

29
00:01:47,950 –> 00:01:51,430
aware, deeply so, of the broken parts and

30
00:01:51,430 –> 00:01:55,190
cracks in the facade of how we address the clearly

31
00:01:55,190 –> 00:01:57,590
broken parts of the human condition.

32
00:01:58,790 –> 00:02:02,590
And it may indicate that noticing the how opens up an

33
00:02:02,590 –> 00:02:06,310
opportunity for all of us as leaders to seek and to

34
00:02:06,310 –> 00:02:09,950
explore and to maybe try to get a glimpse of the

35
00:02:09,950 –> 00:02:13,670
light that lies behind the cracks in the human

36
00:02:13,910 –> 00:02:17,750
condition. So today on the show we are

37
00:02:17,750 –> 00:02:21,470
covering, we’re going to talk about topics from a book

38
00:02:21,470 –> 00:02:25,190
that does its best and the best that it can

39
00:02:25,670 –> 00:02:29,430
to engage in noticing the cracks in the human face of

40
00:02:29,430 –> 00:02:32,710
the problems and challenges of the human condition. And

41
00:02:33,110 –> 00:02:36,910
a book that tries to examine and really talk about what lies behind those cracks

42
00:02:36,910 –> 00:02:40,620
in a coherent fashion. So let’s start on

43
00:02:40,620 –> 00:02:44,380
our journey repeating probably same things

44
00:02:44,380 –> 00:02:48,140
we’ve repeated before on this show, although there’s always new people joining us,

45
00:02:48,140 –> 00:02:51,820
so it’s always new to you. Today we will

46
00:02:51,820 –> 00:02:55,340
be covering the book and the question

47
00:02:55,500 –> 00:02:59,100
posed by it. Oh, I love that. Why don’t we Learn

48
00:02:59,100 –> 00:03:02,780
From History by B.H. liddell

49
00:03:03,180 –> 00:03:07,030
Hart? By the way, my book has a yellow cover.

50
00:03:07,590 –> 00:03:10,470
Your mileage will vary. I do like the yellow cover.

51
00:03:11,670 –> 00:03:15,350
Leaders. The penultimate question of all time, which

52
00:03:15,350 –> 00:03:19,110
is the question of why don’t we learn from history, has nothing

53
00:03:19,110 –> 00:03:21,910
but hard answers to it.

54
00:03:23,830 –> 00:03:27,590
And of course, on my journey through this book,

55
00:03:27,990 –> 00:03:31,420
we will be joined today by our regular co host, Tom Libby.

56
00:03:31,730 –> 00:03:35,490
How you doing today, Tom? Doing well, Jesan. And you know, I gotta be

57
00:03:35,490 –> 00:03:38,850
honest, I’m not surprised in the least that you selected me to help you with

58
00:03:38,850 –> 00:03:41,730
this book because to your point, I think I’ve said

59
00:03:42,530 –> 00:03:45,810
at least two dozen times on this podcast

60
00:03:46,130 –> 00:03:49,570
that phrase that the more things change, the more things stay the same. So I

61
00:03:49,570 –> 00:03:53,370
appreciate, I appreciate you tapping me in on this one because I think we’re

62
00:03:53,370 –> 00:03:57,130
gonna have a lot of fun with the, with this particular subject matter.

63
00:03:57,130 –> 00:04:00,810
So, so appreciate it. Yeah, absolutely. Well, I would encourage everybody to go and

64
00:04:00,810 –> 00:04:04,050
listen to the intro episode where we talk a little bit about

65
00:04:04,690 –> 00:04:06,850
BH Lidell Hart as a writer.

66
00:04:08,530 –> 00:04:12,210
He was, he was an interesting guy just

67
00:04:12,210 –> 00:04:15,730
in general. He was part of the generation of

68
00:04:15,730 –> 00:04:18,210
folks that was born in the,

69
00:04:19,570 –> 00:04:23,250
in the late 19th century, came of age

70
00:04:23,410 –> 00:04:26,130
in the early 20th century,

71
00:04:27,310 –> 00:04:30,990
fought in World War I. And just like C.S. lewis and

72
00:04:30,990 –> 00:04:34,110
J.R.R. tolkien and Eric Remarque

73
00:04:34,750 –> 00:04:38,550
and other notable names like T.E. lawrence, Lawrence of

74
00:04:38,550 –> 00:04:41,150
Arabia, Winston Churchill,

75
00:04:42,589 –> 00:04:46,270
and of course, your friend and mine, Adolf Hitler,

76
00:04:48,590 –> 00:04:52,350
he learned all kinds of interesting lessons

77
00:04:52,350 –> 00:04:55,750
about history, BH Lidelhart did, and about human

78
00:04:55,750 –> 00:04:59,590
nature in the trenches of the Somme

79
00:04:59,670 –> 00:05:03,150
and of Verdun and at Verdun. And we

80
00:05:03,150 –> 00:05:06,750
explore a lot of that history on that intro episode. So I would

81
00:05:06,750 –> 00:05:10,070
encourage you to, to go and take a listen to it. Now. We’re not going

82
00:05:10,070 –> 00:05:13,750
to dive so much into, into Liddell Hart’s as a person today

83
00:05:13,750 –> 00:05:16,710
on the show. Instead, we’re going to really focus on, we’re really going to focus

84
00:05:16,710 –> 00:05:20,190
on the ideas in the book. And, and when you open up, why don’t we

85
00:05:20,190 –> 00:05:23,750
learn from history what you see. And I mentioned this in the intro

86
00:05:23,750 –> 00:05:27,350
episode as well. What you see is that the

87
00:05:27,350 –> 00:05:30,950
preface was written by his son, which is great. The version that I have,

88
00:05:30,950 –> 00:05:34,790
Adrian J. Liddell Hart, who actually made a name for himself in

89
00:05:34,790 –> 00:05:37,950
World War II and in the military as well.

90
00:05:38,510 –> 00:05:41,190
And when you open up the table of contents, you see that the book is

91
00:05:41,190 –> 00:05:44,270
divided into three parts, right? There’s

92
00:05:44,670 –> 00:05:48,030
History and Truth, which is a great way to start a book

93
00:05:48,480 –> 00:05:52,280
entitled why Don’t We Learn from History, Government and Freedom and then

94
00:05:52,280 –> 00:05:56,120
War and Peace. And so as a military historian, you would think that he would

95
00:05:56,120 –> 00:05:59,520
start with War and Peace first, but no, no, he saves that for the end.

96
00:06:00,160 –> 00:06:02,640
And each One of the

97
00:06:03,520 –> 00:06:07,120
sections features a short essay, usually no more than four

98
00:06:07,120 –> 00:06:10,800
pages, where he lays out his ideas very succinctly.

99
00:06:11,040 –> 00:06:14,080
And this book really, Tom, really

100
00:06:15,570 –> 00:06:19,370
sort of puts me in mind of the essay

101
00:06:19,370 –> 00:06:23,050
by George Orwell that we covered where he talked about English literature.

102
00:06:23,050 –> 00:06:26,890
Right. It’s, it’s sort of the clearest example of clear writing that I’ve

103
00:06:26,890 –> 00:06:30,649
seen in a while. And, and partially that’s because they’re both, you

104
00:06:30,649 –> 00:06:33,570
know, they’re both English. They both came out of the English, you know, writing,

105
00:06:35,570 –> 00:06:39,410
thinking, literature structure that was built in Europe.

106
00:06:39,410 –> 00:06:43,220
Right. So they’re products of that. They’re products of that European and

107
00:06:43,220 –> 00:06:46,860
English tradition that goes back, you know, well over

108
00:06:46,940 –> 00:06:50,300
1500 years of just getting clarity in writing.

109
00:06:52,220 –> 00:06:55,580
And as a military historian, Liddell Hart really pursued

110
00:06:56,140 –> 00:06:59,860
clarity in thinking and clarity in writing. And so when you read

111
00:06:59,860 –> 00:07:03,500
these essays, he doesn’t mince words. There’s no

112
00:07:03,500 –> 00:07:05,660
fat in anything in here.

113
00:07:07,670 –> 00:07:11,350
So I’m going to open up with

114
00:07:11,670 –> 00:07:15,150
the section Restraints of Democracy and Power

115
00:07:15,150 –> 00:07:18,990
Politics in a Democracy. And this essay is in Government

116
00:07:18,990 –> 00:07:22,390
and Freedom, which is in the, the second

117
00:07:22,710 –> 00:07:26,550
section of his essays. And each one of the

118
00:07:26,550 –> 00:07:29,350
essays does build on the other one. So we’re going to pull them out separately,

119
00:07:29,350 –> 00:07:32,390
we’re going to talk about them in relation to why don’t we learn from history?

120
00:07:32,550 –> 00:07:36,310
But there’s lessons for leaders in all of this, you know, whether you’re a

121
00:07:36,310 –> 00:07:40,060
civic leader, leader of a nonprofit, leader of a for profit, or

122
00:07:40,060 –> 00:07:42,100
even just a leader of your family or community.

123
00:07:43,700 –> 00:07:47,380
So let’s go to the book Restraints

124
00:07:47,380 –> 00:07:51,020
of Democracy, and I quote, we

125
00:07:51,020 –> 00:07:54,300
learn from history that democracy has commonly put a premium on

126
00:07:54,300 –> 00:07:58,100
conventionality. By its nature, it prefers those who keep

127
00:07:58,100 –> 00:08:01,940
in step with the slowest march of thought and frowns on

128
00:08:01,940 –> 00:08:05,540
those who may disturb, quote, unquote, the conspiracy for mutual

129
00:08:05,540 –> 00:08:09,260
inefficiency. Thereby, this system

130
00:08:09,260 –> 00:08:12,820
of government tends to result in the triumph of mediocrity and

131
00:08:12,820 –> 00:08:16,380
entails the exclusion of first rate ability. If this is combined

132
00:08:16,380 –> 00:08:20,180
with honesty. But the alternative to it,

133
00:08:20,180 –> 00:08:23,580
despotism, almost inevitably means the triumph of

134
00:08:23,660 –> 00:08:27,420
stupidity. And of the two evils, the former

135
00:08:27,660 –> 00:08:31,420
is less. Hence it is better that ability should

136
00:08:31,420 –> 00:08:35,220
consent to its own sacrifice and subordination to the regime of

137
00:08:35,220 –> 00:08:38,660
mediocrity. Rather than assist in establishing a regime where,

138
00:08:38,900 –> 00:08:42,340
in the light of past experience, brute stupidity will be enthroned

139
00:08:42,820 –> 00:08:46,340
and ability may only preserve its footing at the price of

140
00:08:46,340 –> 00:08:49,780
dishonesty. What is the value

141
00:08:50,020 –> 00:08:53,700
in England and America and what is of value

142
00:08:53,700 –> 00:08:57,300
in England and America and worth defending its tradition of freedom, the

143
00:08:57,300 –> 00:09:01,020
guarantee of its vitality. Our civilization, like the Greek, has,

144
00:09:01,020 –> 00:09:04,620
for all its blundering way, taught the value of freedom, of criticism, of authority,

145
00:09:04,620 –> 00:09:08,460
and of harmonizing this with order. Anyone who urges a different

146
00:09:08,460 –> 00:09:11,780
system for efficiency’s sake is betraying the vital

147
00:09:11,780 –> 00:09:15,580
tradition. Then he switches to power

148
00:09:15,580 –> 00:09:19,020
politics in a democracy. And I want to point this out.

149
00:09:19,340 –> 00:09:23,070
He says this talking about how decisions get

150
00:09:23,070 –> 00:09:26,830
made in a democratic government. He says this while committee meetings are

151
00:09:26,830 –> 00:09:30,390
not so frequently held in the late afternoon as in the morning, dinner

152
00:09:30,390 –> 00:09:34,070
itself provides both an opportunity and an atmosphere suited to the informal kind of committee

153
00:09:34,230 –> 00:09:37,749
that tends to be more influential than those which are formally constituted.

154
00:09:38,150 –> 00:09:41,430
The informal type is usually small, and the smaller it is, the more

155
00:09:41,430 –> 00:09:45,150
influential it may be. The two or three gathered

156
00:09:45,150 –> 00:09:48,950
together may outweigh a formal committee of 20 or 30 members, to which

157
00:09:48,950 –> 00:09:52,710
it may often be related under the blanket, where it is assembled by

158
00:09:52,710 –> 00:09:56,550
someone who has a leading voice in a larger official committee. For it

159
00:09:56,550 –> 00:10:00,310
will represent his personal selection in the way of consultants. And

160
00:10:00,310 –> 00:10:03,990
its members being chosen for their congeniality as well as for their advisory

161
00:10:03,990 –> 00:10:07,470
value is likely to reach clear cut conclusions, which in turn may be

162
00:10:07,470 –> 00:10:11,110
translated into the decisions of a formal committee. For at any

163
00:10:11,110 –> 00:10:14,590
gathering of 20 or 30 men there is likely to be so much diversity

164
00:10:14,830 –> 00:10:18,390
and nebulosity of views that the consent of the majority can

165
00:10:18,390 –> 00:10:21,730
generally be gained for any conclusion that is sufficiently definite and

166
00:10:21,880 –> 00:10:25,480
impressively backed by well considered arguments and sponsored by a heavyweight

167
00:10:25,480 –> 00:10:28,760
member, especially if the presentation is

168
00:10:28,760 –> 00:10:32,360
carefully stage managed.

169
00:10:36,840 –> 00:10:37,640
I love that

170
00:10:40,440 –> 00:10:43,280
we don’t have to talk about power politics because we, we don’t like to talk

171
00:10:43,280 –> 00:10:47,080
about that. But this still, especially in the

172
00:10:47,080 –> 00:10:50,720
current political landscape. What you just

173
00:10:50,720 –> 00:10:54,380
read could probably blow up half the people’s brains in country right

174
00:10:54,380 –> 00:10:58,060
now. Well, so let’s open

175
00:10:58,060 –> 00:11:00,820
up with that. The very first question that I have from here,

176
00:11:01,860 –> 00:11:04,420
Tom, why don’t we learn from history?

177
00:11:05,940 –> 00:11:09,620
Get right into it. Good lord. So I,

178
00:11:09,620 –> 00:11:12,900
I think there’s so many factors here and I, I. One thing that I will

179
00:11:12,900 –> 00:11:16,580
say, and I know we’re not going to get into Liddell’s life per

180
00:11:16,580 –> 00:11:20,110
se, but I wonder how

181
00:11:20,110 –> 00:11:23,910
much more, how much more impact his

182
00:11:23,910 –> 00:11:27,430
book would have had if he had today’s access to

183
00:11:27,430 –> 00:11:31,070
psychology and like the psychological research, like some of the

184
00:11:31,070 –> 00:11:34,630
psychological research behind some of the stuff that he talks about is like,

185
00:11:35,190 –> 00:11:38,630
it’s actually because, because of his book, I’ve, I’ve seen several

186
00:11:39,590 –> 00:11:40,950
research, research

187
00:11:44,870 –> 00:11:48,470
papers been done basically because of this. Right. So in,

188
00:11:48,630 –> 00:11:52,430
so to go back to your, to your point and, and I don’t know what

189
00:11:52,430 –> 00:11:55,830
the technical terms for them are, but like there is something to be said about,

190
00:11:55,830 –> 00:11:57,510
like, about

191
00:11:59,670 –> 00:12:03,510
memory bias, right? So like, so we sometimes don’t

192
00:12:03,510 –> 00:12:07,190
learn from history because quite honestly we’re only, we’re

193
00:12:07,190 –> 00:12:10,790
very biased to the history that we read, right? So take, I mean

194
00:12:10,790 –> 00:12:14,560
US History is a very good example of this when you go and you start

195
00:12:14,560 –> 00:12:16,080
reading, if you were to read.

196
00:12:18,240 –> 00:12:20,480
And the other part of it too, and I think he talks about it a

197
00:12:20,480 –> 00:12:24,000
little bit in the book where the, because the victor

198
00:12:24,000 –> 00:12:27,760
usually writes the majority of, of passages when it comes

199
00:12:27,760 –> 00:12:31,360
to enter your subject matter here, whether it’s

200
00:12:31,600 –> 00:12:35,040
War one, World War two, it could be the cola wars for all I care.

201
00:12:35,120 –> 00:12:38,880
It doesn’t matter. Like whatever, whatever conflict

202
00:12:38,880 –> 00:12:42,580
or situation that you’re talking about, it’s usually the, to

203
00:12:42,580 –> 00:12:46,180
the victor goes the spoils, right? So somebody who wins that fight or

204
00:12:46,180 –> 00:12:50,020
wins that race or wins that whatever is going to write.

205
00:12:50,020 –> 00:12:53,860
You’re going to pay more attention to their writings. Therefore, you’re going

206
00:12:53,860 –> 00:12:57,460
to see the results of the victory and you’re not going to, you have a

207
00:12:57,460 –> 00:13:01,020
very, you have a very, very conscious bias of what

208
00:13:01,020 –> 00:13:04,820
history looks like. So you tend to, to not worry about.

209
00:13:04,980 –> 00:13:08,740
Good example. Again, you are talking about the current political landscape that

210
00:13:08,740 –> 00:13:12,220
we’re talking about. Everybody. You can go back in history

211
00:13:12,220 –> 00:13:16,060
and pinpoint times in history where you can say

212
00:13:16,300 –> 00:13:19,420
Hitler gained control of

213
00:13:19,900 –> 00:13:23,380
the political catastrophe that

214
00:13:23,380 –> 00:13:27,140
happened in Germany. We view it now as

215
00:13:27,140 –> 00:13:30,780
a political catastrophe. Well, guess what? Because all the Allies wrote all the,

216
00:13:31,260 –> 00:13:34,780
wrote all the, all the history, right? So but

217
00:13:35,350 –> 00:13:39,070
at the moment and in the time frame of the, in Germany

218
00:13:39,070 –> 00:13:42,670
when all that was happening, the, the people in the moment did not view this

219
00:13:42,670 –> 00:13:46,390
as a, as a, necessarily a bad thing. Were there people that were

220
00:13:46,390 –> 00:13:50,150
like, hey, wait a second, should this really be happening? Maybe,

221
00:13:50,150 –> 00:13:53,990
but they, their voice was never heard because a vast

222
00:13:53,990 –> 00:13:57,670
majority of people, he was a charismatic speaker, people he followed,

223
00:13:57,670 –> 00:14:01,350
people followed him, et cetera, et cetera. And nobody ever saw it coming, so to

224
00:14:01,350 –> 00:14:04,890
speak. Yeah, everybody thought they could do

225
00:14:05,050 –> 00:14:08,410
a deal with that guy. I mean even, even

226
00:14:09,850 –> 00:14:13,650
not Joseph Kennedy. Charles Lindbergh who ran for President. Charles Lindbergh

227
00:14:13,650 –> 00:14:17,410
who ran for President. Henry Ford, right. Who wanted

228
00:14:17,410 –> 00:14:21,130
to. Not wanted to, but helped. I

229
00:14:21,130 –> 00:14:24,930
believe it was either Mercedes or it

230
00:14:24,930 –> 00:14:28,210
might have been IG Farben. I can’t remember who he helped out, but he went

231
00:14:28,210 –> 00:14:30,890
over to, he did, he went over to Germany and he helped him set up

232
00:14:31,150 –> 00:14:34,950
factories to like for the purposes of re. Industrialization after

233
00:14:34,950 –> 00:14:38,110
the, to your point, the disaster of the Weimar Republic and inflation.

234
00:14:38,430 –> 00:14:42,190
Right. But Henry Ford was another guy who thought,

235
00:14:42,270 –> 00:14:45,950
yeah, okay, you know, we can, we can deal with this guy. Even Joe

236
00:14:45,950 –> 00:14:49,790
Kennedy, John Kennedy’s and Ted Kennedy’s dad,

237
00:14:50,270 –> 00:14:54,110
Joe Kennedy got in trouble. I believe he was the ambassador

238
00:14:54,190 –> 00:14:57,870
to England from America during

239
00:14:57,870 –> 00:15:00,830
the Roosevelt administration. He got in trouble

240
00:15:01,990 –> 00:15:05,710
just before the war kicked off by basically saying, hey, you know what?

241
00:15:05,710 –> 00:15:09,350
This Hitler guy, he’s not terrible. Like, we

242
00:15:09,350 –> 00:15:12,830
could probably do a deal with him. It’s fine. Hell, Stalin did a deal with

243
00:15:12,830 –> 00:15:16,510
Hitler. So, so what is the rest of the world looking at our

244
00:15:16,510 –> 00:15:20,270
current administration as? And there’s. There are people in our

245
00:15:20,270 –> 00:15:24,110
country that thinks that that’s, that, that we’re watching history repeat itself as

246
00:15:24,110 –> 00:15:27,910
we speak now. There are. And I want to, I want

247
00:15:27,910 –> 00:15:30,710
to talk a little about that today, because reading this book in the context of

248
00:15:30,710 –> 00:15:34,540
the current political climate that we are in was extremely

249
00:15:34,540 –> 00:15:37,460
interesting. Yeah. There’s something that I, I think

250
00:15:38,260 –> 00:15:41,980
there’s stark differences that. Oh, yeah, what was in place? Like, we have

251
00:15:41,980 –> 00:15:45,540
some checks and balances and we have things that like that are, that our government.

252
00:15:45,940 –> 00:15:49,700
I, I don’t see, I don’t see our current

253
00:15:49,700 –> 00:15:53,460
administration turning into Hitler, but there are people in our country that think that

254
00:15:53,860 –> 00:15:57,580
there are. There are. Correct. Right. Because. Because some of the signs and stuff are

255
00:15:57,580 –> 00:16:01,150
there. But, but again, where Germany didn’t have checks and

256
00:16:01,150 –> 00:16:04,790
balances, we do. The science can be there all they want as long

257
00:16:04,790 –> 00:16:08,110
as our government operates as the way that they’re supposed to

258
00:16:08,110 –> 00:16:11,750
operate. We don’t, we, we’re not going to have that. It’s, it’s,

259
00:16:11,750 –> 00:16:14,789
it’s also, whether you like. Him or not is not my point. I could care

260
00:16:14,789 –> 00:16:17,190
less whether you like him or not. That’s not, that’s not what I’m getting at

261
00:16:17,190 –> 00:16:20,190
here. Yeah, that’s not what we’re getting at here. What we’re talking about also. And

262
00:16:20,190 –> 00:16:24,030
this gets to, to what I just read there. So the

263
00:16:24,030 –> 00:16:26,990
ways in which most European

264
00:16:27,890 –> 00:16:31,570
governments were set up even after World War I,

265
00:16:31,570 –> 00:16:34,610
everything cracked apart after World War I, but

266
00:16:35,490 –> 00:16:38,930
certain ways still struggled on. Even 80 years

267
00:16:39,250 –> 00:16:42,770
later in our time, there’s still evidence of this.

268
00:16:43,010 –> 00:16:46,770
So Europe and England and specifically

269
00:16:47,010 –> 00:16:48,610
European countries like France,

270
00:16:51,250 –> 00:16:54,910
Russia, not so

271
00:16:54,910 –> 00:16:58,430
much Spain, although you could throw a spade in there. Italy for sure.

272
00:16:58,510 –> 00:17:02,270
And of course, Germany come out of a

273
00:17:02,350 –> 00:17:05,230
concept or have a, have an inbuilt concept of

274
00:17:05,470 –> 00:17:08,910
aristocratic rule that we don’t have.

275
00:17:10,350 –> 00:17:14,190
We explicitly rejected that. And so

276
00:17:14,350 –> 00:17:18,190
in the United States, our founding explicitly rejects aristocratic rule.

277
00:17:18,830 –> 00:17:22,450
This is where the, and these

278
00:17:22,450 –> 00:17:25,690
people, maybe they had a point, maybe they didn’t, but that the protesters this summer

279
00:17:25,690 –> 00:17:29,250
were talking about no kings or whatever. Like. Well,

280
00:17:29,570 –> 00:17:33,330
I mean, if you look at the Constitution and if you look at the three,

281
00:17:34,610 –> 00:17:38,210
the three branches of government, like they’re

282
00:17:38,210 –> 00:17:42,010
functioning exactly constitutionally as they should be. You just don’t like

283
00:17:42,010 –> 00:17:45,770
the decisions that they’re making. Which is, which is, which is why you

284
00:17:45,770 –> 00:17:49,490
get to vote. That’s why you, that’s why, that’s

285
00:17:49,490 –> 00:17:53,200
why, like we were talking about with, with Charlie Kirk’s assassination

286
00:17:53,200 –> 00:17:56,680
and around that this is why freedom of speech matters. Because which,

287
00:17:56,920 –> 00:18:00,320
which by the way, we also don’t have a tradition of in a European context.

288
00:18:00,320 –> 00:18:03,400
And so how decisions get made in an

289
00:18:03,400 –> 00:18:07,240
aristocratic, with an aristocratic mindset, in an aristocratic

290
00:18:07,240 –> 00:18:11,080
manner is fundamentally different than how decisions get made in a more. And this is

291
00:18:11,080 –> 00:18:14,720
point that Liddell Hart’s making in a more democratic mindset. So when he

292
00:18:14,720 –> 00:18:18,440
talks about a small cadre of people making decisions and then basically stage

293
00:18:18,440 –> 00:18:22,020
managing them for 20 or 30 other people, that comes. And

294
00:18:22,020 –> 00:18:25,300
Liddell Hart was English. Comes out of a specific aristocratic

295
00:18:25,300 –> 00:18:28,780
mindset. Yeah, yeah. That we don’t have.

296
00:18:29,340 –> 00:18:32,860
So, so as I started this conversation where I said

297
00:18:33,100 –> 00:18:36,860
I wish Lidell had access to some of the research that, that

298
00:18:36,860 –> 00:18:40,580
has, that psych psychology has come leaps and bounds over the last, you know,

299
00:18:40,580 –> 00:18:44,140
80 years since, since World War II. So, so there’s, there’s that,

300
00:18:44,220 –> 00:18:47,980
there’s that, that, you know, that hindsight bias that we have

301
00:18:48,320 –> 00:18:52,160
because, because history is written by the victors most times.

302
00:18:52,560 –> 00:18:56,280
And again, we’re getting better at that, but not. Still not great. But then

303
00:18:56,280 –> 00:18:59,960
there you have the other. I, I was once told that the difference

304
00:18:59,960 –> 00:19:03,760
between right and wrong is the majority. And that

305
00:19:03,760 –> 00:19:07,440
really, that really hit me hard too because now you’re saying to me

306
00:19:07,440 –> 00:19:11,040
that I could know that one plus one, we go back to Orwell, right?

307
00:19:11,120 –> 00:19:14,080
I could know that one plus one or two plus two equals four. But you’re

308
00:19:14,080 –> 00:19:16,960
going to say that the majority of people say it’s five. So now it’s just

309
00:19:16,960 –> 00:19:20,180
right. Five is right from now on. But that in.

310
00:19:20,740 –> 00:19:24,180
When it comes to something so linear as

311
00:19:24,180 –> 00:19:27,980
math, maybe you can make arguments against it. But when it comes

312
00:19:27,980 –> 00:19:31,220
to something that’s either opinion or, or

313
00:19:32,500 –> 00:19:36,300
consensus based or like, there’s a lot, there are a

314
00:19:36,300 –> 00:19:40,060
lot of situations where the right thing to do and this

315
00:19:40,060 –> 00:19:43,300
is where we get the whole Democratic vote, right? So you’re going to vote 100

316
00:19:43,300 –> 00:19:47,090
people vote or you know, 200 people vote and, or

317
00:19:47,090 –> 00:19:50,130
100 people vote and 51 of them say this. So we’re just going to do

318
00:19:50,130 –> 00:19:53,810
that. And 49 of them can know damn well that it’s the wrong thing to

319
00:19:53,810 –> 00:19:57,650
do. But the right. The difference between right and wrong is the

320
00:19:57,650 –> 00:20:01,490
majority right. Like, so there’s also some of that that happens throughout the

321
00:20:01,490 –> 00:20:04,810
course of history. And then there’s the final one that

322
00:20:05,290 –> 00:20:07,130
I think of, quite honestly is

323
00:20:09,050 –> 00:20:12,880
there’s a disassociation of time that happens. Right. The further

324
00:20:12,880 –> 00:20:16,160
away from something we get, we become more arrogant that

325
00:20:16,800 –> 00:20:19,800
we can see it, we know it’s happening. We’re not going to make the same

326
00:20:19,800 –> 00:20:23,520
mistakes because we know they’re there, but yet we do because we have a bias

327
00:20:23,920 –> 00:20:27,680
of disassociation of time. And I’ll give you an example of this one.

328
00:20:27,920 –> 00:20:31,720
Now, for those of you who can’t see me on the video here, I’m

329
00:20:31,720 –> 00:20:35,480
not a woman, so I’m not speaking as a woman. But childbirth, to me is

330
00:20:35,480 –> 00:20:39,320
a very good example of this on an individual basis. You’re a parent,

331
00:20:39,320 –> 00:20:42,200
I’m a parent. I don’t know if you spent time in the delivery room with

332
00:20:42,200 –> 00:20:45,640
your wife, but I did. Yeah. I watched what she went through

333
00:20:45,800 –> 00:20:49,440
and I went. Why would anybody do this more than once? Like,

334
00:20:49,440 –> 00:20:52,960
why? Like, honestly, like, the amount of the, the. The. The mental

335
00:20:52,960 –> 00:20:56,520
anguish, the physical anguish, the, the pain they go through. Like,

336
00:20:57,160 –> 00:21:00,880
it, it. Childbirth, to me, is one of the most fascinating things

337
00:21:00,880 –> 00:21:04,440
on the. In the entire natural world. Because

338
00:21:05,530 –> 00:21:08,810
women decide to do this again. Like, Right. I know, I know.

339
00:21:09,210 –> 00:21:12,410
Let me just say this for, for the record. I know,

340
00:21:12,650 –> 00:21:16,330
guys, if we went through that once, we’d be like, hell, no, we’re

341
00:21:16,330 –> 00:21:19,970
done. Nope, because we wouldn’t do that

342
00:21:19,970 –> 00:21:22,810
again. Like, it’s like we don’t have the same mindset. But. But to get back

343
00:21:22,810 –> 00:21:25,810
to Liddell and this. The reason I say it this way is because there’s a

344
00:21:25,810 –> 00:21:29,530
disassociation with time. So when women have. If you ever notice, like,

345
00:21:30,090 –> 00:21:33,630
two years later, they didn’t remember the pain the same

346
00:21:33,630 –> 00:21:36,830
way that you, observer, do. Right?

347
00:21:37,150 –> 00:21:40,790
Right. So, like. Right. By the way, guys, I’m totally kidding. Because we do

348
00:21:40,790 –> 00:21:44,550
stupid. We still do stupid stuff all the time. And we continue. You

349
00:21:44,550 –> 00:21:47,349
fall off a ladder, you still climb the ladder. I mean, you know, I’m just

350
00:21:47,349 –> 00:21:50,710
saying, like, as an observer, as an observer watching

351
00:21:50,710 –> 00:21:54,510
childbirth, and you think to yourself, why would anybody ever do this again? But

352
00:21:54,590 –> 00:21:57,870
as a woman watches her child grow up and she gets

353
00:21:57,870 –> 00:22:01,670
disassociated from the time of it, she decides to

354
00:22:01,670 –> 00:22:05,470
have another child. And I think that is Another symptom of

355
00:22:05,470 –> 00:22:09,190
what we’re talking about here. The longer we go from this, from the.

356
00:22:09,190 –> 00:22:12,710
Again, take this, our current political landscape, to

357
00:22:12,710 –> 00:22:16,550
Hitler. You can make all the associations you want. We feel

358
00:22:16,550 –> 00:22:19,310
like it’s not going to happen again because of the checks and balances that we

359
00:22:19,310 –> 00:22:22,990
have in place. But, but if you’re just a simple observer looking

360
00:22:22,990 –> 00:22:26,590
in and you’re looking at this going, holy crap, it’s happening again

361
00:22:26,590 –> 00:22:30,380
because these people have had so much time in between that they didn’t realize

362
00:22:30,380 –> 00:22:33,580
it. And by the way, go backwards in time. And you can say the same

363
00:22:33,580 –> 00:22:37,300
thing about people like Napoleon, about Hannibal, about, like, just

364
00:22:37,300 –> 00:22:40,900
keep going. Like you. There’s. There’s plenty of instances where that

365
00:22:41,140 –> 00:22:44,980
singular person is. Is that dynamic shift in the

366
00:22:44,980 –> 00:22:48,780
power of. And the balance of power. So it’s

367
00:22:48,780 –> 00:22:52,620
happened several times. I, Again, I don’t think. I think

368
00:22:52,620 –> 00:22:55,380
we have. I think we figured out some checks and balances. But I could see

369
00:22:55,380 –> 00:22:59,030
this happening again. Look at the current situation in Russia.

370
00:22:59,030 –> 00:23:02,750
Putin. Oh, yeah, right. Like, he, he’s proving that this could

371
00:23:02,750 –> 00:23:06,350
potentially happen again in Russia for sure. Like that. Like, I don’t know if the

372
00:23:06,350 –> 00:23:10,150
Russian government has checks and balances to, to make sure that he doesn’t do that,

373
00:23:10,150 –> 00:23:13,630
but I don’t think they do. They don’t. He’s been. He’s been running things

374
00:23:13,790 –> 00:23:17,470
pretty well. Well, well, you know, pretty consistently, I would say. Well, pretty

375
00:23:17,470 –> 00:23:20,910
consistently. So I’m not claiming to know their political. But I’m just saying, like,

376
00:23:21,150 –> 00:23:24,390
he’s working. I think it’s fairly consistently for the last 30 years there. As an

377
00:23:24,390 –> 00:23:28,060
observer of history, I see this happening again.

378
00:23:28,060 –> 00:23:31,060
That’s, that’s, that’s the point. But it’s the observer. Yeah, but I think. I think.

379
00:23:31,060 –> 00:23:34,460
So between those three things, I think if Liddell had access

380
00:23:34,620 –> 00:23:38,300
to really deep research in those psychological profiles,

381
00:23:38,380 –> 00:23:41,980
I think his book would have been even more impactful. I, I’m not

382
00:23:41,980 –> 00:23:44,700
suggesting it’s not impactful. I. And I, I think it’s really.

383
00:23:45,740 –> 00:23:49,380
I think it’s a really good book. But. And of course, it answers a

384
00:23:49,380 –> 00:23:52,960
question. I think it answers the question pretty well for its time. Yeah.

385
00:23:52,960 –> 00:23:56,760
But think about a guy like Liddell having access to the current psychology

386
00:23:56,840 –> 00:24:00,000
research that we have and how much more impactful he could have been with that

387
00:24:00,000 –> 00:24:02,440
book. So a couple things there.

388
00:24:05,080 –> 00:24:08,600
So we use

389
00:24:08,760 –> 00:24:12,560
history to. And Lidell Hart talks about

390
00:24:12,560 –> 00:24:15,960
this in his book, too. A variation of this. And again, to your point, pre.

391
00:24:16,280 –> 00:24:19,960
Not pre. But the depth of psychological research. When he wrote this, and he wrote

392
00:24:19,960 –> 00:24:23,520
it in, I believe it was the. Yeah. Published in

393
00:24:23,520 –> 00:24:27,180
1944. Ye. Yeah, exactly. So, you

394
00:24:27,180 –> 00:24:30,460
know, the, the degree to which, to your point, the degree to which psychological research

395
00:24:30,460 –> 00:24:33,940
has come along since 1944 is. Is leaps and

396
00:24:33,940 –> 00:24:36,500
bounds ahead of what he. He had in his time.

397
00:24:38,260 –> 00:24:41,900
But even then he understood something about human nature, which gets

398
00:24:41,900 –> 00:24:45,420
to a couple of different things that we’ve talked about on this show. So we

399
00:24:45,420 –> 00:24:49,220
talked about it in our extra episode where we. Where we discussed the

400
00:24:49,220 –> 00:24:52,020
movie Oppenheimer, which interestingly enough, I watched again last night,

401
00:24:53,820 –> 00:24:57,660
kind of just weirdly lined up with this. And then.

402
00:24:58,300 –> 00:25:02,140
And then we also talked about this in the Orwell episode, not only

403
00:25:02,220 –> 00:25:05,580
on his. In his essay on literature in the English language,

404
00:25:06,140 –> 00:25:09,740
but also in his books, you know, animal farm in 1984.

405
00:25:09,740 –> 00:25:13,580
Right. There’s a through line here. Right.

406
00:25:14,060 –> 00:25:17,660
And even books that we’ve covered, a couple books we cover

407
00:25:18,460 –> 00:25:22,230
from Theodore Roosevelt. We covered the

408
00:25:22,230 –> 00:25:24,750
book that he wrote way back when he was a

409
00:25:26,430 –> 00:25:29,950
representative, I think, or a senator in New York State.

410
00:25:30,110 –> 00:25:33,070
He wrote a book about power politics. We talked about that with Libby Younger.

411
00:25:34,910 –> 00:25:37,990
And the through line that Tom is getting to, and I do agree with it,

412
00:25:37,990 –> 00:25:41,830
is this. We have to figure out, and

413
00:25:41,830 –> 00:25:45,430
this is part of the radical experiment of America, but we have to

414
00:25:45,430 –> 00:25:48,800
figure out how to negotiate

415
00:25:48,800 –> 00:25:51,880
reality with each other. So to get back to two plus two equals five

416
00:25:54,360 –> 00:25:58,040
and this, this, by the way, can start happening in the last 10 years here,

417
00:25:58,040 –> 00:26:00,840
because I’m going to go ahead and step on this third rail, because why not?

418
00:26:07,000 –> 00:26:10,360
If you want to say that two plus two equals five in a dynamic

419
00:26:12,210 –> 00:26:15,970
environment where you’re not abrogating my speech, you’re just providing me

420
00:26:15,970 –> 00:26:19,730
consequences for that speech, which is a whole other kind of discussion.

421
00:26:20,770 –> 00:26:23,930
If you want to go ahead and say two plus two equals five, knock yourself

422
00:26:23,930 –> 00:26:26,930
right the hell out. Sure, go ahead.

423
00:26:28,210 –> 00:26:31,570
But I’m going to. I’m going to channel Ben Shapiro here

424
00:26:31,970 –> 00:26:35,010
and I’m going to say facts don’t care about your feelings.

425
00:26:36,690 –> 00:26:40,530
So, like, you could say two plus two equals five all day, but

426
00:26:40,530 –> 00:26:44,330
I have two things. Then I put together two more

427
00:26:44,330 –> 00:26:47,850
things and invariably I’m going to have four things.

428
00:26:48,170 –> 00:26:52,010
Sorry. Like, this is just. This is just reality. This is the

429
00:26:52,010 –> 00:26:55,809
ceiling of, like, your logic. Right. You could feel any way you

430
00:26:55,809 –> 00:26:58,570
want about it. You could feel that it’s. And here we go. I’m going to

431
00:26:58,570 –> 00:27:02,170
step on the. Step on the rail. You could feel that it’s white supremacist.

432
00:27:02,410 –> 00:27:06,180
You could feel that it’s a sign of the patriarchy. You can feel

433
00:27:06,180 –> 00:27:09,900
that it’s oppressive. You can feel that it’s a sign of

434
00:27:09,900 –> 00:27:13,620
systemic oppression. You could feel that it’s a sign of

435
00:27:13,620 –> 00:27:16,420
white fragility. You could feel all of these things.

436
00:27:19,540 –> 00:27:23,340
And I hold up two fingers, and then I

437
00:27:23,340 –> 00:27:26,740
hold up two more fingers and I still only have four

438
00:27:26,740 –> 00:27:30,460
fingers. That’s

439
00:27:30,460 –> 00:27:33,460
it. That’s. That’s it. That’s it. That’s the whole thing. There’s no, there’s no more

440
00:27:33,460 –> 00:27:36,850
argument that I need to. Right. And so this, this has started happening, this

441
00:27:36,850 –> 00:27:40,250
renegotiation of reality that you’re talking about, which is

442
00:27:40,490 –> 00:27:44,130
fine with history. I don’t have a problem with renegotiating

443
00:27:44,130 –> 00:27:47,530
reality with history. And, and Liddell Hart talks a little bit about this in his

444
00:27:47,530 –> 00:27:51,290
book. You know, when myths become stronger than the history, the myth becomes

445
00:27:51,290 –> 00:27:55,050
the history. He, he mentions this, right. And it’s fine. I

446
00:27:55,050 –> 00:27:58,810
have no problem taking a scientific approach. He does. He has a

447
00:27:58,810 –> 00:28:01,770
problem with taking a scientific approach to history. He talks about it in the first

448
00:28:01,770 –> 00:28:05,050
part of his book because he thinks that it drains and denud

449
00:28:06,630 –> 00:28:09,790
of all of its emotive value. And he doesn’t think that you should be able

450
00:28:09,790 –> 00:28:13,550
to do that. He thinks you should balance the emotive value with the scientific

451
00:28:13,550 –> 00:28:17,030
pursuit of the truth. I. Okay, I can see his

452
00:28:17,030 –> 00:28:20,830
argument, But I also think that human beings are going

453
00:28:20,830 –> 00:28:23,470
to look for that emotive value and they’re going to look forward in myths. You

454
00:28:23,470 –> 00:28:25,710
can’t take that out of the human because again, that’s one of those, like two

455
00:28:25,710 –> 00:28:29,270
plus two equals four things. It’s one of the things that’s built into human understanding

456
00:28:29,270 –> 00:28:32,970
of reality. But if we want to renegotiate history, sure,

457
00:28:32,970 –> 00:28:36,250
let’s renegotiate history. Let’s go ahead and dig into the

458
00:28:36,250 –> 00:28:39,010
Tuskegee Project. Let’s go ahead and dig into

459
00:28:40,370 –> 00:28:43,970
Black Wall street in Oklahoma. Let’s go ahead and dig

460
00:28:43,970 –> 00:28:47,730
into the decimation of the Native American

461
00:28:47,730 –> 00:28:51,410
tribes in, in, in North America

462
00:28:51,570 –> 00:28:55,250
and in Mexico and in South America. Let’s go ahead and dig into that

463
00:28:55,810 –> 00:28:58,370
and understand that when you do that,

464
00:29:00,140 –> 00:29:03,780
and this is something that we also forget with history, I think, and

465
00:29:03,780 –> 00:29:07,540
it’s notorious in warfare. Understand that when you do

466
00:29:07,540 –> 00:29:11,100
that, the enemy you are fighting or you are opposing

467
00:29:11,820 –> 00:29:14,380
also gets a vote in a free speech society.

468
00:29:16,300 –> 00:29:19,900
So when you start pulling apart people’s myths, they’re going

469
00:29:19,900 –> 00:29:23,540
to have a reaction to that and a response to

470
00:29:23,540 –> 00:29:27,270
that. Now, we would like to keep that response and reaction in the space of

471
00:29:27,270 –> 00:29:30,670
speech. So there’s going to be an argument, there’s going to be a

472
00:29:30,670 –> 00:29:33,110
verbal conflict, there’s going to be a fight.

473
00:29:34,710 –> 00:29:37,950
And I think in the structure that we have in America, that’s fine. We should

474
00:29:37,950 –> 00:29:40,950
be doing that. Now you have to talk about the observer from the outside. If

475
00:29:40,950 –> 00:29:43,750
I’m an observer from the outside looking at this, all this looks like is just

476
00:29:43,750 –> 00:29:47,510
massive chaos, just endless, never ending massive chaos

477
00:29:47,830 –> 00:29:51,390
about things that should just be decided by an aristocratic

478
00:29:51,390 –> 00:29:55,210
cadre of individuals. Everybody gets told what the thing is

479
00:29:55,370 –> 00:29:57,970
and then you just move on. By the way, this is what they do in

480
00:29:57,970 –> 00:30:01,290
communist China. They have myths based on

481
00:30:01,290 –> 00:30:04,970
Confucianism and the Chinese communist government just says, we’ve decided

482
00:30:04,970 –> 00:30:06,970
this, you’re welcome.

483
00:30:09,370 –> 00:30:12,970
Everybody sort of goes, yeah, that’s fine. By the way, in those kinds of systems,

484
00:30:13,450 –> 00:30:17,170
to Orwell’s point, if you’re in a to totalizing

485
00:30:17,170 –> 00:30:20,890
system like that, where, where dissension from something is

486
00:30:20,890 –> 00:30:24,540
not allowed, this is the point from 1984. Yeah, the two

487
00:30:24,540 –> 00:30:28,180
plus two does equal five, sure, in

488
00:30:28,180 –> 00:30:32,020
that system. But guess what? That

489
00:30:32,020 –> 00:30:35,540
system has to negotiate with the reality of other systems in a global environment.

490
00:30:35,940 –> 00:30:39,380
Which is part of the reason why the Internet’s like wrecking havoc

491
00:30:39,380 –> 00:30:43,220
everywhere right now and has for the last 30, almost 40 years

492
00:30:43,220 –> 00:30:47,060
and actually more than 40 years now and will continue to wreak havoc, by the

493
00:30:47,060 –> 00:30:49,820
way, for the next hundred years because we’re not going to be able to get

494
00:30:49,820 –> 00:30:53,370
over this. And then the other thing that you’re talking about in there, so there’s

495
00:30:53,370 –> 00:30:56,690
negotiating reality, then there’s the other thing you’re talking about. There is democracy.

496
00:30:56,930 –> 00:31:00,450
And fundamentally, as an aristocrat,

497
00:31:01,250 –> 00:31:04,970
B. H Liddell Hart was opposed to democracy and he’s joined

498
00:31:04,970 –> 00:31:08,650
in good as a fellow traveler, or he walks as a fellow

499
00:31:08,650 –> 00:31:12,290
traveler alongside Thomas Jefferson, who was opposed to democracy,

500
00:31:12,770 –> 00:31:16,570
and George Washington, who was opposed to democracy, and Ben Franklin who was

501
00:31:16,570 –> 00:31:19,970
opposed to democracy, and John Adams who was opposed to democracy.

502
00:31:20,290 –> 00:31:24,030
These guys wanted a. And they built a republic

503
00:31:24,830 –> 00:31:28,590
because democracies are two wolves and a

504
00:31:28,590 –> 00:31:31,390
sheep voting on who gets to be dinner.

505
00:31:32,590 –> 00:31:36,430
That’s a democracy. Congratulations. 66% of the people,

506
00:31:36,590 –> 00:31:40,030
66% of the entities in that group just voted on dinner

507
00:31:40,750 –> 00:31:44,150
and your vote didn’t count. That’s democracy. Democracy is

508
00:31:44,150 –> 00:31:47,990
inevitably, inevitably degrades to mob rule. We saw this, we see this

509
00:31:47,990 –> 00:31:51,600
throughout history, literally. So you build a republic,

510
00:31:52,000 –> 00:31:55,680
and in a republic, the two wolves,

511
00:31:56,880 –> 00:32:00,560
the two wolves have to send one wolf as a representative and the sheep

512
00:32:00,560 –> 00:32:03,520
gets to send two sheeps as a representative. And now we’re going to have a

513
00:32:03,520 –> 00:32:06,320
Fight. And that’s how we keep balance in the system.

514
00:32:07,200 –> 00:32:10,880
And so we’ve got all these dynamics which impact how we view

515
00:32:10,880 –> 00:32:14,600
history in America. But to your point, the through line is

516
00:32:14,600 –> 00:32:18,290
about negotiating reality and who gets to negotiate that and who gets to

517
00:32:18,290 –> 00:32:21,730
say what the other thing too. And through your little

518
00:32:21,730 –> 00:32:25,250
Spiel. Spiel there. I wanted to, I just wanted to jump in and say one

519
00:32:25,250 –> 00:32:28,850
thing before, but I obviously wanted to let you finish. But, but when

520
00:32:28,850 –> 00:32:32,489
you, when you’re referencing the two plus two, and I still have four

521
00:32:32,490 –> 00:32:36,090
and you know, fact. Listen, facts are facts. Right? Like, to your point,

522
00:32:36,410 –> 00:32:39,130
the problem with history is

523
00:32:40,090 –> 00:32:43,650
it’s almost never. Yes, the facts, ma’. Am. Right.

524
00:32:43,650 –> 00:32:47,410
Like, there has to. There’s always human emotion in it. There’s always

525
00:32:47,490 –> 00:32:51,250
like these degrees of champions and like,

526
00:32:51,250 –> 00:32:54,610
like there’s always these. There’s heroism and like we’re

527
00:32:54,610 –> 00:32:58,450
elevating things to, to make it feel more real or to more. Have

528
00:32:58,450 –> 00:33:02,170
more impact. And it’s never like, if you were to ever

529
00:33:02,170 –> 00:33:05,970
read a history book that was literally just factual information, nobody would read the

530
00:33:05,970 –> 00:33:09,330
whole. Nobody would read it. No, like, no, it’s like, wait,

531
00:33:09,570 –> 00:33:13,330
okay, so World War II started in this date, ended in this date. The

532
00:33:13,330 –> 00:33:17,010
battle of this date was this date, this day, the battle. Like, like. But

533
00:33:17,650 –> 00:33:21,370
because we tie everything into some sort of

534
00:33:21,370 –> 00:33:25,210
emotion, we have to feel an emotion for anything that we’re

535
00:33:25,210 –> 00:33:28,850
involved in, whether it’s reading history, creating history,

536
00:33:29,250 –> 00:33:32,924
looking at the future, etc. Etc. So to your point about the 2

537
00:33:32,976 –> 00:33:36,540
+2 is always going to equal 4. Yes, because there’s no

538
00:33:36,540 –> 00:33:40,380
emotion in that that you cannot, you can’t invoke emotion into

539
00:33:40,460 –> 00:33:41,980
math. Right.

540
00:33:44,380 –> 00:33:48,220
Like, I remember, I read recently, I read

541
00:33:48,220 –> 00:33:51,900
recently, somebody, somebody was arguing some political.

542
00:33:52,060 –> 00:33:55,780
Blah, blah, blah, whatever. And the person, literally, the, the argument, what

543
00:33:55,780 –> 00:33:58,620
they were talking about was like, listen, you cannot

544
00:33:58,780 –> 00:34:02,590
distort math. And they literally said

545
00:34:02,590 –> 00:34:06,310
exactly what you said. 2/2 is always going to equal 4. So if you’re looking

546
00:34:06,310 –> 00:34:09,510
at statistical data that is verified

547
00:34:09,990 –> 00:34:12,870
statistical data, I’m not talking about stuff that we just make up.

548
00:34:13,750 –> 00:34:16,990
First time I ever heard this line, it made me laugh so hard. It’s like,

549
00:34:16,990 –> 00:34:20,310
hey, Tom, did you know that 86 of all statistics are made up? And I

550
00:34:20,310 –> 00:34:23,270
went, there you go, what? Like, I had to stop. And I was like,

551
00:34:23,990 –> 00:34:26,550
that’s made up. Exactly. Right. So, like,

552
00:34:28,630 –> 00:34:32,370
so but if you. But if it’s verified statistical data and

553
00:34:32,370 –> 00:34:35,890
you can say that X number of people voted this way,

554
00:34:35,890 –> 00:34:39,370
X number of people voted that way. So 28 of people

555
00:34:39,370 –> 00:34:43,050
voted this Way based on the information they were given that is statistical.

556
00:34:43,050 –> 00:34:46,690
You cannot change that. That. It just is what it is,

557
00:34:46,770 –> 00:34:50,010
right? They were trying to make that argument and the guy was still fighting, like,

558
00:34:50,010 –> 00:34:53,810
pushing back, going, no, you can, you can, you can manipulate statistics.

559
00:34:54,130 –> 00:34:57,719
Well, yes, but then they’ve not. Then they’re not validated. They’re not

560
00:34:57,719 –> 00:35:01,359
verifiable statistics at that point. They’re 86% of

561
00:35:01,359 –> 00:35:04,879
statistics are made up. Like, you have to be able to validate it. If you

562
00:35:04,879 –> 00:35:08,479
can’t validate it, then it’s not, it’s not reasonable to see. So, so

563
00:35:08,479 –> 00:35:10,759
my, my point to all that is that,

564
00:35:11,879 –> 00:35:15,319
yes, mathematical information, things that you can

565
00:35:15,319 –> 00:35:19,039
literally see, two plus two equals four. I, if anybody

566
00:35:19,039 –> 00:35:22,479
tries to argue that, I just think they’re being silly, and all of us are

567
00:35:22,479 –> 00:35:26,110
going to think they’re being silly. But when they start debating on the

568
00:35:26,110 –> 00:35:29,870
merits of historical information that is written by people, that are

569
00:35:29,870 –> 00:35:32,910
written by the winners, so to speak, that are written by the

570
00:35:33,310 –> 00:35:37,030
majority, all of that stuff when

571
00:35:37,030 –> 00:35:40,790
you read that, you feel something when you read that. That’s

572
00:35:40,790 –> 00:35:44,230
another bias that we probably don’t talk about is the emotional

573
00:35:44,230 –> 00:35:47,710
bias that you get when you read some of those things. So therefore,

574
00:35:48,110 –> 00:35:51,830
if you see that the United States and its allies are going to

575
00:35:51,830 –> 00:35:55,390
win World War II based on what you’re reading, and you say, okay, so

576
00:35:55,630 –> 00:35:59,350
why, why do we have the. Why are we trying to prevent World War

577
00:35:59,350 –> 00:36:02,550
iii? We know the new the. The allies are going to win. That’s what happens,

578
00:36:02,550 –> 00:36:05,990
right? So we look throughout history, but, like, can we stop for a second? And

579
00:36:05,990 –> 00:36:08,790
let’s, like, not lose the millions of lives that we’re going to lose because of

580
00:36:08,790 –> 00:36:12,550
it. Like, is there a way to circumvent? Right. Just go from point A to

581
00:36:12,550 –> 00:36:15,470
point C instead of, like, can we skip point B,

582
00:36:16,750 –> 00:36:20,570
skip point D. Yeah, don’t learn. That’s what we don’t learn. Right. Well,

583
00:36:20,570 –> 00:36:23,970
and this is, this is. I mean, again, I watched Oppenheimer last night, right?

584
00:36:24,290 –> 00:36:28,010
And you know, it’s interesting. Like, there’s a movie coming out, and we always

585
00:36:28,010 –> 00:36:31,410
end up talking about movies. This is the moment to do this. Right? Now there’s

586
00:36:31,410 –> 00:36:35,090
a movie coming out from Catherine Bigelow, who directed A Hurt Locker

587
00:36:35,249 –> 00:36:37,810
and a couple other movies on Netflix

588
00:36:38,770 –> 00:36:42,490
about, like, Five Minutes to Nuclear Impact or

589
00:36:42,490 –> 00:36:45,570
something like that. Anyway, it reminds. It’s on. It’s going on Netflix. It’s coming out

590
00:36:45,570 –> 00:36:47,930
in December. I think it’s going to be in the theaters for, like, a brief

591
00:36:47,930 –> 00:36:50,250
minute, and then it’s going to. It’s called House of Fire, I think, or House

592
00:36:50,250 –> 00:36:54,030
of Dynamite or I can’t remember. Anyway, it doesn’t matter what the name

593
00:36:54,030 –> 00:36:57,550
of the movie is. It’s the latest entry into a

594
00:36:57,550 –> 00:37:01,070
genre of films that started, gosh,

595
00:37:01,150 –> 00:37:04,910
back in the 1950s with the movie directed by Stanley Kubrick,

596
00:37:04,910 –> 00:37:08,590
Dr. Strangelove, or how I Learned to Stop Worrying and Love the Bomb,

597
00:37:09,390 –> 00:37:13,230
where artists who are

598
00:37:13,230 –> 00:37:16,430
attempting to, to your point about emotion, who are attempting to

599
00:37:16,510 –> 00:37:19,480
emote into history, because that’s what artists do do

600
00:37:21,640 –> 00:37:25,080
and are attempting to get the viewer or the observer to engage

601
00:37:25,080 –> 00:37:28,040
emotionally with something, an idea that they have.

602
00:37:29,640 –> 00:37:33,280
There’s a long genre of films that began there and

603
00:37:33,280 –> 00:37:37,080
continues on through the Katherine Bigalow film, of which Oppenheimer is one of these films

604
00:37:39,000 –> 00:37:42,360
where to your point about World War 3,

605
00:37:42,840 –> 00:37:46,590
it’s that constant drum beat of warning that nuclear

606
00:37:46,590 –> 00:37:49,790
thermonuclear destruction is the worst thing that can possibly ever happen.

607
00:37:50,190 –> 00:37:53,390
It will destroy all of humanity, all civilization, everywhere,

608
00:37:53,950 –> 00:37:57,310
period, full stop. And it is, it is a

609
00:37:57,470 –> 00:38:00,750
drumbeat that runs through history.

610
00:38:01,710 –> 00:38:04,750
And to your point about forgetting, and this is sort of the last thing I

611
00:38:04,750 –> 00:38:06,710
want to say, and then we can go back to the book. But your point

612
00:38:06,710 –> 00:38:10,350
about forgetting, we’ve forgotten how bad World War

613
00:38:10,350 –> 00:38:13,230
II really was because we are distanced from it.

614
00:38:14,160 –> 00:38:18,000
And my concern also is that we have forgotten how

615
00:38:18,000 –> 00:38:20,800
bad a nuclear bomb can be.

616
00:38:22,240 –> 00:38:26,080
Thus you get crazy talk like in the last

617
00:38:26,080 –> 00:38:29,760
administration from certain people about

618
00:38:30,000 –> 00:38:33,840
arming, arming

619
00:38:33,840 –> 00:38:37,520
bases in Europe with nuclear tipped

620
00:38:37,920 –> 00:38:41,640
whatever, and if Russia does this thing

621
00:38:41,640 –> 00:38:45,080
or that thing in the Ukraine, we’re going to use those

622
00:38:45,080 –> 00:38:48,880
nukes. People in the last administration running around saying, saying

623
00:38:49,120 –> 00:38:52,080
nonsense like this. And of course the Russians are responding with,

624
00:38:52,960 –> 00:38:56,800
listen, we lost 60 some odd million people in

625
00:38:56,800 –> 00:39:00,480
what we call the Great Patriotic War and we didn’t miss

626
00:39:00,480 –> 00:39:03,200
a beat. You really want to go ahead, you really want to go ahead and

627
00:39:03,200 –> 00:39:06,880
pull that smoke wagon, you go right on ahead. We got nukes

628
00:39:06,880 –> 00:39:10,680
too. Yeah, exactly. And this is, this is the hubris to your point. This

629
00:39:10,680 –> 00:39:14,240
is reflect, even, just even allowing that to be said, number one,

630
00:39:14,240 –> 00:39:18,080
moves the Overton window on talking about nuclear war. But it also,

631
00:39:19,280 –> 00:39:22,720
it also reflects a level of hubris and arrogance

632
00:39:22,960 –> 00:39:26,360
that can only come from being disassociated in time and

633
00:39:26,360 –> 00:39:30,080
forgetting the lessons of history. Right. Yeah. And by the

634
00:39:30,080 –> 00:39:33,390
way, but just to circle this back into the whole purpose of this

635
00:39:33,550 –> 00:39:36,430
podcast in general, this

636
00:39:37,070 –> 00:39:40,830
nothing that we’re talking about excludes businesses, by the way. No,

637
00:39:41,150 –> 00:39:44,270
we, we’ve seen the same. Think again,

638
00:39:44,910 –> 00:39:48,710
current Environment not excluded. If you think

639
00:39:48,710 –> 00:39:52,110
about, like, I was actually, I literally just read something

640
00:39:52,270 –> 00:39:55,790
earlier, earlier today about this, which I found fascinating because

641
00:39:56,190 –> 00:39:59,790
those of us in the sales and marketing world have been talking recently about

642
00:40:00,390 –> 00:40:04,030
AI kind of plateauing a little bit. Like there’s, there’s been, like

643
00:40:04,030 –> 00:40:07,630
there was this major rush and there was all this talk about

644
00:40:07,630 –> 00:40:10,950
AI replacing people and taking jobs and all this other stuff,

645
00:40:11,270 –> 00:40:15,070
and yet I just saw a statistic, statistic. I just saw

646
00:40:15,070 –> 00:40:18,510
a statistic this morning that said of the

647
00:40:18,510 –> 00:40:22,030
Fortune 500 companies, the average company is only

648
00:40:22,030 –> 00:40:25,030
deploying AI at about a 2% efficiency rate.

649
00:40:26,000 –> 00:40:29,360
How many people you think they’re replacing at IBM because of this?

650
00:40:29,680 –> 00:40:33,240
How many people do you think at Amazon do you think they’re replacing because of

651
00:40:33,240 –> 00:40:36,520
it? It’s just not happening the way we expected it. Now here’s. Let me back

652
00:40:36,520 –> 00:40:39,840
up to my point a second ago. We saw this once before,

653
00:40:40,160 –> 00:40:43,840
folks, the dot com in the late 90s where

654
00:40:44,400 –> 00:40:48,080
investors were just throwing money upon money upon money at

655
00:40:48,480 –> 00:40:51,680
anything that had dot com at the end of it. And some of them lost

656
00:40:51,680 –> 00:40:54,830
their shirts. Some of them did okay. Some of them did all right. Yeah. The

657
00:40:54,830 –> 00:40:58,630
dot com boom didn’t. It didn’t. Just

658
00:40:58,630 –> 00:41:01,750
because there were a few winners did not mean that a,

659
00:41:02,550 –> 00:41:06,190
an intense amount of businesses failed. Yeah. Oh,

660
00:41:06,190 –> 00:41:09,910
yeah. So. So what did that teach us? Apparently

661
00:41:09,910 –> 00:41:13,630
nothing. Nothing. Because, because, because investors are

662
00:41:13,630 –> 00:41:17,030
starting to throw so much money at AI right now, I guarantee you

663
00:41:18,000 –> 00:41:20,880
a majority of them are going to lose their shirts and there are going to

664
00:41:20,880 –> 00:41:24,640
be a few, a select few that come out on top. And for some reason,

665
00:41:24,720 –> 00:41:28,360
we’re going to remember history as the dot com boom and the AI boom as

666
00:41:28,360 –> 00:41:32,120
being successful. How that happens to me is

667
00:41:32,120 –> 00:41:35,920
beyond. Like, again, business does the same stuff.

668
00:41:36,080 –> 00:41:39,800
Like, I feel, I, I’ve been telling people lately, I feel like

669
00:41:39,800 –> 00:41:43,280
I’m a veteran who’s been through like the fourth World War. Like, I feel like

670
00:41:43,280 –> 00:41:45,400
I’ve been, you know, I feel like I’ve been through four. I was going through

671
00:41:45,400 –> 00:41:48,400
the search revolution. Well, no, the first one was the Internet revolution to your point,

672
00:41:48,400 –> 00:41:51,500
with the dot com bubble. Oh, go back, go back one step further. The email.

673
00:41:51,580 –> 00:41:54,700
When email. Oh, God. Oh, yeah, email. Oh my God.

674
00:41:55,260 –> 00:41:58,020
Email’s gonna change the world. And we’re never gonna, we’re not gonna need a postal

675
00:41:58,020 –> 00:42:01,740
service anymore, right, Exactly. We have Amazon that delivers packages literally

676
00:42:01,740 –> 00:42:05,580
daily, like a day to people’s. Anyway, go ahead. Sorry. Yeah,

677
00:42:05,580 –> 00:42:08,860
yeah, sure. Amazon knows exactly where I live

678
00:42:08,940 –> 00:42:12,740
anyway. They can find me. They’ll have no Problem

679
00:42:12,740 –> 00:42:16,100
finding me, by the way. So does ups. Amazon knows where I live,

680
00:42:16,100 –> 00:42:19,780
UPS knows where I live, the postal service, everybody knows where I live now. Like,

681
00:42:19,780 –> 00:42:23,580
I can’t even get away from it at this point anyway. Good. But yeah, you

682
00:42:23,580 –> 00:42:26,540
got email. Take all that away. Email, Internet.

683
00:42:27,100 –> 00:42:30,780
Oh, God. Social media was going to revolutionize how we were going. Remember

684
00:42:30,780 –> 00:42:34,460
that. Like, we were all just going to live in these, like, these like Facebook

685
00:42:34,540 –> 00:42:37,980
pods and the Instagram Pod and the LinkedIn Pod, and we were never going to.

686
00:42:38,220 –> 00:42:41,900
And then, and then VR ar, which was a brief.

687
00:42:42,140 –> 00:42:45,930
Before that, it was, it was a 3D printing. Oh, yeah. Oh, I haven’t even

688
00:42:45,930 –> 00:42:49,570
gotten to that one yet. 3D printing. Crypto. Oh, my God.

689
00:42:49,570 –> 00:42:53,410
Yeah, crypto. And, and now, now here

690
00:42:53,410 –> 00:42:56,490
we are with AI. And I got to admit

691
00:42:57,290 –> 00:43:00,850
to your point, Tom, I get a little

692
00:43:00,850 –> 00:43:04,570
grizzled and gray when I, when I hear

693
00:43:04,570 –> 00:43:08,250
about it because I’m like. On the one hand, it’s very exciting

694
00:43:08,250 –> 00:43:11,850
because it’s a, it’s, it is a gold rush, exciting kind of, kind of thing.

695
00:43:11,850 –> 00:43:15,310
You can, you could fall into the emotion of it. Right. Because it gets to

696
00:43:15,310 –> 00:43:18,910
be emotive. It’s the bright shiny object thing. It’s the bright shiny object thing.

697
00:43:18,910 –> 00:43:22,110
Exactly. You just talked about. That’s exactly what they were. Now

698
00:43:22,670 –> 00:43:26,190
are some of them. Yeah, I’m not suggesting that they quote, unquote, failed.

699
00:43:26,350 –> 00:43:30,110
We won’t know that for 15 years, though. We won’t know for 15 years. Business

700
00:43:30,110 –> 00:43:33,830
lessons. And then, and then being more calculated about how we project, how

701
00:43:33,830 –> 00:43:37,510
we go forward, business. We learned nothing from all of those things you just said.

702
00:43:37,510 –> 00:43:41,070
The email, the, the dot com boom, the search engine, the,

703
00:43:41,230 –> 00:43:44,830
the, the, the 3D printing, the crypto, the

704
00:43:44,990 –> 00:43:48,750
social media. At all of those. There were

705
00:43:49,310 –> 00:43:52,910
thousands of companies involved in those things, most of which failed.

706
00:43:53,150 –> 00:43:56,989
Right. Like, but there will.

707
00:43:56,990 –> 00:44:00,350
And there will be an AI apocalypse. There will. Yeah, there will be an LLM

708
00:44:00,350 –> 00:44:04,070
apocalypse everywhere. Like, I was reading something of the other day in one of the

709
00:44:04,070 –> 00:44:07,830
startup. One of the startup newsletters that I read for the other project that you

710
00:44:07,830 –> 00:44:09,790
and I are on. And,

711
00:44:11,550 –> 00:44:15,190
and something like all of

712
00:44:15,190 –> 00:44:17,910
the. Oh yeah, I know what it was. All of the investor money, or the

713
00:44:17,910 –> 00:44:21,510
vast majority of investor money in Silicon Valley. If you don’t have something

714
00:44:21,510 –> 00:44:25,150
AI, you can’t get, you can’t even get into the door now.

715
00:44:25,470 –> 00:44:28,830
And I think about the event that we ran this week

716
00:44:29,070 –> 00:44:32,670
with the folks that we ran that with on that other

717
00:44:32,670 –> 00:44:36,080
project. And, and I mean,

718
00:44:36,800 –> 00:44:39,680
there’s way more things happening in the world. I Think we only had like what,

719
00:44:39,680 –> 00:44:43,400
two, maybe three out of the, out of the companies, out of the ones that

720
00:44:43,400 –> 00:44:47,160
we looked at that were. And I’m being on purposely oblique

721
00:44:47,160 –> 00:44:50,400
about this folks, but like three companies that we looked at

722
00:44:50,720 –> 00:44:54,240
that even had an AI play. The vast majority of everybody else

723
00:44:54,560 –> 00:44:58,160
is still trying to do a business the way you do a

724
00:44:58,240 –> 00:45:01,860
business. Now is there going to be an AI play built into that?

725
00:45:01,860 –> 00:45:05,460
Yeah, maybe. Probably because you got to get investors attention. But

726
00:45:05,540 –> 00:45:08,780
that’s insane to me because there’s just so many other businesses that are, that could

727
00:45:08,780 –> 00:45:12,460
operate in the world without, without an LLM. Yeah. And there’s so many other

728
00:45:12,460 –> 00:45:16,220
problems that we solve that LLMs can’t solve. So anyway, yeah, no,

729
00:45:16,220 –> 00:45:19,940
we haven’t learned. And there’ll be another bubble in 10 years. There’ll be another

730
00:45:19,940 –> 00:45:23,460
bubble, I guarantee. And it might be, honestly, it might be the humanoid robot bubble

731
00:45:23,460 –> 00:45:27,020
that I think might be the hardware bubble, that might be the hardware version, that

732
00:45:27,020 –> 00:45:30,660
might be the 3D printing version. The next bubble you.

733
00:45:31,040 –> 00:45:34,840
And so Elon, you know why, Jason? Because the more things change, the more.

734
00:45:34,840 –> 00:45:38,560
Things and the more they stick. There we go. There it

735
00:45:38,560 –> 00:45:42,320
is. There it is. Back to the book. Back

736
00:45:42,320 –> 00:45:45,920
to why don’t we learn from History by B.H. liddell Hart.

737
00:45:46,320 –> 00:45:50,040
This is an open source book, by the way. You can get it online for

738
00:45:50,040 –> 00:45:53,840
free. So. But the version that I’m reading has a

739
00:45:53,840 –> 00:45:57,520
yellow cover and it was edited and with an introduction by Gills

740
00:45:57,520 –> 00:46:01,070
Lauren. But you can grab this book anywhere

741
00:46:01,310 –> 00:46:04,430
online. This is, this is definitely an open source book and I would encourage you

742
00:46:04,510 –> 00:46:07,870
if you are in business or you’re in leadership

743
00:46:08,110 –> 00:46:09,790
or you are in tech,

744
00:46:11,790 –> 00:46:15,310
especially if you’re part of one of those LLM

745
00:46:15,310 –> 00:46:18,750
driven AI startups. I strongly recommend

746
00:46:18,990 –> 00:46:22,270
reading this book. All right, back to the book. So

747
00:46:22,670 –> 00:46:26,510
let’s look at the importance of keeping promises and the importance of care about making

748
00:46:26,510 –> 00:46:30,080
promises. And, and I want to talk about, with Tom, about something that he

749
00:46:30,080 –> 00:46:33,280
mentioned that ties into how we teach history.

750
00:46:34,880 –> 00:46:38,240
Back to the book on the importance of keeping promises.

751
00:46:38,480 –> 00:46:41,440
Civilization is built on the practice of keeping promises.

752
00:46:42,080 –> 00:46:45,840
It may not sound a high attainment, but if trust in its observance

753
00:46:45,840 –> 00:46:49,360
should be shaken, the whole structure cracks and sinks.

754
00:46:49,920 –> 00:46:52,960
Any constructive effort in all human relations, personal,

755
00:46:53,570 –> 00:46:57,370
political and commercial depend on being able

756
00:46:57,370 –> 00:47:01,010
to depend on promises. By the way,

757
00:47:01,010 –> 00:47:04,570
pause for just a minute. That’s genius. I’ve never heard

758
00:47:04,570 –> 00:47:08,010
that. I’ve never heard an argument for high trust, a high trust

759
00:47:08,010 –> 00:47:11,850
society laid out as succinctly as is laid out in those

760
00:47:11,850 –> 00:47:15,170
three Sentences. That’s brilliant. That’s brilliant writing.

761
00:47:15,810 –> 00:47:19,210
Back to the book. This truth has a reflection on the question of

762
00:47:19,210 –> 00:47:22,890
collective security among nations and on the lessons of history in regards to that

763
00:47:22,890 –> 00:47:26,640
subject. In the years before the war, the charge was constantly brought. And

764
00:47:26,640 –> 00:47:30,440
by the way, the war he’s talking about, just pause again, is World War II.

765
00:47:30,440 –> 00:47:33,440
But sometimes he’s also talking about World War I. So just you have to think

766
00:47:33,440 –> 00:47:36,880
about those both in concert with each other. All right. In the years before the

767
00:47:36,880 –> 00:47:40,360
war, the charge was constantly brought that its supporters were courting the risk of war

768
00:47:41,400 –> 00:47:45,160
by their exaggerated respect for covenants. Although

769
00:47:45,160 –> 00:47:48,680
they may have been fools in disregarding the conditions necessary for the

770
00:47:48,680 –> 00:47:52,430
effective fulfillment of pledges, they at least show themselves men

771
00:47:52,430 –> 00:47:56,150
of honor. I have that double underlined, by the way, and in the long

772
00:47:56,150 –> 00:47:59,910
view of more fundamental common sense than those who

773
00:47:59,910 –> 00:48:03,750
argued that we should give aggressors a free hand so long as they left us

774
00:48:03,750 –> 00:48:07,590
alone. History has shown repeatedly

775
00:48:07,750 –> 00:48:11,510
that the hope of buying safety in this way is the greatest of

776
00:48:11,510 –> 00:48:15,350
delusions. The importance of care about

777
00:48:15,350 –> 00:48:19,170
making promises. It is immoral to make promises

778
00:48:19,170 –> 00:48:22,610
that one cannot in practice fulfill in the sense that the

779
00:48:22,610 –> 00:48:26,130
recipient expects on the ground in

780
00:48:26,130 –> 00:48:29,850
1939. This is ahead of World War II. I question the

781
00:48:29,850 –> 00:48:33,570
underlying morality of the Polish guarantee as well as its practicality. If

782
00:48:33,570 –> 00:48:37,250
the Poles had realized the military inability of Britain and France to save them from

783
00:48:37,250 –> 00:48:40,690
defeat and of what such a defeat would mean to them individually and

784
00:48:40,690 –> 00:48:44,320
collectively, it is unlikely that they would have shown such a stubborn opposition to

785
00:48:44,320 –> 00:48:48,120
Germany’s originally modest demands for Danzig and a

786
00:48:48,120 –> 00:48:51,680
passage through the Corridor, since it was obvious to me that they were bound to

787
00:48:51,680 –> 00:48:55,320
lose those points and even much more in the event of a conflict.

788
00:48:55,480 –> 00:48:59,320
It seemed to me wrong on our part to make promises that we

789
00:48:59,320 –> 00:49:03,040
that were bound to encourage false hopes. It

790
00:49:03,040 –> 00:49:05,920
also seemed to me that any such promises were the most certain way to produce

791
00:49:05,920 –> 00:49:09,690
war. Because the inevitable provocativeness of guaranteeing at such

792
00:49:09,690 –> 00:49:13,530
a moment of tension an area which we had hitherto treated as outside our

793
00:49:13,530 –> 00:49:17,210
sphere of interest, because of the manifest temptation which the

794
00:49:17,210 –> 00:49:20,850
guarantee offered to a military minded people like the Germans, to show how

795
00:49:20,850 –> 00:49:24,490
fatuously impractical our guarantee was, and because of its

796
00:49:24,490 –> 00:49:27,850
natural effect on stiffening the attitude of a people, the

797
00:49:27,850 –> 00:49:31,450
Poles who had always shown themselves exceptionally intractable in

798
00:49:31,450 –> 00:49:34,450
negotiating a reasonable settlement of any issue

799
00:49:38,220 –> 00:49:42,020
and historian could not help seeing certain parallels between the long standing aspect of

800
00:49:42,020 –> 00:49:45,780
the Polish German situation and that between Britain and the Boer Republics 40

801
00:49:45,780 –> 00:49:49,460
years earlier, and remembering the effects on us of the attempts of the other European

802
00:49:49,460 –> 00:49:52,860
powers to induce or coerce us into negotiating a settlement with the Boers.

803
00:49:53,340 –> 00:49:57,020
If our own reaction then had been so violent, it could hardly be expected that

804
00:49:57,020 –> 00:50:00,460
the reaction of a nation filled with an even more bellicose spirit would be less

805
00:50:00,460 –> 00:50:04,200
violent, especially as the attempt to compel negotiation was backed by

806
00:50:04,200 –> 00:50:08,000
an actual promise of making war if Poland felt moved to

807
00:50:08,000 –> 00:50:11,040
resist the German conditions.

808
00:50:15,120 –> 00:50:18,600
That is a brilliant piece of analysis. That’s why I’m reading this. That is a

809
00:50:18,600 –> 00:50:22,080
brilliant piece of analysis of the psychology

810
00:50:22,160 –> 00:50:26,000
of nation states, which we don’t often talk about. Like, we try

811
00:50:26,000 –> 00:50:29,790
to pretend that nations are these, and

812
00:50:29,790 –> 00:50:33,390
maybe we do it less so now than we have in the past in history.

813
00:50:33,390 –> 00:50:37,070
We try to pretend that nation states are somehow this amorphous collection of

814
00:50:37,070 –> 00:50:40,790
ideas. But nation states actually do have their own psychology and

815
00:50:40,790 –> 00:50:44,510
their own character. And history reveals that. And

816
00:50:44,510 –> 00:50:47,190
Liddell Hart there is brilliant

817
00:50:48,150 –> 00:50:51,910
in basically saying, if you’re going to make a promise,

818
00:50:51,910 –> 00:50:55,620
keep it, but understand the character as a nation

819
00:50:55,620 –> 00:50:58,980
state of the nation state

820
00:50:59,220 –> 00:51:02,540
that you are making the promise to understand their

821
00:51:02,540 –> 00:51:06,020
character, study them, examine them,

822
00:51:06,660 –> 00:51:10,100
don’t just. And then this gets to our current geopolitical

823
00:51:10,580 –> 00:51:14,420
climate. Don’t just, like in the case of NATO, sign a piece

824
00:51:14,420 –> 00:51:18,180
of paper 80 years ago and then just sort of pretend like everything’s

825
00:51:18,180 –> 00:51:22,030
the same as it was 80 years ago. And nation

826
00:51:22,030 –> 00:51:25,150
states, characters change just like people.

827
00:51:25,550 –> 00:51:28,710
France now is

828
00:51:28,710 –> 00:51:32,390
significantly different than they were at

829
00:51:32,390 –> 00:51:35,230
the back end of World War II after being

830
00:51:35,470 –> 00:51:38,670
humiliated by the Germans. They’re significantly different.

831
00:51:39,950 –> 00:51:42,990
They can pay for the protection of their own

832
00:51:43,150 –> 00:51:46,110
continent, by the way, this is all that Trump is saying, by the way. He’s

833
00:51:46,110 –> 00:51:49,890
been saying this since his first administration. Maybe, maybe you

834
00:51:49,890 –> 00:51:53,210
could pay for the protection of your own continent,

835
00:51:53,370 –> 00:51:57,010
because maybe the character of France as a nation state

836
00:51:57,010 –> 00:52:00,690
has changed. Maybe the character of Germany as a nation state has changed. Maybe the

837
00:52:00,690 –> 00:52:04,370
character of Sweden and Finland and at all has

838
00:52:04,370 –> 00:52:07,530
changed since World War II.

839
00:52:08,090 –> 00:52:11,850
Maybe we don’t just have to keep honoring these guarantees

840
00:52:11,850 –> 00:52:15,500
in perpetuity. Now, the other

841
00:52:15,500 –> 00:52:19,140
idea in there, which I find to be interesting, is this idea of being a

842
00:52:19,140 –> 00:52:22,900
person of honor. And this gets to diplomacy, right? And

843
00:52:22,900 –> 00:52:26,580
so if civilization is built on promises, and

844
00:52:27,140 –> 00:52:30,820
one of the big lessons that Liddell Hart learned from World War I

845
00:52:31,300 –> 00:52:32,260
was how,

846
00:52:35,060 –> 00:52:38,860
how fatuous, to a certain degree, diplomacy

847
00:52:38,860 –> 00:52:42,640
really was in the run up to that war. And so

848
00:52:42,640 –> 00:52:46,400
that mistake where, interestingly enough, has been corrected, we now have more

849
00:52:46,400 –> 00:52:50,160
diplomatic venues to get more people to talk as leaders of

850
00:52:50,160 –> 00:52:53,520
nation states to talk than ever before in the History of the world. It’s kind

851
00:52:53,520 –> 00:52:56,760
of insane that we have a un. That’s actually kind of nuts in the history

852
00:52:56,760 –> 00:53:00,600
of the world, and we don’t appreciate how. Bananas in pajamas, that is.

853
00:53:00,600 –> 00:53:04,440
We just don’t. We just don’t. And the first place that we don’t appreciate

854
00:53:04,440 –> 00:53:08,190
that, I think, is in how we teach history to

855
00:53:08,190 –> 00:53:11,190
the next generation. So Tom hit on this,

856
00:53:12,630 –> 00:53:16,390
and this is an interesting point. The biggest challenge in teaching

857
00:53:16,390 –> 00:53:20,110
history is getting people to care

858
00:53:20,110 –> 00:53:23,030
about it who were born after all of that was done.

859
00:53:24,950 –> 00:53:28,550
And so that’s the question to Tom. How do we get people to care?

860
00:53:29,270 –> 00:53:32,870
How do we teach history to people who were born after

861
00:53:32,870 –> 00:53:36,530
all of that was over? Well, and before

862
00:53:36,530 –> 00:53:40,250
you. Before you get to that challenge, I think what happens even

863
00:53:40,250 –> 00:53:44,010
before that is even if you. Let’s say. Let’s say you don’t get

864
00:53:44,010 –> 00:53:47,290
them excited about it, but you at least get them to read it, there’s another

865
00:53:47,290 –> 00:53:51,090
bias that happens that psychology proves that there’s a. An

866
00:53:52,290 –> 00:53:56,130
in. What’s it. What the hell would they. I forget how they worded it. I

867
00:53:56,130 –> 00:53:59,730
don’t remember the title they gave it, but it’s like a. It’s like an obvious.

868
00:54:00,530 –> 00:54:04,050
An obvious ton ability factor. My point is,

869
00:54:04,450 –> 00:54:07,370
so when you. When you read. When you read the history of World War II

870
00:54:07,370 –> 00:54:11,090
and you see, oh, like, oh, you watched this guy come into power and

871
00:54:11,090 –> 00:54:14,770
nobody really wanted. And then it’s almost like you see the inevitability

872
00:54:15,170 –> 00:54:19,010
without somebody actually teaching it to you. Well, we’re not speaking

873
00:54:19,010 –> 00:54:22,170
German right now, so obviously we won, right? Like, so then, like, there’s like an

874
00:54:22,170 –> 00:54:26,010
obvious factor that happens in reading history that. That

875
00:54:26,010 –> 00:54:29,690
gives you a bias that you already know the outcome. So

876
00:54:29,920 –> 00:54:33,760
another reason why today that. That whole history repeats itself kind of

877
00:54:33,760 –> 00:54:37,400
thing is because we. We have this expectation of

878
00:54:37,400 –> 00:54:40,760
inevitability, right? That because history showed us this

879
00:54:40,760 –> 00:54:44,520
inevitability, now we’re going to expect this inevitability. And whether it happens or

880
00:54:44,520 –> 00:54:47,520
not, which, by the way, it usually does.

881
00:54:48,160 –> 00:54:51,360
That’s why we keep saying the more things stay the same,

882
00:54:51,760 –> 00:54:55,520
but. But we keep making the same mistakes over and over again

883
00:54:55,600 –> 00:54:58,560
because we take the inevitability factor into it

884
00:54:59,130 –> 00:55:02,570
subconsciously. So to your point about getting the next

885
00:55:02,570 –> 00:55:06,250
generation to learn to care

886
00:55:06,250 –> 00:55:09,610
about it, that’s another reason why you and I get

887
00:55:09,690 –> 00:55:13,370
involved so heavily in conversations about film. It’s a

888
00:55:13,370 –> 00:55:17,090
media in which we can get the next generation to understand and learn from some

889
00:55:17,090 –> 00:55:20,930
of those historical events. You produce Band

890
00:55:20,930 –> 00:55:24,330
of Brothers or Saving Private Ryan or.

891
00:55:24,720 –> 00:55:27,920
And you Get a the next generation to watch that movie. And they go, oh

892
00:55:27,920 –> 00:55:31,080
my God. And they think it’s just cinematography. And then they realize it’s a real

893
00:55:31,080 –> 00:55:34,880
thing, it actually happened. They’re like, oh my God, I should go

894
00:55:34,880 –> 00:55:38,680
learn about World War. Like we have mechanisms that we can use

895
00:55:38,680 –> 00:55:41,920
and pull the lever on to get the next generation to care.

896
00:55:43,200 –> 00:55:47,000
Well, we did. I’m not sure they care anymore. They don’t watch it. They don’t

897
00:55:47,000 –> 00:55:50,160
watch all that many more. Maybe we should put them in TikTok videos. Maybe that’s

898
00:55:50,160 –> 00:55:53,830
what it is. Just short 30 second clips of what happened with

899
00:55:53,830 –> 00:55:57,390
World War II. And maybe then we’ll get some the next generation to care about

900
00:55:57,390 –> 00:55:59,710
history the. Way that we do. But

901
00:56:00,910 –> 00:56:04,630
kidding aside though, but that, that’s. We, we’ve got to

902
00:56:04,630 –> 00:56:08,390
stop. We’ve got to get it out of the. We, we have to. We. If

903
00:56:08,390 –> 00:56:11,790
we can change the way the mechanisms in which we teach,

904
00:56:12,270 –> 00:56:16,110
we will be able to touch the hearts of the next generation. Because

905
00:56:16,770 –> 00:56:20,130
when you and I were growing up, it was books, it was

906
00:56:20,450 –> 00:56:24,050
our imagination that reading these books and, and how

907
00:56:24,050 –> 00:56:27,850
they impacted us. This generation doesn’t really care

908
00:56:27,850 –> 00:56:31,570
so much about books, but they care about media. And so like some of these,

909
00:56:31,570 –> 00:56:35,090
like video media and all that. So if we can use that,

910
00:56:36,210 –> 00:56:39,170
all we need to do is get their attention. Once we get their attention and

911
00:56:39,170 –> 00:56:42,520
they care about it, then they’ll go back and, and figure out the rest. Right.

912
00:56:42,520 –> 00:56:46,120
So there’s a lot of talk about teachers in this country, particularly the five years

913
00:56:46,120 –> 00:56:49,680
after Covid. Right. Because you know, we all went on lockdown and then,

914
00:56:50,320 –> 00:56:53,760
you know, everybody who’s a parent who had their kid in the public school system

915
00:56:54,480 –> 00:56:57,919
kind of looked over the shoulder. This is kind of what happened. Actually, not kind

916
00:56:57,919 –> 00:57:01,120
of. This is what happened. Everybody who had a kid in public school

917
00:57:01,680 –> 00:57:05,040
all of a sudden had their kids at home learning off of a laptop

918
00:57:05,360 –> 00:57:09,170
and looking over the shoulder and actually for the

919
00:57:09,170 –> 00:57:10,170
first time in

920
00:57:13,290 –> 00:57:16,890
a long time in America, actually seeing for

921
00:57:17,610 –> 00:57:21,450
four, six, eight hours a day what a teacher is actually teaching

922
00:57:21,450 –> 00:57:25,050
their kids. This is why things have started to crack apart with the K through

923
00:57:25,050 –> 00:57:28,850
12 system, which by the way, the unions were

924
00:57:28,850 –> 00:57:32,610
the ones that insisted on lockdowns and worked

925
00:57:32,610 –> 00:57:36,170
in concert with the government and insisted on and still insist on lock

926
00:57:36,170 –> 00:57:39,820
on, on. What are you teaching students

927
00:57:39,820 –> 00:57:43,660
from, from home in many areas, particularly urban areas of

928
00:57:43,660 –> 00:57:47,340
our country. Even though we know three point about statistical

929
00:57:47,340 –> 00:57:50,220
data, we have good research now that

930
00:57:51,340 –> 00:57:54,460
kids lose a step when they are taught virtually

931
00:57:55,820 –> 00:57:59,340
particularly if they are being switched from being taught in person to

932
00:58:00,060 –> 00:58:03,840
being taught virtually, we know this. This is a fact. Okay?

933
00:58:04,240 –> 00:58:08,080
It’s like that two plus two thing. It just is. This is

934
00:58:08,080 –> 00:58:11,840
the thing. Okay? We could argue about why. We could argue

935
00:58:11,840 –> 00:58:14,840
about what the inputs are. All that, that’s fine. But you can’t argue with the

936
00:58:14,840 –> 00:58:18,680
fact, okay? We talk a lot about

937
00:58:18,680 –> 00:58:21,880
teachers in this country and how much teachers get paid and we lament and we

938
00:58:21,880 –> 00:58:24,400
wring our hands and all of that. I don’t want to get into any of

939
00:58:24,400 –> 00:58:27,120
that. Instead, I want to talk about something that’s a little bit more

940
00:58:28,320 –> 00:58:31,960
egregious, I think, which is the

941
00:58:31,960 –> 00:58:35,600
fact that. And this goes to your point about students and TikTok,

942
00:58:36,320 –> 00:58:40,120
I would agree. However, I

943
00:58:40,120 –> 00:58:43,760
think students. I agree. And I think that students

944
00:58:43,760 –> 00:58:47,520
care about ideas the same way students always

945
00:58:47,600 –> 00:58:51,280
have. And what we lack are

946
00:58:51,280 –> 00:58:54,560
teachers willing to engage

947
00:58:54,640 –> 00:58:56,560
passionately with ideas

948
00:58:58,980 –> 00:59:02,780
about the subject matter they are hired to teach. And

949
00:59:02,780 –> 00:59:06,420
this is hugely important with history, right? So, for instance,

950
00:59:07,380 –> 00:59:11,220
if I am going to teach. No. If

951
00:59:11,220 –> 00:59:15,020
I’m going to going to attend a class on an area that

952
00:59:15,020 –> 00:59:18,780
you know a lot about in history, Native American history, it

953
00:59:18,780 –> 00:59:21,620
doesn’t matter whether I agree with your conclusions or not about that history.

954
00:59:22,420 –> 00:59:26,140
None of that matters. None of that matters. I’m

955
00:59:26,140 –> 00:59:29,700
going there to tie into your passion about

956
00:59:29,700 –> 00:59:33,300
that because your passion is going to

957
00:59:33,300 –> 00:59:37,060
either make me care more or it’s going to make

958
00:59:37,060 –> 00:59:40,580
me disagree. But either way, I will give you attention

959
00:59:40,820 –> 00:59:44,300
because of the passion that I see coming from you. This is the whole plot

960
00:59:44,300 –> 00:59:48,020
of the movie Dead Poets Society. How do you make poetry and Shakespeare

961
00:59:48,020 –> 00:59:51,460
interesting? Right? Well, you make poetry and Shakespeare interesting by having a

962
00:59:51,460 –> 00:59:55,290
dynamic person teaching poetry and Shakespeare. That’s how you make it interesting.

963
00:59:56,160 –> 00:59:58,800
Same thing with history. And yet

964
01:00:00,240 –> 01:00:04,000
when you hear teachers talk about teaching history, it’s

965
01:00:04,000 –> 01:00:07,360
either one of two poles. You either have a person who’s got a personal,

966
01:00:07,920 –> 01:00:11,360
as a teacher, got a personal bugaboo of a thing that they’re drilling in on

967
01:00:11,680 –> 01:00:15,240
that’s interesting to them, but it’s not interesting maybe to 99 of the people that

968
01:00:15,240 –> 01:00:19,080
they’re teaching. You see this in colleges a lot. Or you get

969
01:00:19,080 –> 01:00:21,440
a person who is in a K through 12 space where

970
01:00:23,190 –> 01:00:26,790
they would love to be passionate about history, but they’re fighting uphill

971
01:00:26,790 –> 01:00:30,590
against a lack of comprehension, a lack of preparation, kids

972
01:00:30,590 –> 01:00:34,230
falling behind the policies of the system of the blah, blah, blah, blah,

973
01:00:34,230 –> 01:00:37,829
blah. And they would love to be passionate, but they’re not going to try to

974
01:00:37,829 –> 01:00:41,190
break the system because they’re getting paid $30,000 a year. They just want their pension

975
01:00:41,190 –> 01:00:43,430
at the end of it. They’re just not going to try to break the system.

976
01:00:43,510 –> 01:00:46,830
And by the way, they know that the teacher. Because I know teachers in K

977
01:00:46,830 –> 01:00:50,110
through 12 system over the course of my life. They know that the teacher down

978
01:00:50,110 –> 01:00:53,340
the hall who’s supposed to be teaching English comprehension or

979
01:00:53,580 –> 01:00:57,220
spelling or whatever, is it doing their job. They know

980
01:00:57,220 –> 01:01:00,900
this. They know, they know. And so by the time that

981
01:01:00,900 –> 01:01:03,900
kid shows up in sixth period for history,

982
01:01:05,260 –> 01:01:08,860
they’re fighting uphill against the previous, you know, the other previous five

983
01:01:08,860 –> 01:01:12,340
periods of nonsense. And so they’re just trying to get through the period and get

984
01:01:12,340 –> 01:01:16,020
the kid out. And those are the two poles

985
01:01:16,020 –> 01:01:17,660
the teachers are fighting a pill against.

986
01:01:19,790 –> 01:01:23,390
And history is so critical for our time. Now, I would argue that it’s

987
01:01:23,390 –> 01:01:26,910
probably as critical or more critical even than English than

988
01:01:26,910 –> 01:01:29,630
English, right? Even being able to speak and write well.

989
01:01:31,550 –> 01:01:34,269
Because history is the place where, for

990
01:01:35,790 –> 01:01:39,310
it’s the platform where our political battles are being fought.

991
01:01:39,870 –> 01:01:43,710
It’s the place where ideologies are being

992
01:01:43,710 –> 01:01:46,920
made and are being rendered. It’s even a place where, oddly enough,

993
01:01:47,000 –> 01:01:50,360
identities are being formed of all kinds.

994
01:01:51,160 –> 01:01:55,000
And teachers have a huge, huge responsibility, particularly

995
01:01:55,000 –> 01:01:57,960
in the K12 system. Have a huge responsibility. And I don’t know how you fix

996
01:01:57,960 –> 01:01:58,840
the two poles problem.

997
01:02:02,520 –> 01:02:05,000
Well, and to your point, I think,

998
01:02:07,960 –> 01:02:11,690
I think history is too big of a subject for

999
01:02:11,690 –> 01:02:15,210
that to be the, the thing that they use to try to break the polls.

1000
01:02:15,210 –> 01:02:17,890
Right? So. Because yeah, like

1001
01:02:19,170 –> 01:02:22,610
if you just think about. So I remember, I remember going through The K through

1002
01:02:22,610 –> 01:02:25,650
12 system, whatever in public schools. And I remember,

1003
01:02:26,690 –> 01:02:30,050
I remember like in. So we had to take four years when I got to

1004
01:02:30,050 –> 01:02:33,570
high school. There were four years of high school, right? 9, 10, 11, 12 Y.

1005
01:02:33,570 –> 01:02:37,370
And the, and they, they. This is not a, a

1006
01:02:37,370 –> 01:02:40,960
college environment where you get to select your, you know, your

1007
01:02:40,960 –> 01:02:44,080
course, your courses, right? So they. You dictated

1008
01:02:44,960 –> 01:02:48,800
year one. Year one was world history, year two was US

1009
01:02:48,800 –> 01:02:51,360
history, year three was.

1010
01:02:54,320 –> 01:02:57,960
US. History post World War I, I think it was. So it was

1011
01:02:57,960 –> 01:03:01,440
US history one which was like from the beginning of the country,

1012
01:03:02,400 –> 01:03:05,720
you know, 1500s, uh, up until, you know,

1013
01:03:05,720 –> 01:03:09,160
1918, World War I. And then it was

1014
01:03:09,560 –> 01:03:12,760
history from World War u. S history. Two

1015
01:03:13,240 –> 01:03:16,880
was World War I to present day. And then the, the

1016
01:03:16,880 –> 01:03:20,720
four. The fourth year was, was some sort of like there

1017
01:03:20,720 –> 01:03:24,280
was some nonsense. I say nonsense, but it was because it was like, it was

1018
01:03:24,280 –> 01:03:28,040
like theoretical history. Like it was like a future thing. Like we were, we were

1019
01:03:28,040 –> 01:03:31,830
trying to predict like his history predictions or something. Some craziness like that.

1020
01:03:31,830 –> 01:03:34,630
Which by the way, not a single one of us will write about anything. But

1021
01:03:34,630 –> 01:03:38,310
whatever. But because we’re not driving

1022
01:03:38,310 –> 01:03:42,110
flying cars right now. Yeah. We literally just talked

1023
01:03:42,110 –> 01:03:45,590
about this before we hit the record button. We’re not flying cars. We’re not teleporting

1024
01:03:45,590 –> 01:03:49,430
anywhere. We don’t have the Galactic Empire, the Galactic Federation

1025
01:03:49,430 –> 01:03:53,070
of Planets. We don’t have any of that yet. And of my graduating class in,

1026
01:03:53,310 –> 01:03:56,910
in whatever night, early night, whatever, the 19, whatever,

1027
01:03:56,910 –> 01:04:00,610
whatever. We all thought that was happening because we watched

1028
01:04:00,610 –> 01:04:04,290
the Jetsons growing up. Right. Whatever. Anyway, but, but

1029
01:04:04,290 –> 01:04:08,010
anyway, to your, to your point, the. When they, when they

1030
01:04:08,010 –> 01:04:11,810
said to us, like, you know, freshman year is

1031
01:04:11,810 –> 01:04:15,370
going to be world history. And we were like, okay, cool. Whose world?

1032
01:04:15,690 –> 01:04:18,890
What world? Like, because. And they, they went through these,

1033
01:04:19,370 –> 01:04:23,010
they went through these timelines so fast. Yeah. You

1034
01:04:23,010 –> 01:04:26,480
couldn’t get passionate about any of them even if you wanted to. Right. Like

1035
01:04:26,640 –> 01:04:30,320
I. Most of what I learned about in world history my freshman year, I had

1036
01:04:30,320 –> 01:04:33,160
to go reback and I had to go back and revisit things that I thought

1037
01:04:33,160 –> 01:04:36,280
were even a little bit interesting to see if I was curious about him. The

1038
01:04:36,280 –> 01:04:40,040
main dynasty in China, the Samurai. All the like,

1039
01:04:40,040 –> 01:04:43,120
we, and we talked about some of these things on your podcast, but I had

1040
01:04:43,120 –> 01:04:46,880
to go backwards and after the fact to see, like, was that

1041
01:04:46,880 –> 01:04:50,560
worth my time and effort to dig a little deeper because they had.

1042
01:04:50,640 –> 01:04:54,310
It’s so surface level and to your point, if I’m a history

1043
01:04:54,310 –> 01:04:57,630
teacher, for example. To your point. Exactly.

1044
01:04:58,510 –> 01:05:02,310
And I would make the argument that being a teacher in this country

1045
01:05:02,310 –> 01:05:06,070
is the easiest job on the planet. It is the easiest job on the planet

1046
01:05:06,070 –> 01:05:09,110
because number one, you’re not doing it at all unless you

1047
01:05:09,110 –> 01:05:12,750
absolutely love it and you really want to be there

1048
01:05:12,830 –> 01:05:16,550
because as you know and me being on this. But I love history and there’s

1049
01:05:16,550 –> 01:05:20,280
certain parts of history that I love way more than others. I could never

1050
01:05:20,280 –> 01:05:23,720
be a history teacher because I would want to teach just that. And they won’t

1051
01:05:23,720 –> 01:05:27,200
let me. Right. Right. Because I have to follow some syllabus and.

1052
01:05:27,200 –> 01:05:30,800
Etc. So I have to love teaching more than

1053
01:05:30,800 –> 01:05:33,960
anything else in my being in order to be a history teacher.

1054
01:05:34,200 –> 01:05:38,040
Yep. Which I don’t. Which is why I’m not a history teacher. So

1055
01:05:38,040 –> 01:05:41,080
if I, if I were, that’s why I say, and by the way, any teacher

1056
01:05:41,080 –> 01:05:43,400
out there, you don’t have to come find me to kill me. I’m not suggesting

1057
01:05:43,400 –> 01:05:47,010
you have an easy job. That is not what I said. I didn’t say you’re

1058
01:05:47,330 –> 01:05:50,250
the function of your job is easy. I’m just saying it’s the easiest job in

1059
01:05:50,250 –> 01:05:53,730
the planet because you’re not doing it unless you love it. And they always

1060
01:05:53,810 –> 01:05:56,450
say if you love what you do, you never work a day in your life.

1061
01:05:56,450 –> 01:06:00,290
So there you go, Ex. So you have the easiest job

1062
01:06:00,290 –> 01:06:03,050
if you, you’re doing it because you love it. And if you love it, then

1063
01:06:03,050 –> 01:06:06,650
it’s not really work. And if you, if it’s not really work, then stop complaining

1064
01:06:06,650 –> 01:06:10,010
about it anyway. But no, I’m kidding. I’m totally, totally

1065
01:06:10,010 –> 01:06:13,730
kidding. No, no, it’s fine. I, I already nuked. I already nuked teachers already.

1066
01:06:13,730 –> 01:06:17,390
I nuked the unions. I’m in far more trouble than you are. Oh,

1067
01:06:17,390 –> 01:06:19,870
yeah, because you went after the union. I’m just going after the individual. Oh, yeah,

1068
01:06:20,030 –> 01:06:23,790
yeah, yeah, yeah, right. More than me. They’re gonna. Yeah, Randy Baumgarten’s gonna

1069
01:06:23,790 –> 01:06:27,590
find me in about 10 seconds. Anyway, sorry, go ahead, keep going. Yeah, no, but,

1070
01:06:27,590 –> 01:06:30,750
but anyway. But the point, like, to your point, it’s, it’s part of the problem

1071
01:06:30,750 –> 01:06:34,190
that we have with, with, especially with subject matter like this book,

1072
01:06:34,590 –> 01:06:37,150
is that the subject matter is so

1073
01:06:38,190 –> 01:06:41,270
intense, so wide, so varied, so

1074
01:06:41,270 –> 01:06:45,040
expansive, that there’s no way that a single individual is

1075
01:06:45,040 –> 01:06:48,160
going to be able to teach all of it. And if we focus on one

1076
01:06:48,160 –> 01:06:52,000
thing or another, it’s, it’s not going to be. It’s, it’s so isolating

1077
01:06:52,000 –> 01:06:55,360
that there. You’re not understanding all of the other factors involved. For

1078
01:06:55,360 –> 01:06:59,200
example, and to your, to your point about, like, if you really

1079
01:06:59,200 –> 01:07:02,280
go deep, if you could literally be a teacher about

1080
01:07:02,760 –> 01:07:06,600
the preamble to World War II, never mind the actual war, but

1081
01:07:06,600 –> 01:07:10,200
the, the events that led up to World War II, you could teach an entire

1082
01:07:10,280 –> 01:07:13,040
semester college syllabus on just that

1083
01:07:14,080 –> 01:07:17,840
if somebody was willing to take it. But they don’t

1084
01:07:17,840 –> 01:07:21,600
find that valuable enough to take that course, like that one course about

1085
01:07:21,920 –> 01:07:25,600
the, you know, the, the. The era between World War I and World

1086
01:07:25,600 –> 01:07:29,040
War II that basically caused World War II to happen in the first place.

1087
01:07:29,360 –> 01:07:33,120
All the geopolitical landscaping that happened, all the political power shifting that

1088
01:07:33,120 –> 01:07:36,800
happened, all that stuff in one instance. And by the

1089
01:07:36,800 –> 01:07:40,410
way, not just the time frame, forget about just the time frame,

1090
01:07:40,490 –> 01:07:44,250
but what was happening in Germany was different than what was happening

1091
01:07:44,250 –> 01:07:47,730
in the rest of Europe that was different. That was happening in China. Did China

1092
01:07:47,730 –> 01:07:50,490
have anything to do with this? Did the United States being so,

1093
01:07:51,690 –> 01:07:55,490
so arrogant and egotistical, thinking that they were too strong to even

1094
01:07:55,490 –> 01:07:59,210
bother with it in the first place. What was that geopolitical landscape?

1095
01:07:59,210 –> 01:08:02,970
How was that economy working for you, by the way? Black. Black Friday was in

1096
01:08:02,970 –> 01:08:06,680
there. In. Involved in that as well. Like there. There. There’s so many different

1097
01:08:06,680 –> 01:08:10,480
components to what just brought us to war. One that. Sorry.

1098
01:08:10,480 –> 01:08:14,040
To. To World War II, that you could not be an expert at all of

1099
01:08:14,040 –> 01:08:17,840
it. You just can’t. So. Right. That’s another reason why we don’t learn from the

1100
01:08:17,840 –> 01:08:21,480
mistakes of history. It’s too vast and we just don’t

1101
01:08:21,480 –> 01:08:25,200
get deeply involved in enough to, like, actually pluck out

1102
01:08:25,200 –> 01:08:28,920
the real lessons. So this goes. But this goes

1103
01:08:28,920 –> 01:08:31,840
to. Okay, so this goes to a basic problem with. With the. With the structure

1104
01:08:31,840 –> 01:08:34,650
of the public school system, which we never actually talked about that on this podcast.

1105
01:08:34,800 –> 01:08:38,640
It’s been a long time coming. The structure of K through 12

1106
01:08:38,720 –> 01:08:42,160
schooling is built, of course, on the

1107
01:08:42,160 –> 01:08:45,840
Henry Ford industrialization model,

1108
01:08:46,240 –> 01:08:50,080
which, by the way, takes its. Talk about history, takes

1109
01:08:50,080 –> 01:08:53,920
its cue, or took its cue from a Prussian model,

1110
01:08:53,920 –> 01:08:57,760
a German Prussian model of education, which was focused on

1111
01:08:58,160 –> 01:09:01,760
getting people just enough information to become

1112
01:09:02,000 –> 01:09:05,600
rote soldiers. Right. To be able

1113
01:09:05,600 –> 01:09:09,120
to. Because this is what. This is what not only Kaiser Wilhelm.

1114
01:09:09,680 –> 01:09:13,440
Well, every German leader from Kaiser Wilhelm all the way to Adolf

1115
01:09:13,440 –> 01:09:16,960
Hitler wanted. Right. Was

1116
01:09:17,040 –> 01:09:20,840
compliant troops that would do what they

1117
01:09:20,840 –> 01:09:24,560
were told. And Henry Ford and

1118
01:09:24,560 –> 01:09:28,280
John D. Rockefeller. And John. John Dewey. Rockefeller, sorry,

1119
01:09:28,280 –> 01:09:31,390
Dewey and Ford and all those other folks got together

1120
01:09:32,110 –> 01:09:34,870
and a bunch of other folks got together and they said, we will build an

1121
01:09:34,870 –> 01:09:37,470
American public education system to take these people from the farm

1122
01:09:38,350 –> 01:09:40,030
who were used to

1123
01:09:41,950 –> 01:09:45,750
walking behind a plow and doing whatever it is that they wanted. Right? And we

1124
01:09:45,750 –> 01:09:49,390
have to turn them into industrial cogs in a room

1125
01:09:49,630 –> 01:09:53,350
so that they don’t get up and leave to turn a person into an

1126
01:09:53,350 –> 01:09:57,140
industrial cog, starting at the age of 4 or

1127
01:09:57,140 –> 01:10:00,940
5 or 6 until they are 17

1128
01:10:00,940 –> 01:10:04,620
or 18. You don’t need to teach

1129
01:10:04,620 –> 01:10:08,380
that person history. So, of course history is taught. As to your

1130
01:10:08,380 –> 01:10:12,100
point, this dizzying array of names

1131
01:10:12,100 –> 01:10:15,020
and dates and nonsense that you just blast through

1132
01:10:15,660 –> 01:10:19,500
because you’re not actually teaching people ideas or how to think, because

1133
01:10:19,500 –> 01:10:22,620
God forbid you do that. You don’t want them to think. You want them to

1134
01:10:22,620 –> 01:10:25,880
shut up and go to the factory. And this is my

1135
01:10:25,960 –> 01:10:29,640
opposition, by the way, one of my core oppositions to the entire public

1136
01:10:29,640 –> 01:10:33,160
school system in America. We need to reform the

1137
01:10:33,160 –> 01:10:36,920
whole thing. I don’t know if that’s possible.

1138
01:10:37,080 –> 01:10:39,959
My wife is way more radical on this than I am, which is why we

1139
01:10:39,959 –> 01:10:43,600
homeschool our kids. Yes, I did just say. But we homeschool our kids. My wife

1140
01:10:43,600 –> 01:10:47,280
is way more radical on this than I am. I think the

1141
01:10:47,280 –> 01:10:50,920
system probably has to be, not probably has to be reformed

1142
01:10:52,040 –> 01:10:55,400
because we no longer live in a world where

1143
01:10:55,400 –> 01:10:58,520
rote factory work,

1144
01:10:59,000 –> 01:11:02,320
even if we bring all the factories back to America, it still won’t be rope

1145
01:11:02,320 –> 01:11:05,400
factory work. It will not be the rope factory work of

1146
01:11:06,120 –> 01:11:09,400
what Henry Ford was wanting his workers to be between

1147
01:11:09,960 –> 01:11:13,720
1910 and 1930.

1148
01:11:14,360 –> 01:11:17,850
It’s not going to be that. That worked really well in the 20th century.

1149
01:11:18,010 –> 01:11:20,650
Mass industrialization worked really well. I’m not knocking it.

1150
01:11:21,690 –> 01:11:23,930
And we don’t live in that world anymore. We don’t live in a world of

1151
01:11:23,930 –> 01:11:27,530
mass industrialization. We need people to think in terms of ideas,

1152
01:11:28,650 –> 01:11:32,250
not in terms of rote responses to the test.

1153
01:11:35,210 –> 01:11:38,890
And so history, which, to your point about

1154
01:11:39,050 –> 01:11:41,450
being an expert in what happened in the world between,

1155
01:11:42,780 –> 01:11:46,340
let’s say, let’s conserve, let’s take, let’s pick a 25 year period between like, you

1156
01:11:46,340 –> 01:11:50,140
know, 1915 and, and you know, our 15 year period,

1157
01:11:50,140 –> 01:11:52,860
1950. Oh, no, no, let’s put 25, 1915 to

1158
01:11:52,860 –> 01:11:56,060
1939. Being an expert just on that

1159
01:11:56,380 –> 01:11:59,980
requires you to be as, as someone once said,

1160
01:11:59,980 –> 01:12:02,980
Richard Dawkins, I think when he was talking to Jordan Peterson, used this phrase, which

1161
01:12:02,980 –> 01:12:06,340
I really like, be drunk on ideas. You have to be drunk on the

1162
01:12:06,340 –> 01:12:10,110
ideas of that period. Because to your point,

1163
01:12:10,110 –> 01:12:13,870
there’s a lot of ideas of that period, economic, social, cultural, and on

1164
01:12:13,870 –> 01:12:17,630
and on and on. I don’t know that you have anybody in

1165
01:12:17,630 –> 01:12:21,070
K through 12 who

1166
01:12:21,070 –> 01:12:24,470
approaches because they’re still doing the rope thing.

1167
01:12:24,710 –> 01:12:28,510
Right, the rope preparation thing. And are there some systems that are changed, some

1168
01:12:28,510 –> 01:12:31,590
parts of the school system, some parts of the country that are changing? Yes. I’m

1169
01:12:31,590 –> 01:12:33,470
sure that if you’re in the sound of my voice, you’re going to think of

1170
01:12:33,470 –> 01:12:37,280
Your K through 12 public education building or your

1171
01:12:37,280 –> 01:12:40,600
student and you’re going to say, well, not my kid. That’s not what’s happening. Blah,

1172
01:12:40,600 –> 01:12:43,880
blah, blah, blah, blah. It’s great. Okay, sure. No one raindrop blames itself for the

1173
01:12:43,880 –> 01:12:47,000
flood. Okay, cool. Yeah, you’re special. Got it. All right.

1174
01:12:48,120 –> 01:12:51,800
But you see the outcomes, by the way, when those

1175
01:12:51,800 –> 01:12:54,800
kids who have gone through the K through 12 system and some of them have

1176
01:12:54,800 –> 01:12:58,520
matriculated to college now come out and

1177
01:12:58,600 –> 01:13:01,720
not only do they not know history, but they know tick tock really well.

1178
01:13:02,930 –> 01:13:06,770
They don’t know how to write a sentence. And now I, as an employer,

1179
01:13:06,770 –> 01:13:10,450
as a leader, have to lead these people around who, who have

1180
01:13:10,450 –> 01:13:14,170
never been given the who’s. Who’s whose thirst has

1181
01:13:14,170 –> 01:13:17,570
never been activated for even the history

1182
01:13:17,809 –> 01:13:21,450
of the business that they are in because their thirst was never

1183
01:13:21,450 –> 01:13:25,250
activated properly for the history of the world and

1184
01:13:25,250 –> 01:13:28,900
country that they live in. That’s a

1185
01:13:28,900 –> 01:13:32,540
tragedy. I said all of that. To say that that is an

1186
01:13:32,860 –> 01:13:35,500
absolute tragedy on how we teach history.

1187
01:13:36,620 –> 01:13:40,420
And as an amateur historian and a person who gets interested

1188
01:13:40,420 –> 01:13:44,140
in all of this stuff, it’s an absolute bugaboo for me. Like,

1189
01:13:45,020 –> 01:13:48,380
we do. We homeschool our kids and so, like, when my wife has a difficult

1190
01:13:48,380 –> 01:13:52,180
question in history that she does not know how to answer or how to propose

1191
01:13:52,180 –> 01:13:55,370
or difficult idea that pops up in history, go ask.

1192
01:13:56,090 –> 01:13:59,730
She told me to go ask. Go ask your father. Go ask him. He’ll tell

1193
01:13:59,730 –> 01:14:02,690
you. And I’ll talk to you for four hours about it. Because I just, I

1194
01:14:02,690 –> 01:14:05,930
just know. I just know the stuff. Like, sure, ask me about the collapse of

1195
01:14:05,930 –> 01:14:09,370
the rise and fall the Ottoman Empire. Like, why do I know about that?

1196
01:14:10,890 –> 01:14:13,530
Why do I need to know about that? Like the sultan here,

1197
01:14:14,490 –> 01:14:16,970
like, why do I need to know about that? Or why do I need to

1198
01:14:16,970 –> 01:14:20,290
know about the monetary system of, you know, the Greek

1199
01:14:20,290 –> 01:14:23,990
islands as described by Heraclitus?

1200
01:14:24,470 –> 01:14:27,710
Why do I need. Why, why do I know this? Because I’m passionate about the

1201
01:14:27,710 –> 01:14:31,270
ideas behind those kinds of things. And there’s always a core idea.

1202
01:14:31,350 –> 01:14:35,110
And of course the idea is of humanity advancing and building civilization

1203
01:14:35,670 –> 01:14:38,550
built on to Lidell Hart’s idea, built on promises.

1204
01:14:39,430 –> 01:14:42,550
But at the end of the day, I’m drunk on ideas. And you need a

1205
01:14:42,550 –> 01:14:46,390
history teacher to be drunk on ideas, particularly at the K12 level, because

1206
01:14:46,390 –> 01:14:50,070
my, the kids will pick up on that passion and then they’ll lock it.

1207
01:14:53,110 –> 01:14:55,270
And by the way, you never want to ask a kid to predict the future.

1208
01:14:56,790 –> 01:15:00,590
Like, that’s ridiculous. That’s. That’s some nonsense. Did you,

1209
01:15:00,590 –> 01:15:03,069
when you were in high school, did you go to, like, not go to. But

1210
01:15:03,069 –> 01:15:05,310
were you like, in that? Did you. You don’t have to say the year you

1211
01:15:05,310 –> 01:15:09,030
graduated, but like, in the four years of high school, looking back,

1212
01:15:09,110 –> 01:15:12,910
was that part of like the years in this country where we were doing experimental

1213
01:15:12,910 –> 01:15:16,270
learning? Oh, God, yes. Yeah, yeah, yeah. Okay. All right, so you, you.

1214
01:15:16,510 –> 01:15:19,390
I’m sorry, I’m sorry.

1215
01:15:20,030 –> 01:15:23,710
Yeah, it was just. And again, it was like. But

1216
01:15:24,430 –> 01:15:27,910
there was. They were trying to teach us like,

1217
01:15:27,910 –> 01:15:31,750
predictive analytics. I. They shouldn’t have called it a history class, put it that

1218
01:15:31,750 –> 01:15:35,150
way. But, but it fell under, under the history department

1219
01:15:35,790 –> 01:15:39,360
because it was supposed to be historical data that was supposed to be predictive analytics.

1220
01:15:41,030 –> 01:15:43,990
I always thought that that class should have been under the math department, but that’s

1221
01:15:45,110 –> 01:15:48,950
whatever. So, but anyway, either way, nothing,

1222
01:15:48,950 –> 01:15:51,990
nothing that we had. Not a single one of us got a damn thing right.

1223
01:15:52,870 –> 01:15:56,630
But, but okay, so. Well, one last question.

1224
01:15:56,710 –> 01:15:59,590
So you, you brought up something that’s very interesting. So

1225
01:16:00,390 –> 01:16:04,070
I’ll ask, I’ll frame my observation the four of a question without any

1226
01:16:04,070 –> 01:16:07,760
run up. How’s your factual history? Why

1227
01:16:07,760 –> 01:16:08,880
is that so popular?

1228
01:16:12,080 –> 01:16:13,680
Well, I, I, I think,

1229
01:16:18,320 –> 01:16:22,160
I, I think it gives you retrospect and, and I think that

1230
01:16:22,160 –> 01:16:25,760
the, we talked a lot about history invoking emotions here.

1231
01:16:26,240 –> 01:16:29,960
I think sometimes I, I think sometimes

1232
01:16:29,960 –> 01:16:33,670
that counterfactual information you get a different sense,

1233
01:16:33,670 –> 01:16:37,430
you get a different emotion coming from it and then it’s up to you to

1234
01:16:37,430 –> 01:16:40,270
kind of balance that that, that too. Again,

1235
01:16:41,150 –> 01:16:43,950
this is, this is a very, this to me is a very clear

1236
01:16:45,070 –> 01:16:48,510
example of. There’s always three sides to every story.

1237
01:16:48,670 –> 01:16:51,630
His, hers and the truth. That’s why to me,

1238
01:16:51,870 –> 01:16:55,510
counterfactual information that can be very, very important when you

1239
01:16:55,510 –> 01:16:59,340
start seeing. Again, I, I’m not talking about

1240
01:16:59,340 –> 01:17:02,980
statistical, verifiable statistical data. I’m talking about

1241
01:17:03,700 –> 01:17:07,540
like, I’ll just take our, an example

1242
01:17:07,540 –> 01:17:10,980
from, from my, my own, my own

1243
01:17:11,140 –> 01:17:14,580
repertoire. Here, right here in Massachusetts, out in western part of

1244
01:17:14,580 –> 01:17:18,100
Massachusetts, there’s a small town that originally was a native community.

1245
01:17:18,820 –> 01:17:21,620
And the, the,

1246
01:17:23,710 –> 01:17:27,310
the facts of the case, the statistical data behind it

1247
01:17:27,310 –> 01:17:31,150
was a hundred armed men walked

1248
01:17:31,150 –> 01:17:34,590
into that village and killed 400 unarmed people.

1249
01:17:34,990 –> 01:17:38,750
Period. Okay. Okay. The factual versus

1250
01:17:38,750 –> 01:17:41,950
counterfactual details to it were

1251
01:17:42,830 –> 01:17:46,470
that the armed men were colonialists. The

1252
01:17:46,470 –> 01:17:50,160
native community was comprised at the time of the

1253
01:17:50,160 –> 01:17:53,840
massacre or at the time of the, the incursion. The, the time of the

1254
01:17:53,840 –> 01:17:56,120
event was mostly women and children.

1255
01:17:57,640 –> 01:18:01,480
So you had a hundred armed men going in and killing 400

1256
01:18:01,720 –> 01:18:05,120
unarmed women and children. Okay, the

1257
01:18:05,120 –> 01:18:08,960
counter. So that’s, that’s all factual information that’s supposed to invoke

1258
01:18:08,960 –> 01:18:12,680
a certain kind of feeling to you. And I’m, as I’m even saying it, I

1259
01:18:12,680 –> 01:18:16,090
get riled up a little bit inside. But so then the alternate, the

1260
01:18:16,090 –> 01:18:19,850
counterfactual information of that was the armed men had

1261
01:18:19,850 –> 01:18:23,690
intel that told them that warriors were going to be there and

1262
01:18:23,690 –> 01:18:27,210
that they should shoot on site. And it didn’t matter whether they were

1263
01:18:27,370 –> 01:18:30,810
just shoot the people and we’ll, it’s like basically drop the bodies and we’ll figure

1264
01:18:30,810 –> 01:18:34,410
it out later. Most of those armed men,

1265
01:18:35,210 –> 01:18:39,050
years later had tremendous

1266
01:18:39,050 –> 01:18:42,810
mental breakdowns. Over it. These guys did not live

1267
01:18:42,810 –> 01:18:46,530
normal, healthy lives after the fact. So they did. They were

1268
01:18:46,530 –> 01:18:49,850
basically just taking. Let me, Let me ask you if you heard this story before.

1269
01:18:50,730 –> 01:18:54,170
We killed innocent people, but we were just taking orders. We’re a military

1270
01:18:54,890 –> 01:18:58,450
faction that we’re just, we’re just doing our job. We’re just taking orders. So you

1271
01:18:58,450 –> 01:19:01,370
shouldn’t vilify us over this. Right.

1272
01:19:02,330 –> 01:19:06,170
So again, when you talk about counterfactual information or,

1273
01:19:06,170 –> 01:19:09,670
or. And again, it’s not because the.

1274
01:19:09,750 –> 01:19:13,550
Again, facts are facts, you can’t really change them, but sometimes when

1275
01:19:13,550 –> 01:19:17,110
you hear only one side of the facts and you don’t hear both

1276
01:19:17,190 –> 01:19:20,950
sides of, again, factual information, albeit. And again,

1277
01:19:21,750 –> 01:19:24,990
the point is now you have his, hers in the truth. Right? So now you

1278
01:19:24,990 –> 01:19:28,590
have the, the. The. We also don’t

1279
01:19:28,590 –> 01:19:32,110
know, the warriors that were expected to be there, were they

1280
01:19:32,110 –> 01:19:35,720
deemed, you know, criminals

1281
01:19:35,720 –> 01:19:39,480
or some sort of. What is the

1282
01:19:39,480 –> 01:19:43,280
word that I’m thinking of here? Somebody that you’re

1283
01:19:43,280 –> 01:19:46,960
targeting very specifically in warfare. I can’t remember there’s a particular name for it.

1284
01:19:47,280 –> 01:19:51,080
Oh, insurgents. Yeah. Yeah. So. So though. And

1285
01:19:51,080 –> 01:19:54,000
they were known. They. We don’t know. There were certain facts that we don’t know.

1286
01:19:54,000 –> 01:19:57,440
Now after the fact, after the deed was done, after, you know, and this came

1287
01:19:57,440 –> 01:20:00,970
out probably 100 years later,

1288
01:20:01,450 –> 01:20:05,050
that those particular warriors were in fact involved

1289
01:20:05,050 –> 01:20:08,890
in raids on, you know, on the colonial. And this, this was

1290
01:20:09,290 –> 01:20:12,970
pre. Before the time of the state of Massachusetts even existed.

1291
01:20:13,370 –> 01:20:17,090
So this is like in the 1600s. And. But so that

1292
01:20:17,090 –> 01:20:20,610
you find out after the fact that those warriors were indeed, in fact part

1293
01:20:20,610 –> 01:20:24,330
of. Of raids that were on, you know, colonial

1294
01:20:24,330 –> 01:20:28,050
towns and they, they did do some stuff. They. That could have warranted them

1295
01:20:28,530 –> 01:20:32,330
being singled out and, and going after. So

1296
01:20:32,330 –> 01:20:36,010
does that justify the hundred armed men going in there and killing 400

1297
01:20:36,010 –> 01:20:39,570
unarmed and innocent women and children? Probably not that, but

1298
01:20:39,730 –> 01:20:41,810
let me say. Let me rephrase that. No.

1299
01:20:44,610 –> 01:20:48,050
Does that negate them from. From responsibility of quote,

1300
01:20:48,050 –> 01:20:51,770
unquote, just taking orders? No. But if you

1301
01:20:51,770 –> 01:20:55,170
were in that moment and you were one of those soldiers, would you have done

1302
01:20:55,170 –> 01:20:57,800
something different? My guess is no.

1303
01:20:58,760 –> 01:21:02,280
Oh, does something different. No. Everybody’s a hero after the fact

1304
01:21:03,800 –> 01:21:07,600
or. Or a villain after the fact. Everybody’s a hero or a villain. Right. But

1305
01:21:07,600 –> 01:21:11,320
in that moment, can you truly crucify that per those. Those

1306
01:21:11,320 –> 01:21:15,120
hundred guys? Some could say yes to a

1307
01:21:15,120 –> 01:21:18,760
degree. Because again, again, walking into a village of

1308
01:21:19,000 –> 01:21:22,560
400 innocent women and children, by the way, I keep saying the word innocent because

1309
01:21:22,560 –> 01:21:26,290
that’s the way it’s Depicted, you can simply say 400 women

1310
01:21:26,290 –> 01:21:29,810
and children. Right. Yeah. You don’t. Yeah, you don’t need another adjective on there.

1311
01:21:30,770 –> 01:21:34,330
But. But again, but the word innocent is. Is purposeful.

1312
01:21:34,330 –> 01:21:38,050
It’s intended to invoke emotion to you. Right, Right.

1313
01:21:38,050 –> 01:21:41,650
So that. That’s. Again, that’s kind of what I’m. What I’m getting at here. And

1314
01:21:41,650 –> 01:21:44,850
when you’re describing the 100 men, it’s 100 armed

1315
01:21:45,090 –> 01:21:48,450
military men. Like, again, you’re supposed to get a feeling

1316
01:21:48,710 –> 01:21:52,070
of what you’re. What you’re seeing here. And you’re supposed to develop this

1317
01:21:52,230 –> 01:21:55,430
vision in your brain as to what was happening at the moment. Now, there are

1318
01:21:55,430 –> 01:21:59,230
people that will argue that at some point One of those

1319
01:21:59,230 –> 01:22:03,070
100 should have realized that they weren’t armed insurgents, that they were actually

1320
01:22:03,070 –> 01:22:06,710
women and children. Stopped the massacre. Maybe you killed

1321
01:22:06,710 –> 01:22:10,150
100 of them, but you’re not going to kill all 400.

1322
01:22:10,790 –> 01:22:14,510
Right. But that’s. That’s the argument that’s made in most of these cases

1323
01:22:14,510 –> 01:22:18,360
where you’re no longer just following orders. You are now just a

1324
01:22:18,360 –> 01:22:21,760
murderer because you’ve. You’ve walked into this village, you started shooting.

1325
01:22:21,920 –> 01:22:25,280
Nobody stopped to actually think or do or whatever. But again,

1326
01:22:25,840 –> 01:22:29,680
that’s the whole. Like, to your point about. Or the question about car. Counterfactual

1327
01:22:29,680 –> 01:22:33,520
information is. To me, it’s always that.

1328
01:22:33,840 –> 01:22:37,600
That version of his, hers and the truth. Always. There’s always some semblance

1329
01:22:37,600 –> 01:22:41,360
of. You’re gonna get a story from the victor, a story from the loser.

1330
01:22:41,360 –> 01:22:45,130
And the. The actual story is probably somewhere in the middle. Somewhere in the

1331
01:22:45,130 –> 01:22:48,330
middle. Yeah. Yeah, yeah, yeah. Well. And

1332
01:22:52,650 –> 01:22:56,250
Liddell Hart has an answer for this. Back to the book. There’s a thought about

1333
01:22:56,250 –> 01:23:00,010
this. Back to the book. Back to. Why don’t we learn from history?

1334
01:23:00,970 –> 01:23:04,330
So this is in the War and Peace section. The

1335
01:23:04,330 –> 01:23:06,650
dilemma of the Intellectual.

1336
01:23:10,900 –> 01:23:13,060
We did not plan this, folks, so this is good.

1337
01:23:15,940 –> 01:23:19,420
Neither intellectuals nor their critics appear to recognize the

1338
01:23:19,420 –> 01:23:23,060
inherent dilemma of the thinking man and its inevitability.

1339
01:23:24,020 –> 01:23:27,100
The dilemma should be faced, for it is a natural part of the growth of

1340
01:23:27,100 –> 01:23:30,860
any human mind. An intellectual ought to realize the extent to

1341
01:23:30,860 –> 01:23:34,580
which the world is shaped by human emotions, emotions uncontrolled

1342
01:23:34,580 –> 01:23:37,900
by reason. His thinking must have been shallow and his

1343
01:23:37,900 –> 01:23:41,620
observations narrow. If he fails to realize that having once

1344
01:23:41,620 –> 01:23:45,460
learned to think and to use reason as a guide, however, he cannot possibly

1345
01:23:45,460 –> 01:23:49,139
float with the current of popular emotion and fluctuate with its violent

1346
01:23:49,139 –> 01:23:52,980
changes unless he himself ceases to think or is deliberately false in

1347
01:23:52,980 –> 01:23:56,780
his own thought. And in the Latter case, it

1348
01:23:56,780 –> 01:24:00,520
is likely that he will commit intellectual suicide gradually, quote,

1349
01:24:00,520 –> 01:24:04,040
unquote, by the death of a thousand cuts. A deeper

1350
01:24:04,040 –> 01:24:07,840
diagnosis of the malady from which left wing intellectuals have suffered in the

1351
01:24:07,840 –> 01:24:11,520
past might suggest that their troubles have come not from following reason too far,

1352
01:24:11,920 –> 01:24:14,960
but from not following it far enough to realize the general power of

1353
01:24:14,960 –> 01:24:18,760
unreason. That’s an interesting point. Many

1354
01:24:18,760 –> 01:24:22,600
of them also seem to have suffered from failing to apply reason internally as

1355
01:24:22,600 –> 01:24:26,290
well as externally, through not using it with the control of their own

1356
01:24:26,290 –> 01:24:30,090
emotions. In that way, they unwittingly helped to get this country into the mess

1357
01:24:30,090 –> 01:24:33,770
of the last war and then found themselves in an intellectual mess as a

1358
01:24:33,770 –> 01:24:37,450
result. In one of the more penetrating criticisms

1359
01:24:37,450 –> 01:24:39,930
written on this subject, George Orwell

1360
01:24:41,690 –> 01:24:45,330
exposed a profound truth in saying that, quote, the energy that

1361
01:24:45,330 –> 01:24:49,170
actually shapes the world springs from emotions. He referred to the deep

1362
01:24:49,170 –> 01:24:53,010
seated and dynamic power of racial pride, leader worship, religious belief, love

1363
01:24:53,010 –> 01:24:56,580
of war. There are powerful emotions beyond these, however,

1364
01:24:56,740 –> 01:25:00,540
the energy of the intellectual himself springs from emotion, the love of

1365
01:25:00,540 –> 01:25:04,260
truth, the desire for wider knowledge and understanding. That emotion

1366
01:25:04,260 –> 01:25:06,980
has done quite a lot to shape the world, as a study of world history

1367
01:25:07,220 –> 01:25:10,900
amply shows. In the thinking man, that

1368
01:25:10,900 –> 01:25:14,540
source of energy dries up only when he ceases to believe

1369
01:25:14,540 –> 01:25:18,100
in the guiding power of thought and allows himself to become merely a

1370
01:25:18,100 –> 01:25:21,390
vehicle for the prevailing popular emotions of

1371
01:25:21,790 –> 01:25:25,590
the moment. I’m going to skip

1372
01:25:25,590 –> 01:25:28,990
down because he’s going to talk about Bertrand Russell, he says, and I’m going to

1373
01:25:28,990 –> 01:25:32,270
read this paragraph. History bears witness to the vital part that

1374
01:25:32,270 –> 01:25:36,110
prophets, quote, unquote, have played in human progress, which is evidence of

1375
01:25:36,110 –> 01:25:39,950
the ultimate practical value of expressing unreservedly the truth as

1376
01:25:39,950 –> 01:25:43,790
one sees it. Yet it also becomes clear that the acceptance and spreading

1377
01:25:43,790 –> 01:25:46,840
of their vision has always depended on another class of men, leaders

1378
01:25:47,560 –> 01:25:51,040
who had to be philosophical strategists striking a

1379
01:25:51,040 –> 01:25:54,280
compromise between truth. I think you’ve said this actually,

1380
01:25:54,760 –> 01:25:58,440
Tom, striking a compromise between truth and

1381
01:25:58,600 –> 01:26:02,199
men’s receptivity to it. Their effect

1382
01:26:02,199 –> 01:26:05,840
is often depended as much on their own limitations in perceiving the

1383
01:26:05,840 –> 01:26:09,160
truth as on their practical wisdom in proclaiming it.

1384
01:26:09,960 –> 01:26:13,800
The prophet must be stoned. That is their lot and the test of their self

1385
01:26:13,800 –> 01:26:17,440
fulfillment. A leader who is stoned, however, may merely prove that he has failed in

1386
01:26:17,440 –> 01:26:21,080
his function through a deficiency of wisdom or through confusion, confusing his

1387
01:26:21,080 –> 01:26:24,880
function with that of a prophet. Time alone can tell whether the effect of

1388
01:26:24,880 –> 01:26:28,680
such a sacrifice redeems the apparent failure as a leader that does honor to

1389
01:26:28,680 –> 01:26:32,480
him as a man. And then, finally, I’m going to Skip

1390
01:26:32,480 –> 01:26:36,080
a couple paragraphs to go down. Opposition to the truth is inevitable,

1391
01:26:36,160 –> 01:26:39,400
especially if it takes the form of a new idea. But the degree of resistance

1392
01:26:39,400 –> 01:26:42,440
can be diminished by giving thought not only to the aim but to the method

1393
01:26:42,440 –> 01:26:45,470
of approach. This is an important tactic.

1394
01:26:46,190 –> 01:26:48,830
Avoid a frontal attack on a long established position.

1395
01:26:49,630 –> 01:26:53,470
Instead, seek to turn it by a flank movement so that a more

1396
01:26:53,470 –> 01:26:56,430
penetrable side is exposed to the thrust of truth.

1397
01:26:57,390 –> 01:27:01,070
But any such indirect approach. But in any such indirect

1398
01:27:01,070 –> 01:27:04,830
approach, take care not to diverge from the truth. For nothing is

1399
01:27:04,830 –> 01:27:08,430
more fatal to its real advancement than to lapse

1400
01:27:09,060 –> 01:27:10,980
into untruth.

1401
01:27:16,420 –> 01:27:19,860
I wish this guy was still alive. I’d love to have him on the show.

1402
01:27:26,180 –> 01:27:29,700
The dilemma of the intellectual. Or,

1403
01:27:29,860 –> 01:27:33,140
you know, I mean, I. There’s a clip floating around from one of our episodes

1404
01:27:33,140 –> 01:27:36,950
of our show. We were talking about Sam Altman. If we’re so smart,

1405
01:27:36,950 –> 01:27:40,590
why aren’t we rich? Tom, you know. Yeah, I watch that a lot,

1406
01:27:41,630 –> 01:27:42,430
that clip.

1407
01:27:50,350 –> 01:27:54,190
I mean, the big

1408
01:27:54,190 –> 01:27:57,790
separation in life is between the thinker and the doer, right? And,

1409
01:27:59,230 –> 01:28:02,910
and the intellect is hobbled by his mind and the working of his faculties.

1410
01:28:03,380 –> 01:28:06,580
He’s hobbled into not doing or not acting.

1411
01:28:08,980 –> 01:28:12,420
However, the theory of how people operate can only take you so far.

1412
01:28:12,820 –> 01:28:16,620
You actually have to go into practice. You actually have to. You actually

1413
01:28:16,620 –> 01:28:20,460
have to actually interact with people, right? And the dilemma of the intellectual,

1414
01:28:20,460 –> 01:28:24,020
the, the practical dilemma that Liddell Hart is talking about

1415
01:28:24,020 –> 01:28:27,780
here relates exactly to what we’ve been talking about here on,

1416
01:28:27,860 –> 01:28:30,520
on the show. When we over

1417
01:28:30,520 –> 01:28:34,240
intellectualize history and we lack

1418
01:28:34,240 –> 01:28:37,600
the emotive where we. Or we deny the emotive power of it, or even worse.

1419
01:28:37,600 –> 01:28:40,320
And I see a lot of this happening particularly in the last, I would say

1420
01:28:40,320 –> 01:28:43,720
10 to 15 years in our country when we are. When we allow

1421
01:28:43,800 –> 01:28:47,440
history to fall into political or ideological

1422
01:28:47,440 –> 01:28:51,200
capture. Now we’re tying emotions

1423
01:28:51,200 –> 01:28:54,920
to ideology to wind people up because we don’t know

1424
01:28:55,770 –> 01:28:59,530
another way to get them to act like.

1425
01:28:59,530 –> 01:29:03,170
One of the knocks in our time is how few people, particularly

1426
01:29:03,170 –> 01:29:06,010
young people, protest

1427
01:29:06,410 –> 01:29:09,850
injustices like previous generations

1428
01:29:10,170 –> 01:29:13,930
did in the 60s and 70s. And there’s a lot of

1429
01:29:13,930 –> 01:29:17,770
different factors that go into why young people don’t protest injustices at

1430
01:29:17,770 –> 01:29:21,500
as high a rate as was

1431
01:29:21,500 –> 01:29:25,300
done by previous generations. There’s a lot of different factors that go into this.

1432
01:29:26,260 –> 01:29:30,020
But the biggest thing that I see is the protests from young people

1433
01:29:30,420 –> 01:29:34,100
that are about injustices that may or may not be

1434
01:29:34,100 –> 01:29:37,340
happening are probably some of the most ill

1435
01:29:37,340 –> 01:29:39,060
informed protests

1436
01:29:41,460 –> 01:29:44,260
in recent history because

1437
01:29:46,590 –> 01:29:50,350
they have not been given the intellectual foundation

1438
01:29:51,790 –> 01:29:54,990
They’ve just the intellectual foundation to think through their positions.

1439
01:29:55,870 –> 01:29:58,310
So I’ll give you a case in point. I can pro, this is a big

1440
01:29:58,310 –> 01:30:01,990
example and I can point to this when I see a protest on

1441
01:30:01,990 –> 01:30:05,830
an Ivy League college campus. Columbia is

1442
01:30:05,830 –> 01:30:07,630
too easy to pick on, so I’ll pick on Stanford.

1443
01:30:11,790 –> 01:30:15,620
And, and, and just for the record. Stanford is not an Ivy League school,

1444
01:30:15,620 –> 01:30:19,380
but go ahead. Well, they, they think they are, but anyway, it’s fine. I

1445
01:30:19,380 –> 01:30:23,220
know, I know they’re not, but they think they are anyway, so. Yeah,

1446
01:30:25,140 –> 01:30:28,460
and, and, and students unroll a banner in

1447
01:30:28,460 –> 01:30:32,260
protest and the banner on

1448
01:30:33,060 –> 01:30:35,700
says Gays for G.

1449
01:30:39,950 –> 01:30:43,710
There’s a fundamental, there’s several fundamental. And I’m

1450
01:30:43,710 –> 01:30:47,310
not the only person to point this out. This is why it’s so easy. There’s

1451
01:30:47,310 –> 01:30:51,030
several fundamental steps in

1452
01:30:51,030 –> 01:30:54,749
the ladder to that thought that you just

1453
01:30:54,749 –> 01:30:58,510
haven’t been educated on gay people

1454
01:30:58,510 –> 01:31:01,550
in Gaza? Well, there are none

1455
01:31:03,390 –> 01:31:07,060
because in that part of the world, Islam has a

1456
01:31:07,380 –> 01:31:10,660
tradition as a religion

1457
01:31:12,420 –> 01:31:16,020
that is not as, shall we say, tolerant as

1458
01:31:16,020 –> 01:31:19,220
Western Christianity has been. And Western

1459
01:31:19,220 –> 01:31:21,460
secularism has been around

1460
01:31:23,300 –> 01:31:26,820
the public open practice out of the closet of

1461
01:31:27,140 –> 01:31:30,700
homosexuality. And so if you

1462
01:31:30,700 –> 01:31:34,510
are unrolling a banner that proclaims that you are

1463
01:31:34,510 –> 01:31:38,230
gay and you, you are protesting against a

1464
01:31:38,230 –> 01:31:41,990
perceived injustice happening during a war in Gaza,

1465
01:31:42,470 –> 01:31:45,830
so thus you are pro the people in Gaza, you have to understand

1466
01:31:46,310 –> 01:31:50,150
that the history of the people that you claim to be in, in solidarity

1467
01:31:50,150 –> 01:31:53,990
with, those people don’t want to be in solidarity

1468
01:31:53,990 –> 01:31:57,070
with you. And by the way, this is not something that they’ve recently like come

1469
01:31:57,070 –> 01:32:00,590
to a conclusion about. They don’t want to be in

1470
01:32:00,590 –> 01:32:02,270
solidarity with you going back to like 600,

1471
01:32:04,510 –> 01:32:08,230
the 6th century. And by the way, they’ve been very clear about

1472
01:32:08,230 –> 01:32:10,110
this. They’re not hiding it.

1473
01:32:12,030 –> 01:32:15,470
So that reveals an intellectual poverty

1474
01:32:16,110 –> 01:32:19,870
that is incredibly troubling. That could be an intellectual poverty that

1475
01:32:19,870 –> 01:32:23,610
again could be solved through an actual understanding

1476
01:32:23,680 –> 01:32:27,360
understanding of the facts of history. So would you, would you

1477
01:32:27,360 –> 01:32:30,960
been, would you have been more understanding of the

1478
01:32:30,960 –> 01:32:34,520
rolling out of the banner if it said gays for Gaza and then in

1479
01:32:34,520 –> 01:32:36,960
parentheses it said we understand the irony?

1480
01:32:38,880 –> 01:32:41,960
Yes, I would have been. Yes, that, that would actually, that would actually have worked

1481
01:32:41,960 –> 01:32:45,240
for me. Yes, actually, that, that would, that would have worked for me. Absolutely. And

1482
01:32:45,240 –> 01:32:47,440
I picked that one because it’s, it’s a hot bit one, but you can see

1483
01:32:47,440 –> 01:32:50,130
it everywhere on social media. I picked that. So I picked that one. One. It’s

1484
01:32:50,130 –> 01:32:53,930
also very blatant. Like, I knew exactly where you were going with

1485
01:32:53,930 –> 01:32:57,610
this as soon as you said, the. What the banner said. Yeah, it’s very blatant.

1486
01:32:57,610 –> 01:33:00,450
I do, I do agree with you in, in a sense, like, so

1487
01:33:01,730 –> 01:33:05,570
there’s. Just a lack of, a lack of historical understanding behind that. Right, right.

1488
01:33:05,970 –> 01:33:09,730
And, and to my point, it’s not whether you are gay or

1489
01:33:09,730 –> 01:33:13,330
not. Shouldn’t have, like, who cares? I, I don’t. It’s not

1490
01:33:13,890 –> 01:33:16,530
if they’re gay or not. But if you’re going to write something like that,

1491
01:33:17,680 –> 01:33:21,480
to your point, or to my point a second ago, be clever

1492
01:33:21,480 –> 01:33:25,240
about it and make sure you let the world know you understand the irony

1493
01:33:25,240 –> 01:33:29,040
behind it. Like, we, we know that the people in Gaza don’t give a

1494
01:33:29,040 –> 01:33:32,880
crap about us, but we, we care about them. We care about the

1495
01:33:32,880 –> 01:33:36,680
right, we care about the X, Y, whatever, whatever. If you feel

1496
01:33:36,680 –> 01:33:40,400
it’s an atrocity or if you feel like it’s a humanitarian thing,

1497
01:33:40,640 –> 01:33:44,470
whatever, whatever you’re. But, but to your point, and I, I don’t disagree

1498
01:33:44,470 –> 01:33:47,990
with you, by the way, that there is a fundamental disconnect

1499
01:33:48,230 –> 01:33:51,430
with some of the, the ways in which people are

1500
01:33:52,070 –> 01:33:55,670
project that they think versus what they actually think or what

1501
01:33:55,670 –> 01:33:59,430
or, or why they think that way. There’s, there’s disconnect there. And I, I think,

1502
01:33:59,510 –> 01:34:02,390
yeah, when I saw, I, I saw something similar

1503
01:34:03,270 –> 01:34:06,590
several times. And it’s not just Stanford that has done it. There’s been. The

1504
01:34:06,590 –> 01:34:10,160
LGBTQ community has done several things like that.

1505
01:34:10,640 –> 01:34:14,400
And, and Right. I just go, but do you,

1506
01:34:15,040 –> 01:34:18,720
like, do you understand who you’re supporting? And by the way,

1507
01:34:18,800 –> 01:34:22,000
I’m not suggesting that we shouldn’t stop the, the,

1508
01:34:22,640 –> 01:34:26,280
this craziness that’s happening in the Middle East. I. No, no, no. I am for

1509
01:34:26,280 –> 01:34:30,080
stopping killing people. Believe me. I don’t care what country you’re in.

1510
01:34:30,080 –> 01:34:33,400
I had. Stop killing people. Like, why are we doing this? This doesn’t make any

1511
01:34:33,400 –> 01:34:36,940
sense. You, you referenced a few minutes ago or earlier in this

1512
01:34:36,940 –> 01:34:40,740
podcast, this, this podcast episode that we today have more

1513
01:34:40,820 –> 01:34:44,580
avenues for diplomacy than we ever have yet. Ever.

1514
01:34:44,900 –> 01:34:48,100
Yet we still have. These things happen. Like,

1515
01:34:48,660 –> 01:34:52,380
absolutely. Because. Because human nature will. Human nature will. Will out.

1516
01:34:52,380 –> 01:34:55,980
Even Ladell Hart would say this human nature will out. Like, it just, it just

1517
01:34:55,980 –> 01:34:59,780
will. People will go to war. I’m reading John Keegan’s, which we’re

1518
01:34:59,780 –> 01:35:03,220
going to cover that on the podcast too. Coming up here in an episode, John

1519
01:35:03,220 –> 01:35:06,230
Keegan’s History of the First World War War. And it is

1520
01:35:07,510 –> 01:35:11,270
the thing, the reasons he talks about for going to war are

1521
01:35:11,270 –> 01:35:15,030
the same reasons that people go to war now. It has not changed.

1522
01:35:15,110 –> 01:35:18,790
National pride. A leader’s personal hurt feelings

1523
01:35:18,950 –> 01:35:22,750
and arrogance and ego. Right. They think they can get away with it or

1524
01:35:22,750 –> 01:35:26,590
they have hurt pride as a nation or as a leader. Arrogance,

1525
01:35:26,590 –> 01:35:30,310
ego, pride, power. Right, exactly.

1526
01:35:30,310 –> 01:35:33,950
And, and so they’re just going to. There’s going to do it. So I, I

1527
01:35:33,950 –> 01:35:37,470
would. And yes, if there were a, there were a banner that were rolled out

1528
01:35:37,470 –> 01:35:41,310
underneath there, absolutely. I would, I would be

1529
01:35:41,310 –> 01:35:44,470
like, oh, okay, so you, you know what the game is. Okay, cool. I can

1530
01:35:44,470 –> 01:35:47,670
leave that. But, but because there’s not a banner there

1531
01:35:48,150 –> 01:35:51,910
now, I have to treat you or I have to treat that message. I have

1532
01:35:51,910 –> 01:35:55,190
to deal with that message. I have to examine that message critically

1533
01:35:55,750 –> 01:35:59,600
in the realm of ideas and in the realm of ideas,

1534
01:36:00,080 –> 01:36:01,120
that message.

1535
01:36:04,240 –> 01:36:07,120
It may take a brain surgeon. Normally you say it doesn’t take a brain surgeon.

1536
01:36:07,440 –> 01:36:10,960
No, no. It may take a brain surgeon to, to

1537
01:36:11,280 –> 01:36:15,120
parse critically what’s happening inside of that

1538
01:36:15,120 –> 01:36:18,800
message and to realize, again, the poverty of idea that’s at the

1539
01:36:18,800 –> 01:36:21,200
bottom of it. And by the way, there’s a lot of these messages that are

1540
01:36:21,200 –> 01:36:24,680
floating around now, like, okay, you want to be in solidarity with the Palestinians for

1541
01:36:24,680 –> 01:36:28,480
whatever identity group you come from in America. Cool beings, that’s fine. You have the

1542
01:36:28,480 –> 01:36:32,320
freedom to do that. You live in America. You should be thanking whatever deity

1543
01:36:32,320 –> 01:36:35,920
you pray to that you live in America. Cool. Fine. Understand,

1544
01:36:36,480 –> 01:36:39,280
though, that your particular affinity group,

1545
01:36:41,840 –> 01:36:44,640
the history of your particular affinity group in this country

1546
01:36:45,920 –> 01:36:49,520
is a history that is, that is uniquely situated

1547
01:36:50,160 –> 01:36:52,900
in this historical context of this, this country.

1548
01:36:54,100 –> 01:36:56,980
So now I’m going to pick on the Columbia students. Now I’m going to go

1549
01:36:56,980 –> 01:37:00,660
for them. So when you’re talking about

1550
01:37:05,540 –> 01:37:08,100
bringing the Intifada to Columbia University,

1551
01:37:09,460 –> 01:37:13,260
you don’t know what you’re talking about. You

1552
01:37:13,260 –> 01:37:16,900
just don’t know what you’re talking about. You have no idea. You, you, you’ve now,

1553
01:37:17,300 –> 01:37:21,020
you’ve now taken a historical event that you may feel emotion

1554
01:37:21,020 –> 01:37:24,780
about or a series of historical events that you may feel emotion about

1555
01:37:24,780 –> 01:37:27,220
because of the way they were taught to you by a person who’s an expert

1556
01:37:27,220 –> 01:37:30,940
on those historical events. You’ve taken that emotion, you’ve

1557
01:37:30,940 –> 01:37:34,100
channeled it into political ideology, and you’ve come up with a

1558
01:37:34,100 –> 01:37:36,820
sloganeering idea that sounds good on paper.

1559
01:37:37,700 –> 01:37:41,140
Right. We’re going to internationalize or globalize the Intifada. Right.

1560
01:37:42,260 –> 01:37:45,910
And you don’t understand what

1561
01:37:45,910 –> 01:37:48,750
you’re asking for because someone like me

1562
01:37:49,630 –> 01:37:52,750
looks at that message and goes, okay.

1563
01:37:53,390 –> 01:37:56,830
And by the way, there’s no irony underneath that you’re deadly Sincere.

1564
01:37:57,070 –> 01:38:00,790
Okay, that’s cool. I’m deadly sincere about not having the

1565
01:38:00,790 –> 01:38:04,470
Intifada here because I know exactly what that word means. And the

1566
01:38:04,470 –> 01:38:08,030
history that that word represents inside of that historical

1567
01:38:08,030 –> 01:38:11,070
context over there. That has, by the way,

1568
01:38:11,950 –> 01:38:15,350
some of it does have something to do with us as America, but it has

1569
01:38:15,350 –> 01:38:19,190
a vast majority, vast majority of that has more

1570
01:38:19,190 –> 01:38:22,750
to do with how the European colonial powers dealt with the

1571
01:38:22,750 –> 01:38:26,270
breakup of the Ottoman Empire and dealt with Turkey

1572
01:38:26,270 –> 01:38:29,830
going all the way back to the rise of the Ottoman empire in like

1573
01:38:29,830 –> 01:38:31,310
1086 or something.

1574
01:38:33,870 –> 01:38:37,110
That’s history. That’s what I’m talking about. That poverty of

1575
01:38:37,110 –> 01:38:40,830
understanding, that inability to link ideas together and have second thoughts.

1576
01:38:41,250 –> 01:38:45,090
That’s the part that’s most troubling to me around why we don’t learn

1577
01:38:45,090 –> 01:38:47,970
from history. That’s the most troubling thing to me. And of course I’m going to

1578
01:38:47,970 –> 01:38:51,050
have people who are going to email me and say, well, hey, son, yes, we

1579
01:38:51,050 –> 01:38:54,690
absolutely do understand that. It goes to 1086. And you have to understand that colonialism,

1580
01:38:54,770 –> 01:38:58,250
daba dabba, dabba, systemic oppression. And I’m going to say, well, that’s one

1581
01:38:58,250 –> 01:39:01,770
interpretation. That’s just like your interpretation,

1582
01:39:01,770 –> 01:39:03,890
man. I have a different interpretation.

1583
01:39:05,970 –> 01:39:09,290
I want to see that you have an actual depth to your

1584
01:39:09,290 –> 01:39:13,090
interpretation. And I don’t think you do, because I don’t think you studied

1585
01:39:13,090 –> 01:39:13,370
history.

1586
01:39:17,210 –> 01:39:21,010
Yeah. Oh, I’m going all in on this one. I’m going

1587
01:39:21,010 –> 01:39:24,690
all in on this episode. I got nothing on this one. I’m going all in

1588
01:39:24,690 –> 01:39:28,290
on this one. I am, because I’m passionate about history and I’m

1589
01:39:28,290 –> 01:39:31,770
passionate about people actually learning these things and actually caring

1590
01:39:32,170 –> 01:39:35,610
in sincere ways about what the words are that they say

1591
01:39:35,970 –> 01:39:39,570
and understanding the historical context that, that they exist and that they exist

1592
01:39:39,570 –> 01:39:41,570
in. And also understanding that

1593
01:39:43,730 –> 01:39:47,530
the history of this country is not the history of everywhere else. And so

1594
01:39:47,530 –> 01:39:51,370
the kinds of quirks and things that you get to get away with here, if

1595
01:39:51,370 –> 01:39:54,890
I put you anywhere else, I mean, we’re talking about the Middle east right now,

1596
01:39:54,890 –> 01:39:58,130
but let’s talk about any, any country on the African continent.

1597
01:39:58,610 –> 01:40:00,530
If I put you there, you’re going to have, you’re going to have, you’re going

1598
01:40:00,530 –> 01:40:04,180
to have a problem because it’s a different thing there. And

1599
01:40:04,180 –> 01:40:07,860
so outsourcing that or

1600
01:40:07,860 –> 01:40:11,460
insourcing those ideas to here doesn’t. Doesn’t work. And

1601
01:40:11,460 –> 01:40:15,260
then the other, the other way doesn’t work either. Outsourcing those ideas to other

1602
01:40:15,260 –> 01:40:19,020
places. Like the conversations that we have in America about social justice

1603
01:40:19,100 –> 01:40:22,940
based on our particular historical understanding of what that means don’t work

1604
01:40:22,940 –> 01:40:26,740
in France and they don’t work in England because they come

1605
01:40:26,740 –> 01:40:30,300
from a different lineage and a different heritage and a different way of thinking about

1606
01:40:30,300 –> 01:40:34,100
both society and justice which gets

1607
01:40:34,100 –> 01:40:36,940
back to historical consciousness anyway.

1608
01:40:37,900 –> 01:40:41,700
And, and you know, so, so listen, this is, this is summed up very

1609
01:40:41,700 –> 01:40:44,900
easily, right? This is subbed up very easily. Yeah, go ahead. We just did two

1610
01:40:44,900 –> 01:40:48,540
hours. We did. And, and the idea, the concept

1611
01:40:48,540 –> 01:40:52,340
here is very simply we’re never going to learn

1612
01:40:52,340 –> 01:40:55,500
from history. It’s never going to happen because

1613
01:40:55,980 –> 01:40:59,830
of the, the, the emotions behind pride, power,

1614
01:41:00,070 –> 01:41:03,670
ego. That, that’s part of, it’s,

1615
01:41:03,670 –> 01:41:06,750
it’s human nature. It’s who we are as a people, as, as a, as a,

1616
01:41:06,750 –> 01:41:10,590
as an entity, as a being. We, we, we get our, we get

1617
01:41:10,590 –> 01:41:14,350
our pride hurt, we want to fight, we get, we, we, we,

1618
01:41:14,350 –> 01:41:17,950
we see something that we don’t have that somebody else has, we want it, we’re

1619
01:41:17,950 –> 01:41:21,270
gonna. Fight. Like it, it,

1620
01:41:21,670 –> 01:41:23,510
we’re never. Now that being said,

1621
01:41:25,520 –> 01:41:29,280
if we ever get to the point where, you know, there’s

1622
01:41:29,280 –> 01:41:32,080
been lots of talks about how, you know, over

1623
01:41:32,480 –> 01:41:36,120
evolution where like you know, we’re losing our pinky, we lost our

1624
01:41:36,120 –> 01:41:39,440
appendix because certain things happen and there’s some sci fi

1625
01:41:39,679 –> 01:41:43,240
theory out there that, that says that as we, as our

1626
01:41:43,240 –> 01:41:47,080
brains get bigger and we are urged to actually that that

1627
01:41:47,080 –> 01:41:50,650
fight or flight mechanism will go away and

1628
01:41:50,810 –> 01:41:54,650
the urge to fight for power will go away because we’ll have the intellect

1629
01:41:54,970 –> 01:41:57,850
to not have to do that anymore. Until that happens,

1630
01:41:59,130 –> 01:42:02,170
we’re never going to learn from history because we are going to fight over things

1631
01:42:02,170 –> 01:42:05,810
that are prideful for us. We’re going to fight over things that we want power

1632
01:42:05,810 –> 01:42:09,650
over. We’re going to fight over things that you know, our leadership

1633
01:42:09,650 –> 01:42:13,290
has a ego, ego trip over

1634
01:42:14,090 –> 01:42:17,480
because of their own, their own self interest or wor. Like

1635
01:42:17,640 –> 01:42:21,400
that’s going to continue to happen. So and, and

1636
01:42:21,400 –> 01:42:24,840
I don’t think any, there, there was a reference to,

1637
01:42:25,400 –> 01:42:29,160
there was a reference to the, the, the, the, the, the

1638
01:42:29,160 –> 01:42:32,600
military strategy room that goes a little beyond

1639
01:42:32,840 –> 01:42:36,040
the Joint Chiefs of Staffs in, in World War II in thinking about

1640
01:42:37,320 –> 01:42:41,000
Japan would never in a million years bomb Pearl harbor because they just know how

1641
01:42:41,330 –> 01:42:43,970
big and strong the United States is. And there was like a,

1642
01:42:45,010 –> 01:42:48,850
a believability bias that happened in there where even if one person said

1643
01:42:48,930 –> 01:42:52,730
but wait, the data shows this and they

1644
01:42:52,730 –> 01:42:56,370
go yeah, I don’t care what the data says, Japan will never do it. Right?

1645
01:42:57,169 –> 01:42:59,930
Japan’s never going to do that. We’re too big, we’re too big, we’re too big,

1646
01:42:59,930 –> 01:43:03,610
etc. Etc. We’re too strong, we’re too powerful, etc. Whatever. And

1647
01:43:03,610 –> 01:43:07,110
there’s a, there’s a report even of saying that when Japan’s,

1648
01:43:08,220 –> 01:43:12,060
when Japan’s planes were in, in route,

1649
01:43:12,700 –> 01:43:16,420
a radar radar technician went

1650
01:43:16,420 –> 01:43:20,140
to his superior officer and said hey, there’s a blip on this map. It looks

1651
01:43:20,140 –> 01:43:23,980
a little funny. And the radar got the, the, the superior officer said,

1652
01:43:23,980 –> 01:43:26,700
oh well, we’re, I, I think we’re expecting our

1653
01:43:26,700 –> 01:43:29,500
B17s coming back. Don’t worry about it.

1654
01:43:31,020 –> 01:43:34,750
An hour later, Pearl harbor doesn’t exist or you know, the rest is history.

1655
01:43:34,750 –> 01:43:38,590
So yeah, so there there. So even when there is a

1656
01:43:38,590 –> 01:43:40,670
singular voice of reason

1657
01:43:42,350 –> 01:43:46,150
because of like that consensus bias that happened

1658
01:43:46,150 –> 01:43:49,990
in that we still don’t listen to it. So even if there is an

1659
01:43:49,990 –> 01:43:53,790
intellect that comes above and beyond or goes beyond or our thinking and

1660
01:43:54,350 –> 01:43:57,390
again all the things we talked about today, whether it’s military,

1661
01:43:57,390 –> 01:44:01,160
geopolitical, whether it’s business related because you don’t think that the,

1662
01:44:01,240 –> 01:44:04,880
that somebody out there is going to an like the, the

1663
01:44:04,880 –> 01:44:08,160
VC world going hey can we put a little bit of a break on this,

1664
01:44:08,160 –> 01:44:11,680
on spending on these AI companies? Because there’s going to be a bubble somewhere. And

1665
01:44:11,680 –> 01:44:14,600
then the, the consensus bias still happens and goes

1666
01:44:15,080 –> 01:44:17,880
nah, you don’t know. What you’re talking about. Nah, you don’t know what you’re talking

1667
01:44:17,880 –> 01:44:19,400
about. You don’t know what you’re talking about.

1668
01:44:22,040 –> 01:44:25,850
So foreign. We’re

1669
01:44:25,850 –> 01:44:29,570
not solving this problem today, Hasan. No, we’re not solving this problem today.

1670
01:44:29,570 –> 01:44:33,090
But maybe, maybe, maybe that we can. Identify it

1671
01:44:33,170 –> 01:44:36,890
is probably tells me that there is much smarter people than us that

1672
01:44:36,890 –> 01:44:39,650
have already identified it and they couldn’t do anything about it.

1673
01:44:41,890 –> 01:44:45,690
Well so at least, at least what we can say is this. I think we

1674
01:44:45,690 –> 01:44:49,490
can say this. I think we could sum up the our two hour long conversation

1675
01:44:49,650 –> 01:44:53,290
in this, this, this two hour long episode around why We Don’t Learn

1676
01:44:53,290 –> 01:44:55,810
from History by BH Lart with this

1677
01:44:57,570 –> 01:45:01,290
I am a proponent and I think probably you are Tom, as

1678
01:45:01,290 –> 01:45:04,930
you Tom, are as well. I am a proponent of a. More,

1679
01:45:06,050 –> 01:45:09,650
some would say tragic but maybe that’s not it. A more

1680
01:45:10,210 –> 01:45:13,810
small C. Small C conservative view of history

1681
01:45:14,530 –> 01:45:18,330
rather than a large capital P progressive view of history

1682
01:45:18,730 –> 01:45:22,410
which is the kind of view of history that quite frankly

1683
01:45:22,730 –> 01:45:26,370
folks in the west have had since the Enlightenment going back

1684
01:45:26,370 –> 01:45:30,130
to the 17th century. This idea that everything will

1685
01:45:30,130 –> 01:45:33,970
just somehow consistently get better and that

1686
01:45:33,970 –> 01:45:36,250
history is merely the Tale of

1687
01:45:37,610 –> 01:45:41,460
endless improvement going forward. I mean, you see this with the techno future optimistic

1688
01:45:41,690 –> 01:45:45,370
right now, like Mark Andreessen and others that are. To your point, they’re sitting around

1689
01:45:45,370 –> 01:45:49,210
in all those rooms dumping all this money into AI because like, the future

1690
01:45:49,210 –> 01:45:52,490
will just be better. Maybe it will, maybe it won’t.

1691
01:45:53,370 –> 01:45:56,570
The tragic view of human nature. They also serve another

1692
01:45:57,450 –> 01:46:01,210
very vital problem with the. The whole theory of history teaching us

1693
01:46:01,290 –> 01:46:04,490
is because they feel if they throw enough money at it, it will eventually get

1694
01:46:04,490 –> 01:46:08,340
better regardless. Right? You’re right. Exactly. No matter how tragic human nature is,

1695
01:46:08,490 –> 01:46:11,930
is right. Right. And. And the thing is you can’t.

1696
01:46:13,850 –> 01:46:17,530
In taking a small c. Conservative view of. Of. Of human nature.

1697
01:46:17,530 –> 01:46:21,370
Not of human nature, sorry, of history, which I, I believe

1698
01:46:21,370 –> 01:46:25,130
I’m a proponent of. You.

1699
01:46:25,370 –> 01:46:29,210
You understand that no matter how many trillions of dollars you throw at human

1700
01:46:29,210 –> 01:46:32,410
nature. And yes, I did use trillions.

1701
01:46:34,180 –> 01:46:37,660
It doesn’t matter. Human nature is impervious to your

1702
01:46:37,660 –> 01:46:41,180
trillions. And human nature is a whack a

1703
01:46:41,180 –> 01:46:44,700
mole. It pops up in all kinds of odd

1704
01:46:44,700 –> 01:46:48,380
places and it. It pushes

1705
01:46:48,380 –> 01:46:52,220
and expands boundaries, usually against your

1706
01:46:52,220 –> 01:46:55,380
will and the will of your dollars. And by the way, the will of your

1707
01:46:55,380 –> 01:46:59,220
power. You know, you could say that

1708
01:46:59,870 –> 01:47:03,670
the trillions of dollars will do it, or rules and regulations will

1709
01:47:03,670 –> 01:47:07,110
do it, or laws will do it, or punishments and

1710
01:47:07,110 –> 01:47:10,030
consequences will do it. And the fact of the matter is,

1711
01:47:11,310 –> 01:47:14,990
a small c conservative reading of history

1712
01:47:16,350 –> 01:47:18,510
says that you’re actually incorrect.

1713
01:47:19,950 –> 01:47:23,790
You’re wrong. And you’re wrong with

1714
01:47:24,510 –> 01:47:27,720
a. You’re wrong. And you’re. And you’re one more.

1715
01:47:29,240 –> 01:47:32,920
You’re one more wealthy person or wealthy system

1716
01:47:33,960 –> 01:47:35,560
or wealthy set of institutions

1717
01:47:37,720 –> 01:47:41,000
that will wind up on the ash sheep of history

1718
01:47:41,800 –> 01:47:43,720
being read about by a bunch of people

1719
01:47:45,720 –> 01:47:49,320
four or five hundred years from now going, why those people didn’t get it.

1720
01:47:50,200 –> 01:47:50,680
Yeah.

1721
01:47:54,290 –> 01:47:58,130
And so with that, there’s no upper.

1722
01:47:58,290 –> 01:48:01,650
Upper on this episode of the podcast. There’s a way to end with this. I

1723
01:48:01,650 –> 01:48:04,210
think we’re done. I think we’re just. We’re just done. Wait, hold on. I think

1724
01:48:04,210 –> 01:48:04,770
we’re done here

1725
01:48:10,930 –> 01:48:14,210
with that. We’re. We’re out.