Why Don’t We Learn From History by B.H. Liddell Hart w/Tom Libby & Jesan Sorrells
—
Delving into B.H. Liddell Hart’s Why Don’t We Learn From History, hosted by Jesan Sorrells and Tom Libby, examines why societies—and leaders in business, politics, and communities—struggle to absorb history’s lessons. They discuss the enduring influence of emotional bias, the pitfalls of democratic and aristocratic decision-making, and the critical role of passionate history teaching in shaping informed citizens and leaders. The episode draws connections from military history to modern business trends, showing how historical context and human nature impact leadership choices and organizational outcomes.
- Book Title: Why Don’t We Learn From History
- Author: B.H. Liddell Hart
- Guest Names: Jesan Sorrells (host), Tom Libby (co-host)
—
Time Stamped Overview
—
00:00 Welcome and Introduction – Why Don’t We Learn From History by B.H. Liddell Hart.
01:00 “Cracks in the Human Condition.”
09:17 Informal Committees Shape Decisions.
11:07 “Psychology, Memory, and Bias”
18:29 “Majority Shapes Right and Wrong.”
25:14 “Negotiating Reality in America.”
30:30 “Systems, Democracy, and Global Chaos.”
35:17 Historical Bias and Emotional Impact.
39:29 AI Adoption: Reality vs Expectations.
44:20 “AI Essential for Business Success.”
48:27 “Pre-WWII Polish Guarantee Debate.”
56:42 Post-COVID Education System Concerns.
59:43 “Making Poetry and History Alive.”
01:06:26 Teaching Complexity in History.
01:11:23 “Drunk on Ideas, Not Rote.”
01:16:12 “Perspective and Counterfactual Insights.”
01:23:39 “The Perils of Intellectual Neglect.”
01:24:54 “Emotion, Truth, and Progress.”
01:31:28 “Misguided Solidarity and History.”
01:39:23 Passion for History and Context.
01:41:20 “Cycle of Pride and Conflict.”
01:46:57 “Wealth Fails to Ensure Legacy.”
—
Opening and closing themes composed by Brian Sanyshyn of Brian Sanyshyn Music.
—
- Pick up your copy of 12 Rules for Leaders: The Foundation of Intentional Leadership NOW on AMAZON!
- Check out the 2022 Leadership Lessons From the Great Books podcast reading list!
—
★ Support this podcast on Patreon ★
00:00:01,360 –> 00:00:05,000
Hello, my name is Jesan Sorrells, and
2
00:00:05,000 –> 00:00:08,480
this is the Leadership Lessons from the Great Books podcast,
3
00:00:08,720 –> 00:00:12,560
episode number 167,
4
00:00:13,760 –> 00:00:17,600
which puts us at about. Let’s see, how
5
00:00:17,600 –> 00:00:20,560
many episodes is that? Eight.
6
00:00:21,040 –> 00:00:24,640
Yeah. Eight episodes from our 175th
7
00:00:25,120 –> 00:00:28,910
episode. So we should get there by the end of this year. And thank you
8
00:00:28,910 –> 00:00:31,430
all for joining us on this journey. All right,
9
00:00:32,710 –> 00:00:35,990
so on this show, my co host
10
00:00:36,390 –> 00:00:39,670
who’s joining me today, Tom and I have
11
00:00:40,070 –> 00:00:43,830
circled around and around and around, coming back
12
00:00:44,070 –> 00:00:47,750
similar to the Nietzschean Myth of Return, to
13
00:00:47,990 –> 00:00:51,670
a core idea that is reflected in all of the books
14
00:00:51,800 –> 00:00:55,480
that we talk about on this show. And it is
15
00:00:55,480 –> 00:00:58,880
the idea that, and Tom is probably going to say it again today, the more
16
00:00:58,880 –> 00:01:02,680
things change, the more they stay the same.
17
00:01:03,560 –> 00:01:06,920
And this is true. It’s. It’s what people in Washington,
18
00:01:06,920 –> 00:01:10,760
D.C. pre Covid used to call a true fact.
19
00:01:10,920 –> 00:01:13,480
I don’t know what they call it now, but it is definitely true,
20
00:01:14,600 –> 00:01:18,430
which is why we keep returning to this idea repeatedly, no matter the
21
00:01:18,430 –> 00:01:21,630
book, no matter the genre, and no matter the author.
22
00:01:22,990 –> 00:01:26,790
In one way, if you’re listening to this, you could conclude that
23
00:01:26,790 –> 00:01:30,150
this fact of return reflects something inherent in the human condition
24
00:01:30,150 –> 00:01:33,989
itself. I think Richard Messing, who
25
00:01:33,989 –> 00:01:37,590
came on and talked with us about man’s search for meaning, and Viktor Frankl
26
00:01:37,590 –> 00:01:41,070
might say that. Right. In another
27
00:01:41,070 –> 00:01:44,470
way, you could say that this fact of noticing and
28
00:01:44,470 –> 00:01:47,950
recognizing such a conclusion reflects the idea that we are
29
00:01:47,950 –> 00:01:51,430
aware, deeply so, of the broken parts and
30
00:01:51,430 –> 00:01:55,190
cracks in the facade of how we address the clearly
31
00:01:55,190 –> 00:01:57,590
broken parts of the human condition.
32
00:01:58,790 –> 00:02:02,590
And it may indicate that noticing the how opens up an
33
00:02:02,590 –> 00:02:06,310
opportunity for all of us as leaders to seek and to
34
00:02:06,310 –> 00:02:09,950
explore and to maybe try to get a glimpse of the
35
00:02:09,950 –> 00:02:13,670
light that lies behind the cracks in the human
36
00:02:13,910 –> 00:02:17,750
condition. So today on the show we are
37
00:02:17,750 –> 00:02:21,470
covering, we’re going to talk about topics from a book
38
00:02:21,470 –> 00:02:25,190
that does its best and the best that it can
39
00:02:25,670 –> 00:02:29,430
to engage in noticing the cracks in the human face of
40
00:02:29,430 –> 00:02:32,710
the problems and challenges of the human condition. And
41
00:02:33,110 –> 00:02:36,910
a book that tries to examine and really talk about what lies behind those cracks
42
00:02:36,910 –> 00:02:40,620
in a coherent fashion. So let’s start on
43
00:02:40,620 –> 00:02:44,380
our journey repeating probably same things
44
00:02:44,380 –> 00:02:48,140
we’ve repeated before on this show, although there’s always new people joining us,
45
00:02:48,140 –> 00:02:51,820
so it’s always new to you. Today we will
46
00:02:51,820 –> 00:02:55,340
be covering the book and the question
47
00:02:55,500 –> 00:02:59,100
posed by it. Oh, I love that. Why don’t we Learn
48
00:02:59,100 –> 00:03:02,780
From History by B.H. liddell
49
00:03:03,180 –> 00:03:07,030
Hart? By the way, my book has a yellow cover.
50
00:03:07,590 –> 00:03:10,470
Your mileage will vary. I do like the yellow cover.
51
00:03:11,670 –> 00:03:15,350
Leaders. The penultimate question of all time, which
52
00:03:15,350 –> 00:03:19,110
is the question of why don’t we learn from history, has nothing
53
00:03:19,110 –> 00:03:21,910
but hard answers to it.
54
00:03:23,830 –> 00:03:27,590
And of course, on my journey through this book,
55
00:03:27,990 –> 00:03:31,420
we will be joined today by our regular co host, Tom Libby.
56
00:03:31,730 –> 00:03:35,490
How you doing today, Tom? Doing well, Jesan. And you know, I gotta be
57
00:03:35,490 –> 00:03:38,850
honest, I’m not surprised in the least that you selected me to help you with
58
00:03:38,850 –> 00:03:41,730
this book because to your point, I think I’ve said
59
00:03:42,530 –> 00:03:45,810
at least two dozen times on this podcast
60
00:03:46,130 –> 00:03:49,570
that phrase that the more things change, the more things stay the same. So I
61
00:03:49,570 –> 00:03:53,370
appreciate, I appreciate you tapping me in on this one because I think we’re
62
00:03:53,370 –> 00:03:57,130
gonna have a lot of fun with the, with this particular subject matter.
63
00:03:57,130 –> 00:04:00,810
So, so appreciate it. Yeah, absolutely. Well, I would encourage everybody to go and
64
00:04:00,810 –> 00:04:04,050
listen to the intro episode where we talk a little bit about
65
00:04:04,690 –> 00:04:06,850
BH Lidell Hart as a writer.
66
00:04:08,530 –> 00:04:12,210
He was, he was an interesting guy just
67
00:04:12,210 –> 00:04:15,730
in general. He was part of the generation of
68
00:04:15,730 –> 00:04:18,210
folks that was born in the,
69
00:04:19,570 –> 00:04:23,250
in the late 19th century, came of age
70
00:04:23,410 –> 00:04:26,130
in the early 20th century,
71
00:04:27,310 –> 00:04:30,990
fought in World War I. And just like C.S. lewis and
72
00:04:30,990 –> 00:04:34,110
J.R.R. tolkien and Eric Remarque
73
00:04:34,750 –> 00:04:38,550
and other notable names like T.E. lawrence, Lawrence of
74
00:04:38,550 –> 00:04:41,150
Arabia, Winston Churchill,
75
00:04:42,589 –> 00:04:46,270
and of course, your friend and mine, Adolf Hitler,
76
00:04:48,590 –> 00:04:52,350
he learned all kinds of interesting lessons
77
00:04:52,350 –> 00:04:55,750
about history, BH Lidelhart did, and about human
78
00:04:55,750 –> 00:04:59,590
nature in the trenches of the Somme
79
00:04:59,670 –> 00:05:03,150
and of Verdun and at Verdun. And we
80
00:05:03,150 –> 00:05:06,750
explore a lot of that history on that intro episode. So I would
81
00:05:06,750 –> 00:05:10,070
encourage you to, to go and take a listen to it. Now. We’re not going
82
00:05:10,070 –> 00:05:13,750
to dive so much into, into Liddell Hart’s as a person today
83
00:05:13,750 –> 00:05:16,710
on the show. Instead, we’re going to really focus on, we’re really going to focus
84
00:05:16,710 –> 00:05:20,190
on the ideas in the book. And, and when you open up, why don’t we
85
00:05:20,190 –> 00:05:23,750
learn from history what you see. And I mentioned this in the intro
86
00:05:23,750 –> 00:05:27,350
episode as well. What you see is that the
87
00:05:27,350 –> 00:05:30,950
preface was written by his son, which is great. The version that I have,
88
00:05:30,950 –> 00:05:34,790
Adrian J. Liddell Hart, who actually made a name for himself in
89
00:05:34,790 –> 00:05:37,950
World War II and in the military as well.
90
00:05:38,510 –> 00:05:41,190
And when you open up the table of contents, you see that the book is
91
00:05:41,190 –> 00:05:44,270
divided into three parts, right? There’s
92
00:05:44,670 –> 00:05:48,030
History and Truth, which is a great way to start a book
93
00:05:48,480 –> 00:05:52,280
entitled why Don’t We Learn from History, Government and Freedom and then
94
00:05:52,280 –> 00:05:56,120
War and Peace. And so as a military historian, you would think that he would
95
00:05:56,120 –> 00:05:59,520
start with War and Peace first, but no, no, he saves that for the end.
96
00:06:00,160 –> 00:06:02,640
And each One of the
97
00:06:03,520 –> 00:06:07,120
sections features a short essay, usually no more than four
98
00:06:07,120 –> 00:06:10,800
pages, where he lays out his ideas very succinctly.
99
00:06:11,040 –> 00:06:14,080
And this book really, Tom, really
100
00:06:15,570 –> 00:06:19,370
sort of puts me in mind of the essay
101
00:06:19,370 –> 00:06:23,050
by George Orwell that we covered where he talked about English literature.
102
00:06:23,050 –> 00:06:26,890
Right. It’s, it’s sort of the clearest example of clear writing that I’ve
103
00:06:26,890 –> 00:06:30,649
seen in a while. And, and partially that’s because they’re both, you
104
00:06:30,649 –> 00:06:33,570
know, they’re both English. They both came out of the English, you know, writing,
105
00:06:35,570 –> 00:06:39,410
thinking, literature structure that was built in Europe.
106
00:06:39,410 –> 00:06:43,220
Right. So they’re products of that. They’re products of that European and
107
00:06:43,220 –> 00:06:46,860
English tradition that goes back, you know, well over
108
00:06:46,940 –> 00:06:50,300
1500 years of just getting clarity in writing.
109
00:06:52,220 –> 00:06:55,580
And as a military historian, Liddell Hart really pursued
110
00:06:56,140 –> 00:06:59,860
clarity in thinking and clarity in writing. And so when you read
111
00:06:59,860 –> 00:07:03,500
these essays, he doesn’t mince words. There’s no
112
00:07:03,500 –> 00:07:05,660
fat in anything in here.
113
00:07:07,670 –> 00:07:11,350
So I’m going to open up with
114
00:07:11,670 –> 00:07:15,150
the section Restraints of Democracy and Power
115
00:07:15,150 –> 00:07:18,990
Politics in a Democracy. And this essay is in Government
116
00:07:18,990 –> 00:07:22,390
and Freedom, which is in the, the second
117
00:07:22,710 –> 00:07:26,550
section of his essays. And each one of the
118
00:07:26,550 –> 00:07:29,350
essays does build on the other one. So we’re going to pull them out separately,
119
00:07:29,350 –> 00:07:32,390
we’re going to talk about them in relation to why don’t we learn from history?
120
00:07:32,550 –> 00:07:36,310
But there’s lessons for leaders in all of this, you know, whether you’re a
121
00:07:36,310 –> 00:07:40,060
civic leader, leader of a nonprofit, leader of a for profit, or
122
00:07:40,060 –> 00:07:42,100
even just a leader of your family or community.
123
00:07:43,700 –> 00:07:47,380
So let’s go to the book Restraints
124
00:07:47,380 –> 00:07:51,020
of Democracy, and I quote, we
125
00:07:51,020 –> 00:07:54,300
learn from history that democracy has commonly put a premium on
126
00:07:54,300 –> 00:07:58,100
conventionality. By its nature, it prefers those who keep
127
00:07:58,100 –> 00:08:01,940
in step with the slowest march of thought and frowns on
128
00:08:01,940 –> 00:08:05,540
those who may disturb, quote, unquote, the conspiracy for mutual
129
00:08:05,540 –> 00:08:09,260
inefficiency. Thereby, this system
130
00:08:09,260 –> 00:08:12,820
of government tends to result in the triumph of mediocrity and
131
00:08:12,820 –> 00:08:16,380
entails the exclusion of first rate ability. If this is combined
132
00:08:16,380 –> 00:08:20,180
with honesty. But the alternative to it,
133
00:08:20,180 –> 00:08:23,580
despotism, almost inevitably means the triumph of
134
00:08:23,660 –> 00:08:27,420
stupidity. And of the two evils, the former
135
00:08:27,660 –> 00:08:31,420
is less. Hence it is better that ability should
136
00:08:31,420 –> 00:08:35,220
consent to its own sacrifice and subordination to the regime of
137
00:08:35,220 –> 00:08:38,660
mediocrity. Rather than assist in establishing a regime where,
138
00:08:38,900 –> 00:08:42,340
in the light of past experience, brute stupidity will be enthroned
139
00:08:42,820 –> 00:08:46,340
and ability may only preserve its footing at the price of
140
00:08:46,340 –> 00:08:49,780
dishonesty. What is the value
141
00:08:50,020 –> 00:08:53,700
in England and America and what is of value
142
00:08:53,700 –> 00:08:57,300
in England and America and worth defending its tradition of freedom, the
143
00:08:57,300 –> 00:09:01,020
guarantee of its vitality. Our civilization, like the Greek, has,
144
00:09:01,020 –> 00:09:04,620
for all its blundering way, taught the value of freedom, of criticism, of authority,
145
00:09:04,620 –> 00:09:08,460
and of harmonizing this with order. Anyone who urges a different
146
00:09:08,460 –> 00:09:11,780
system for efficiency’s sake is betraying the vital
147
00:09:11,780 –> 00:09:15,580
tradition. Then he switches to power
148
00:09:15,580 –> 00:09:19,020
politics in a democracy. And I want to point this out.
149
00:09:19,340 –> 00:09:23,070
He says this talking about how decisions get
150
00:09:23,070 –> 00:09:26,830
made in a democratic government. He says this while committee meetings are
151
00:09:26,830 –> 00:09:30,390
not so frequently held in the late afternoon as in the morning, dinner
152
00:09:30,390 –> 00:09:34,070
itself provides both an opportunity and an atmosphere suited to the informal kind of committee
153
00:09:34,230 –> 00:09:37,749
that tends to be more influential than those which are formally constituted.
154
00:09:38,150 –> 00:09:41,430
The informal type is usually small, and the smaller it is, the more
155
00:09:41,430 –> 00:09:45,150
influential it may be. The two or three gathered
156
00:09:45,150 –> 00:09:48,950
together may outweigh a formal committee of 20 or 30 members, to which
157
00:09:48,950 –> 00:09:52,710
it may often be related under the blanket, where it is assembled by
158
00:09:52,710 –> 00:09:56,550
someone who has a leading voice in a larger official committee. For it
159
00:09:56,550 –> 00:10:00,310
will represent his personal selection in the way of consultants. And
160
00:10:00,310 –> 00:10:03,990
its members being chosen for their congeniality as well as for their advisory
161
00:10:03,990 –> 00:10:07,470
value is likely to reach clear cut conclusions, which in turn may be
162
00:10:07,470 –> 00:10:11,110
translated into the decisions of a formal committee. For at any
163
00:10:11,110 –> 00:10:14,590
gathering of 20 or 30 men there is likely to be so much diversity
164
00:10:14,830 –> 00:10:18,390
and nebulosity of views that the consent of the majority can
165
00:10:18,390 –> 00:10:21,730
generally be gained for any conclusion that is sufficiently definite and
166
00:10:21,880 –> 00:10:25,480
impressively backed by well considered arguments and sponsored by a heavyweight
167
00:10:25,480 –> 00:10:28,760
member, especially if the presentation is
168
00:10:28,760 –> 00:10:32,360
carefully stage managed.
169
00:10:36,840 –> 00:10:37,640
I love that
170
00:10:40,440 –> 00:10:43,280
we don’t have to talk about power politics because we, we don’t like to talk
171
00:10:43,280 –> 00:10:47,080
about that. But this still, especially in the
172
00:10:47,080 –> 00:10:50,720
current political landscape. What you just
173
00:10:50,720 –> 00:10:54,380
read could probably blow up half the people’s brains in country right
174
00:10:54,380 –> 00:10:58,060
now. Well, so let’s open
175
00:10:58,060 –> 00:11:00,820
up with that. The very first question that I have from here,
176
00:11:01,860 –> 00:11:04,420
Tom, why don’t we learn from history?
177
00:11:05,940 –> 00:11:09,620
Get right into it. Good lord. So I,
178
00:11:09,620 –> 00:11:12,900
I think there’s so many factors here and I, I. One thing that I will
179
00:11:12,900 –> 00:11:16,580
say, and I know we’re not going to get into Liddell’s life per
180
00:11:16,580 –> 00:11:20,110
se, but I wonder how
181
00:11:20,110 –> 00:11:23,910
much more, how much more impact his
182
00:11:23,910 –> 00:11:27,430
book would have had if he had today’s access to
183
00:11:27,430 –> 00:11:31,070
psychology and like the psychological research, like some of the
184
00:11:31,070 –> 00:11:34,630
psychological research behind some of the stuff that he talks about is like,
185
00:11:35,190 –> 00:11:38,630
it’s actually because, because of his book, I’ve, I’ve seen several
186
00:11:39,590 –> 00:11:40,950
research, research
187
00:11:44,870 –> 00:11:48,470
papers been done basically because of this. Right. So in,
188
00:11:48,630 –> 00:11:52,430
so to go back to your, to your point and, and I don’t know what
189
00:11:52,430 –> 00:11:55,830
the technical terms for them are, but like there is something to be said about,
190
00:11:55,830 –> 00:11:57,510
like, about
191
00:11:59,670 –> 00:12:03,510
memory bias, right? So like, so we sometimes don’t
192
00:12:03,510 –> 00:12:07,190
learn from history because quite honestly we’re only, we’re
193
00:12:07,190 –> 00:12:10,790
very biased to the history that we read, right? So take, I mean
194
00:12:10,790 –> 00:12:14,560
US History is a very good example of this when you go and you start
195
00:12:14,560 –> 00:12:16,080
reading, if you were to read.
196
00:12:18,240 –> 00:12:20,480
And the other part of it too, and I think he talks about it a
197
00:12:20,480 –> 00:12:24,000
little bit in the book where the, because the victor
198
00:12:24,000 –> 00:12:27,760
usually writes the majority of, of passages when it comes
199
00:12:27,760 –> 00:12:31,360
to enter your subject matter here, whether it’s
200
00:12:31,600 –> 00:12:35,040
War one, World War two, it could be the cola wars for all I care.
201
00:12:35,120 –> 00:12:38,880
It doesn’t matter. Like whatever, whatever conflict
202
00:12:38,880 –> 00:12:42,580
or situation that you’re talking about, it’s usually the, to
203
00:12:42,580 –> 00:12:46,180
the victor goes the spoils, right? So somebody who wins that fight or
204
00:12:46,180 –> 00:12:50,020
wins that race or wins that whatever is going to write.
205
00:12:50,020 –> 00:12:53,860
You’re going to pay more attention to their writings. Therefore, you’re going
206
00:12:53,860 –> 00:12:57,460
to see the results of the victory and you’re not going to, you have a
207
00:12:57,460 –> 00:13:01,020
very, you have a very, very conscious bias of what
208
00:13:01,020 –> 00:13:04,820
history looks like. So you tend to, to not worry about.
209
00:13:04,980 –> 00:13:08,740
Good example. Again, you are talking about the current political landscape that
210
00:13:08,740 –> 00:13:12,220
we’re talking about. Everybody. You can go back in history
211
00:13:12,220 –> 00:13:16,060
and pinpoint times in history where you can say
212
00:13:16,300 –> 00:13:19,420
Hitler gained control of
213
00:13:19,900 –> 00:13:23,380
the political catastrophe that
214
00:13:23,380 –> 00:13:27,140
happened in Germany. We view it now as
215
00:13:27,140 –> 00:13:30,780
a political catastrophe. Well, guess what? Because all the Allies wrote all the,
216
00:13:31,260 –> 00:13:34,780
wrote all the, all the history, right? So but
217
00:13:35,350 –> 00:13:39,070
at the moment and in the time frame of the, in Germany
218
00:13:39,070 –> 00:13:42,670
when all that was happening, the, the people in the moment did not view this
219
00:13:42,670 –> 00:13:46,390
as a, as a, necessarily a bad thing. Were there people that were
220
00:13:46,390 –> 00:13:50,150
like, hey, wait a second, should this really be happening? Maybe,
221
00:13:50,150 –> 00:13:53,990
but they, their voice was never heard because a vast
222
00:13:53,990 –> 00:13:57,670
majority of people, he was a charismatic speaker, people he followed,
223
00:13:57,670 –> 00:14:01,350
people followed him, et cetera, et cetera. And nobody ever saw it coming, so to
224
00:14:01,350 –> 00:14:04,890
speak. Yeah, everybody thought they could do
225
00:14:05,050 –> 00:14:08,410
a deal with that guy. I mean even, even
226
00:14:09,850 –> 00:14:13,650
not Joseph Kennedy. Charles Lindbergh who ran for President. Charles Lindbergh
227
00:14:13,650 –> 00:14:17,410
who ran for President. Henry Ford, right. Who wanted
228
00:14:17,410 –> 00:14:21,130
to. Not wanted to, but helped. I
229
00:14:21,130 –> 00:14:24,930
believe it was either Mercedes or it
230
00:14:24,930 –> 00:14:28,210
might have been IG Farben. I can’t remember who he helped out, but he went
231
00:14:28,210 –> 00:14:30,890
over to, he did, he went over to Germany and he helped him set up
232
00:14:31,150 –> 00:14:34,950
factories to like for the purposes of re. Industrialization after
233
00:14:34,950 –> 00:14:38,110
the, to your point, the disaster of the Weimar Republic and inflation.
234
00:14:38,430 –> 00:14:42,190
Right. But Henry Ford was another guy who thought,
235
00:14:42,270 –> 00:14:45,950
yeah, okay, you know, we can, we can deal with this guy. Even Joe
236
00:14:45,950 –> 00:14:49,790
Kennedy, John Kennedy’s and Ted Kennedy’s dad,
237
00:14:50,270 –> 00:14:54,110
Joe Kennedy got in trouble. I believe he was the ambassador
238
00:14:54,190 –> 00:14:57,870
to England from America during
239
00:14:57,870 –> 00:15:00,830
the Roosevelt administration. He got in trouble
240
00:15:01,990 –> 00:15:05,710
just before the war kicked off by basically saying, hey, you know what?
241
00:15:05,710 –> 00:15:09,350
This Hitler guy, he’s not terrible. Like, we
242
00:15:09,350 –> 00:15:12,830
could probably do a deal with him. It’s fine. Hell, Stalin did a deal with
243
00:15:12,830 –> 00:15:16,510
Hitler. So, so what is the rest of the world looking at our
244
00:15:16,510 –> 00:15:20,270
current administration as? And there’s. There are people in our
245
00:15:20,270 –> 00:15:24,110
country that thinks that that’s, that, that we’re watching history repeat itself as
246
00:15:24,110 –> 00:15:27,910
we speak now. There are. And I want to, I want
247
00:15:27,910 –> 00:15:30,710
to talk a little about that today, because reading this book in the context of
248
00:15:30,710 –> 00:15:34,540
the current political climate that we are in was extremely
249
00:15:34,540 –> 00:15:37,460
interesting. Yeah. There’s something that I, I think
250
00:15:38,260 –> 00:15:41,980
there’s stark differences that. Oh, yeah, what was in place? Like, we have
251
00:15:41,980 –> 00:15:45,540
some checks and balances and we have things that like that are, that our government.
252
00:15:45,940 –> 00:15:49,700
I, I don’t see, I don’t see our current
253
00:15:49,700 –> 00:15:53,460
administration turning into Hitler, but there are people in our country that think that
254
00:15:53,860 –> 00:15:57,580
there are. There are. Correct. Right. Because. Because some of the signs and stuff are
255
00:15:57,580 –> 00:16:01,150
there. But, but again, where Germany didn’t have checks and
256
00:16:01,150 –> 00:16:04,790
balances, we do. The science can be there all they want as long
257
00:16:04,790 –> 00:16:08,110
as our government operates as the way that they’re supposed to
258
00:16:08,110 –> 00:16:11,750
operate. We don’t, we, we’re not going to have that. It’s, it’s,
259
00:16:11,750 –> 00:16:14,789
it’s also, whether you like. Him or not is not my point. I could care
260
00:16:14,789 –> 00:16:17,190
less whether you like him or not. That’s not, that’s not what I’m getting at
261
00:16:17,190 –> 00:16:20,190
here. Yeah, that’s not what we’re getting at here. What we’re talking about also. And
262
00:16:20,190 –> 00:16:24,030
this gets to, to what I just read there. So the
263
00:16:24,030 –> 00:16:26,990
ways in which most European
264
00:16:27,890 –> 00:16:31,570
governments were set up even after World War I,
265
00:16:31,570 –> 00:16:34,610
everything cracked apart after World War I, but
266
00:16:35,490 –> 00:16:38,930
certain ways still struggled on. Even 80 years
267
00:16:39,250 –> 00:16:42,770
later in our time, there’s still evidence of this.
268
00:16:43,010 –> 00:16:46,770
So Europe and England and specifically
269
00:16:47,010 –> 00:16:48,610
European countries like France,
270
00:16:51,250 –> 00:16:54,910
Russia, not so
271
00:16:54,910 –> 00:16:58,430
much Spain, although you could throw a spade in there. Italy for sure.
272
00:16:58,510 –> 00:17:02,270
And of course, Germany come out of a
273
00:17:02,350 –> 00:17:05,230
concept or have a, have an inbuilt concept of
274
00:17:05,470 –> 00:17:08,910
aristocratic rule that we don’t have.
275
00:17:10,350 –> 00:17:14,190
We explicitly rejected that. And so
276
00:17:14,350 –> 00:17:18,190
in the United States, our founding explicitly rejects aristocratic rule.
277
00:17:18,830 –> 00:17:22,450
This is where the, and these
278
00:17:22,450 –> 00:17:25,690
people, maybe they had a point, maybe they didn’t, but that the protesters this summer
279
00:17:25,690 –> 00:17:29,250
were talking about no kings or whatever. Like. Well,
280
00:17:29,570 –> 00:17:33,330
I mean, if you look at the Constitution and if you look at the three,
281
00:17:34,610 –> 00:17:38,210
the three branches of government, like they’re
282
00:17:38,210 –> 00:17:42,010
functioning exactly constitutionally as they should be. You just don’t like
283
00:17:42,010 –> 00:17:45,770
the decisions that they’re making. Which is, which is, which is why you
284
00:17:45,770 –> 00:17:49,490
get to vote. That’s why you, that’s why, that’s
285
00:17:49,490 –> 00:17:53,200
why, like we were talking about with, with Charlie Kirk’s assassination
286
00:17:53,200 –> 00:17:56,680
and around that this is why freedom of speech matters. Because which,
287
00:17:56,920 –> 00:18:00,320
which by the way, we also don’t have a tradition of in a European context.
288
00:18:00,320 –> 00:18:03,400
And so how decisions get made in an
289
00:18:03,400 –> 00:18:07,240
aristocratic, with an aristocratic mindset, in an aristocratic
290
00:18:07,240 –> 00:18:11,080
manner is fundamentally different than how decisions get made in a more. And this is
291
00:18:11,080 –> 00:18:14,720
point that Liddell Hart’s making in a more democratic mindset. So when he
292
00:18:14,720 –> 00:18:18,440
talks about a small cadre of people making decisions and then basically stage
293
00:18:18,440 –> 00:18:22,020
managing them for 20 or 30 other people, that comes. And
294
00:18:22,020 –> 00:18:25,300
Liddell Hart was English. Comes out of a specific aristocratic
295
00:18:25,300 –> 00:18:28,780
mindset. Yeah, yeah. That we don’t have.
296
00:18:29,340 –> 00:18:32,860
So, so as I started this conversation where I said
297
00:18:33,100 –> 00:18:36,860
I wish Lidell had access to some of the research that, that
298
00:18:36,860 –> 00:18:40,580
has, that psych psychology has come leaps and bounds over the last, you know,
299
00:18:40,580 –> 00:18:44,140
80 years since, since World War II. So, so there’s, there’s that,
300
00:18:44,220 –> 00:18:47,980
there’s that, that, you know, that hindsight bias that we have
301
00:18:48,320 –> 00:18:52,160
because, because history is written by the victors most times.
302
00:18:52,560 –> 00:18:56,280
And again, we’re getting better at that, but not. Still not great. But then
303
00:18:56,280 –> 00:18:59,960
there you have the other. I, I was once told that the difference
304
00:18:59,960 –> 00:19:03,760
between right and wrong is the majority. And that
305
00:19:03,760 –> 00:19:07,440
really, that really hit me hard too because now you’re saying to me
306
00:19:07,440 –> 00:19:11,040
that I could know that one plus one, we go back to Orwell, right?
307
00:19:11,120 –> 00:19:14,080
I could know that one plus one or two plus two equals four. But you’re
308
00:19:14,080 –> 00:19:16,960
going to say that the majority of people say it’s five. So now it’s just
309
00:19:16,960 –> 00:19:20,180
right. Five is right from now on. But that in.
310
00:19:20,740 –> 00:19:24,180
When it comes to something so linear as
311
00:19:24,180 –> 00:19:27,980
math, maybe you can make arguments against it. But when it comes
312
00:19:27,980 –> 00:19:31,220
to something that’s either opinion or, or
313
00:19:32,500 –> 00:19:36,300
consensus based or like, there’s a lot, there are a
314
00:19:36,300 –> 00:19:40,060
lot of situations where the right thing to do and this
315
00:19:40,060 –> 00:19:43,300
is where we get the whole Democratic vote, right? So you’re going to vote 100
316
00:19:43,300 –> 00:19:47,090
people vote or you know, 200 people vote and, or
317
00:19:47,090 –> 00:19:50,130
100 people vote and 51 of them say this. So we’re just going to do
318
00:19:50,130 –> 00:19:53,810
that. And 49 of them can know damn well that it’s the wrong thing to
319
00:19:53,810 –> 00:19:57,650
do. But the right. The difference between right and wrong is the
320
00:19:57,650 –> 00:20:01,490
majority right. Like, so there’s also some of that that happens throughout the
321
00:20:01,490 –> 00:20:04,810
course of history. And then there’s the final one that
322
00:20:05,290 –> 00:20:07,130
I think of, quite honestly is
323
00:20:09,050 –> 00:20:12,880
there’s a disassociation of time that happens. Right. The further
324
00:20:12,880 –> 00:20:16,160
away from something we get, we become more arrogant that
325
00:20:16,800 –> 00:20:19,800
we can see it, we know it’s happening. We’re not going to make the same
326
00:20:19,800 –> 00:20:23,520
mistakes because we know they’re there, but yet we do because we have a bias
327
00:20:23,920 –> 00:20:27,680
of disassociation of time. And I’ll give you an example of this one.
328
00:20:27,920 –> 00:20:31,720
Now, for those of you who can’t see me on the video here, I’m
329
00:20:31,720 –> 00:20:35,480
not a woman, so I’m not speaking as a woman. But childbirth, to me is
330
00:20:35,480 –> 00:20:39,320
a very good example of this on an individual basis. You’re a parent,
331
00:20:39,320 –> 00:20:42,200
I’m a parent. I don’t know if you spent time in the delivery room with
332
00:20:42,200 –> 00:20:45,640
your wife, but I did. Yeah. I watched what she went through
333
00:20:45,800 –> 00:20:49,440
and I went. Why would anybody do this more than once? Like,
334
00:20:49,440 –> 00:20:52,960
why? Like, honestly, like, the amount of the, the. The. The mental
335
00:20:52,960 –> 00:20:56,520
anguish, the physical anguish, the, the pain they go through. Like,
336
00:20:57,160 –> 00:21:00,880
it, it. Childbirth, to me, is one of the most fascinating things
337
00:21:00,880 –> 00:21:04,440
on the. In the entire natural world. Because
338
00:21:05,530 –> 00:21:08,810
women decide to do this again. Like, Right. I know, I know.
339
00:21:09,210 –> 00:21:12,410
Let me just say this for, for the record. I know,
340
00:21:12,650 –> 00:21:16,330
guys, if we went through that once, we’d be like, hell, no, we’re
341
00:21:16,330 –> 00:21:19,970
done. Nope, because we wouldn’t do that
342
00:21:19,970 –> 00:21:22,810
again. Like, it’s like we don’t have the same mindset. But. But to get back
343
00:21:22,810 –> 00:21:25,810
to Liddell and this. The reason I say it this way is because there’s a
344
00:21:25,810 –> 00:21:29,530
disassociation with time. So when women have. If you ever notice, like,
345
00:21:30,090 –> 00:21:33,630
two years later, they didn’t remember the pain the same
346
00:21:33,630 –> 00:21:36,830
way that you, observer, do. Right?
347
00:21:37,150 –> 00:21:40,790
Right. So, like. Right. By the way, guys, I’m totally kidding. Because we do
348
00:21:40,790 –> 00:21:44,550
stupid. We still do stupid stuff all the time. And we continue. You
349
00:21:44,550 –> 00:21:47,349
fall off a ladder, you still climb the ladder. I mean, you know, I’m just
350
00:21:47,349 –> 00:21:50,710
saying, like, as an observer, as an observer watching
351
00:21:50,710 –> 00:21:54,510
childbirth, and you think to yourself, why would anybody ever do this again? But
352
00:21:54,590 –> 00:21:57,870
as a woman watches her child grow up and she gets
353
00:21:57,870 –> 00:22:01,670
disassociated from the time of it, she decides to
354
00:22:01,670 –> 00:22:05,470
have another child. And I think that is Another symptom of
355
00:22:05,470 –> 00:22:09,190
what we’re talking about here. The longer we go from this, from the.
356
00:22:09,190 –> 00:22:12,710
Again, take this, our current political landscape, to
357
00:22:12,710 –> 00:22:16,550
Hitler. You can make all the associations you want. We feel
358
00:22:16,550 –> 00:22:19,310
like it’s not going to happen again because of the checks and balances that we
359
00:22:19,310 –> 00:22:22,990
have in place. But, but if you’re just a simple observer looking
360
00:22:22,990 –> 00:22:26,590
in and you’re looking at this going, holy crap, it’s happening again
361
00:22:26,590 –> 00:22:30,380
because these people have had so much time in between that they didn’t realize
362
00:22:30,380 –> 00:22:33,580
it. And by the way, go backwards in time. And you can say the same
363
00:22:33,580 –> 00:22:37,300
thing about people like Napoleon, about Hannibal, about, like, just
364
00:22:37,300 –> 00:22:40,900
keep going. Like you. There’s. There’s plenty of instances where that
365
00:22:41,140 –> 00:22:44,980
singular person is. Is that dynamic shift in the
366
00:22:44,980 –> 00:22:48,780
power of. And the balance of power. So it’s
367
00:22:48,780 –> 00:22:52,620
happened several times. I, Again, I don’t think. I think
368
00:22:52,620 –> 00:22:55,380
we have. I think we figured out some checks and balances. But I could see
369
00:22:55,380 –> 00:22:59,030
this happening again. Look at the current situation in Russia.
370
00:22:59,030 –> 00:23:02,750
Putin. Oh, yeah, right. Like, he, he’s proving that this could
371
00:23:02,750 –> 00:23:06,350
potentially happen again in Russia for sure. Like that. Like, I don’t know if the
372
00:23:06,350 –> 00:23:10,150
Russian government has checks and balances to, to make sure that he doesn’t do that,
373
00:23:10,150 –> 00:23:13,630
but I don’t think they do. They don’t. He’s been. He’s been running things
374
00:23:13,790 –> 00:23:17,470
pretty well. Well, well, you know, pretty consistently, I would say. Well, pretty
375
00:23:17,470 –> 00:23:20,910
consistently. So I’m not claiming to know their political. But I’m just saying, like,
376
00:23:21,150 –> 00:23:24,390
he’s working. I think it’s fairly consistently for the last 30 years there. As an
377
00:23:24,390 –> 00:23:28,060
observer of history, I see this happening again.
378
00:23:28,060 –> 00:23:31,060
That’s, that’s, that’s the point. But it’s the observer. Yeah, but I think. I think.
379
00:23:31,060 –> 00:23:34,460
So between those three things, I think if Liddell had access
380
00:23:34,620 –> 00:23:38,300
to really deep research in those psychological profiles,
381
00:23:38,380 –> 00:23:41,980
I think his book would have been even more impactful. I, I’m not
382
00:23:41,980 –> 00:23:44,700
suggesting it’s not impactful. I. And I, I think it’s really.
383
00:23:45,740 –> 00:23:49,380
I think it’s a really good book. But. And of course, it answers a
384
00:23:49,380 –> 00:23:52,960
question. I think it answers the question pretty well for its time. Yeah.
385
00:23:52,960 –> 00:23:56,760
But think about a guy like Liddell having access to the current psychology
386
00:23:56,840 –> 00:24:00,000
research that we have and how much more impactful he could have been with that
387
00:24:00,000 –> 00:24:02,440
book. So a couple things there.
388
00:24:05,080 –> 00:24:08,600
So we use
389
00:24:08,760 –> 00:24:12,560
history to. And Lidell Hart talks about
390
00:24:12,560 –> 00:24:15,960
this in his book, too. A variation of this. And again, to your point, pre.
391
00:24:16,280 –> 00:24:19,960
Not pre. But the depth of psychological research. When he wrote this, and he wrote
392
00:24:19,960 –> 00:24:23,520
it in, I believe it was the. Yeah. Published in
393
00:24:23,520 –> 00:24:27,180
1944. Ye. Yeah, exactly. So, you
394
00:24:27,180 –> 00:24:30,460
know, the, the degree to which, to your point, the degree to which psychological research
395
00:24:30,460 –> 00:24:33,940
has come along since 1944 is. Is leaps and
396
00:24:33,940 –> 00:24:36,500
bounds ahead of what he. He had in his time.
397
00:24:38,260 –> 00:24:41,900
But even then he understood something about human nature, which gets
398
00:24:41,900 –> 00:24:45,420
to a couple of different things that we’ve talked about on this show. So we
399
00:24:45,420 –> 00:24:49,220
talked about it in our extra episode where we. Where we discussed the
400
00:24:49,220 –> 00:24:52,020
movie Oppenheimer, which interestingly enough, I watched again last night,
401
00:24:53,820 –> 00:24:57,660
kind of just weirdly lined up with this. And then.
402
00:24:58,300 –> 00:25:02,140
And then we also talked about this in the Orwell episode, not only
403
00:25:02,220 –> 00:25:05,580
on his. In his essay on literature in the English language,
404
00:25:06,140 –> 00:25:09,740
but also in his books, you know, animal farm in 1984.
405
00:25:09,740 –> 00:25:13,580
Right. There’s a through line here. Right.
406
00:25:14,060 –> 00:25:17,660
And even books that we’ve covered, a couple books we cover
407
00:25:18,460 –> 00:25:22,230
from Theodore Roosevelt. We covered the
408
00:25:22,230 –> 00:25:24,750
book that he wrote way back when he was a
409
00:25:26,430 –> 00:25:29,950
representative, I think, or a senator in New York State.
410
00:25:30,110 –> 00:25:33,070
He wrote a book about power politics. We talked about that with Libby Younger.
411
00:25:34,910 –> 00:25:37,990
And the through line that Tom is getting to, and I do agree with it,
412
00:25:37,990 –> 00:25:41,830
is this. We have to figure out, and
413
00:25:41,830 –> 00:25:45,430
this is part of the radical experiment of America, but we have to
414
00:25:45,430 –> 00:25:48,800
figure out how to negotiate
415
00:25:48,800 –> 00:25:51,880
reality with each other. So to get back to two plus two equals five
416
00:25:54,360 –> 00:25:58,040
and this, this, by the way, can start happening in the last 10 years here,
417
00:25:58,040 –> 00:26:00,840
because I’m going to go ahead and step on this third rail, because why not?
418
00:26:07,000 –> 00:26:10,360
If you want to say that two plus two equals five in a dynamic
419
00:26:12,210 –> 00:26:15,970
environment where you’re not abrogating my speech, you’re just providing me
420
00:26:15,970 –> 00:26:19,730
consequences for that speech, which is a whole other kind of discussion.
421
00:26:20,770 –> 00:26:23,930
If you want to go ahead and say two plus two equals five, knock yourself
422
00:26:23,930 –> 00:26:26,930
right the hell out. Sure, go ahead.
423
00:26:28,210 –> 00:26:31,570
But I’m going to. I’m going to channel Ben Shapiro here
424
00:26:31,970 –> 00:26:35,010
and I’m going to say facts don’t care about your feelings.
425
00:26:36,690 –> 00:26:40,530
So, like, you could say two plus two equals five all day, but
426
00:26:40,530 –> 00:26:44,330
I have two things. Then I put together two more
427
00:26:44,330 –> 00:26:47,850
things and invariably I’m going to have four things.
428
00:26:48,170 –> 00:26:52,010
Sorry. Like, this is just. This is just reality. This is the
429
00:26:52,010 –> 00:26:55,809
ceiling of, like, your logic. Right. You could feel any way you
430
00:26:55,809 –> 00:26:58,570
want about it. You could feel that it’s. And here we go. I’m going to
431
00:26:58,570 –> 00:27:02,170
step on the. Step on the rail. You could feel that it’s white supremacist.
432
00:27:02,410 –> 00:27:06,180
You could feel that it’s a sign of the patriarchy. You can feel
433
00:27:06,180 –> 00:27:09,900
that it’s oppressive. You can feel that it’s a sign of
434
00:27:09,900 –> 00:27:13,620
systemic oppression. You could feel that it’s a sign of
435
00:27:13,620 –> 00:27:16,420
white fragility. You could feel all of these things.
436
00:27:19,540 –> 00:27:23,340
And I hold up two fingers, and then I
437
00:27:23,340 –> 00:27:26,740
hold up two more fingers and I still only have four
438
00:27:26,740 –> 00:27:30,460
fingers. That’s
439
00:27:30,460 –> 00:27:33,460
it. That’s. That’s it. That’s it. That’s the whole thing. There’s no, there’s no more
440
00:27:33,460 –> 00:27:36,850
argument that I need to. Right. And so this, this has started happening, this
441
00:27:36,850 –> 00:27:40,250
renegotiation of reality that you’re talking about, which is
442
00:27:40,490 –> 00:27:44,130
fine with history. I don’t have a problem with renegotiating
443
00:27:44,130 –> 00:27:47,530
reality with history. And, and Liddell Hart talks a little bit about this in his
444
00:27:47,530 –> 00:27:51,290
book. You know, when myths become stronger than the history, the myth becomes
445
00:27:51,290 –> 00:27:55,050
the history. He, he mentions this, right. And it’s fine. I
446
00:27:55,050 –> 00:27:58,810
have no problem taking a scientific approach. He does. He has a
447
00:27:58,810 –> 00:28:01,770
problem with taking a scientific approach to history. He talks about it in the first
448
00:28:01,770 –> 00:28:05,050
part of his book because he thinks that it drains and denud
449
00:28:06,630 –> 00:28:09,790
of all of its emotive value. And he doesn’t think that you should be able
450
00:28:09,790 –> 00:28:13,550
to do that. He thinks you should balance the emotive value with the scientific
451
00:28:13,550 –> 00:28:17,030
pursuit of the truth. I. Okay, I can see his
452
00:28:17,030 –> 00:28:20,830
argument, But I also think that human beings are going
453
00:28:20,830 –> 00:28:23,470
to look for that emotive value and they’re going to look forward in myths. You
454
00:28:23,470 –> 00:28:25,710
can’t take that out of the human because again, that’s one of those, like two
455
00:28:25,710 –> 00:28:29,270
plus two equals four things. It’s one of the things that’s built into human understanding
456
00:28:29,270 –> 00:28:32,970
of reality. But if we want to renegotiate history, sure,
457
00:28:32,970 –> 00:28:36,250
let’s renegotiate history. Let’s go ahead and dig into the
458
00:28:36,250 –> 00:28:39,010
Tuskegee Project. Let’s go ahead and dig into
459
00:28:40,370 –> 00:28:43,970
Black Wall street in Oklahoma. Let’s go ahead and dig
460
00:28:43,970 –> 00:28:47,730
into the decimation of the Native American
461
00:28:47,730 –> 00:28:51,410
tribes in, in, in North America
462
00:28:51,570 –> 00:28:55,250
and in Mexico and in South America. Let’s go ahead and dig into that
463
00:28:55,810 –> 00:28:58,370
and understand that when you do that,
464
00:29:00,140 –> 00:29:03,780
and this is something that we also forget with history, I think, and
465
00:29:03,780 –> 00:29:07,540
it’s notorious in warfare. Understand that when you do
466
00:29:07,540 –> 00:29:11,100
that, the enemy you are fighting or you are opposing
467
00:29:11,820 –> 00:29:14,380
also gets a vote in a free speech society.
468
00:29:16,300 –> 00:29:19,900
So when you start pulling apart people’s myths, they’re going
469
00:29:19,900 –> 00:29:23,540
to have a reaction to that and a response to
470
00:29:23,540 –> 00:29:27,270
that. Now, we would like to keep that response and reaction in the space of
471
00:29:27,270 –> 00:29:30,670
speech. So there’s going to be an argument, there’s going to be a
472
00:29:30,670 –> 00:29:33,110
verbal conflict, there’s going to be a fight.
473
00:29:34,710 –> 00:29:37,950
And I think in the structure that we have in America, that’s fine. We should
474
00:29:37,950 –> 00:29:40,950
be doing that. Now you have to talk about the observer from the outside. If
475
00:29:40,950 –> 00:29:43,750
I’m an observer from the outside looking at this, all this looks like is just
476
00:29:43,750 –> 00:29:47,510
massive chaos, just endless, never ending massive chaos
477
00:29:47,830 –> 00:29:51,390
about things that should just be decided by an aristocratic
478
00:29:51,390 –> 00:29:55,210
cadre of individuals. Everybody gets told what the thing is
479
00:29:55,370 –> 00:29:57,970
and then you just move on. By the way, this is what they do in
480
00:29:57,970 –> 00:30:01,290
communist China. They have myths based on
481
00:30:01,290 –> 00:30:04,970
Confucianism and the Chinese communist government just says, we’ve decided
482
00:30:04,970 –> 00:30:06,970
this, you’re welcome.
483
00:30:09,370 –> 00:30:12,970
Everybody sort of goes, yeah, that’s fine. By the way, in those kinds of systems,
484
00:30:13,450 –> 00:30:17,170
to Orwell’s point, if you’re in a to totalizing
485
00:30:17,170 –> 00:30:20,890
system like that, where, where dissension from something is
486
00:30:20,890 –> 00:30:24,540
not allowed, this is the point from 1984. Yeah, the two
487
00:30:24,540 –> 00:30:28,180
plus two does equal five, sure, in
488
00:30:28,180 –> 00:30:32,020
that system. But guess what? That
489
00:30:32,020 –> 00:30:35,540
system has to negotiate with the reality of other systems in a global environment.
490
00:30:35,940 –> 00:30:39,380
Which is part of the reason why the Internet’s like wrecking havoc
491
00:30:39,380 –> 00:30:43,220
everywhere right now and has for the last 30, almost 40 years
492
00:30:43,220 –> 00:30:47,060
and actually more than 40 years now and will continue to wreak havoc, by the
493
00:30:47,060 –> 00:30:49,820
way, for the next hundred years because we’re not going to be able to get
494
00:30:49,820 –> 00:30:53,370
over this. And then the other thing that you’re talking about in there, so there’s
495
00:30:53,370 –> 00:30:56,690
negotiating reality, then there’s the other thing you’re talking about. There is democracy.
496
00:30:56,930 –> 00:31:00,450
And fundamentally, as an aristocrat,
497
00:31:01,250 –> 00:31:04,970
B. H Liddell Hart was opposed to democracy and he’s joined
498
00:31:04,970 –> 00:31:08,650
in good as a fellow traveler, or he walks as a fellow
499
00:31:08,650 –> 00:31:12,290
traveler alongside Thomas Jefferson, who was opposed to democracy,
500
00:31:12,770 –> 00:31:16,570
and George Washington, who was opposed to democracy, and Ben Franklin who was
501
00:31:16,570 –> 00:31:19,970
opposed to democracy, and John Adams who was opposed to democracy.
502
00:31:20,290 –> 00:31:24,030
These guys wanted a. And they built a republic
503
00:31:24,830 –> 00:31:28,590
because democracies are two wolves and a
504
00:31:28,590 –> 00:31:31,390
sheep voting on who gets to be dinner.
505
00:31:32,590 –> 00:31:36,430
That’s a democracy. Congratulations. 66% of the people,
506
00:31:36,590 –> 00:31:40,030
66% of the entities in that group just voted on dinner
507
00:31:40,750 –> 00:31:44,150
and your vote didn’t count. That’s democracy. Democracy is
508
00:31:44,150 –> 00:31:47,990
inevitably, inevitably degrades to mob rule. We saw this, we see this
509
00:31:47,990 –> 00:31:51,600
throughout history, literally. So you build a republic,
510
00:31:52,000 –> 00:31:55,680
and in a republic, the two wolves,
511
00:31:56,880 –> 00:32:00,560
the two wolves have to send one wolf as a representative and the sheep
512
00:32:00,560 –> 00:32:03,520
gets to send two sheeps as a representative. And now we’re going to have a
513
00:32:03,520 –> 00:32:06,320
Fight. And that’s how we keep balance in the system.
514
00:32:07,200 –> 00:32:10,880
And so we’ve got all these dynamics which impact how we view
515
00:32:10,880 –> 00:32:14,600
history in America. But to your point, the through line is
516
00:32:14,600 –> 00:32:18,290
about negotiating reality and who gets to negotiate that and who gets to
517
00:32:18,290 –> 00:32:21,730
say what the other thing too. And through your little
518
00:32:21,730 –> 00:32:25,250
Spiel. Spiel there. I wanted to, I just wanted to jump in and say one
519
00:32:25,250 –> 00:32:28,850
thing before, but I obviously wanted to let you finish. But, but when
520
00:32:28,850 –> 00:32:32,489
you, when you’re referencing the two plus two, and I still have four
521
00:32:32,490 –> 00:32:36,090
and you know, fact. Listen, facts are facts. Right? Like, to your point,
522
00:32:36,410 –> 00:32:39,130
the problem with history is
523
00:32:40,090 –> 00:32:43,650
it’s almost never. Yes, the facts, ma’. Am. Right.
524
00:32:43,650 –> 00:32:47,410
Like, there has to. There’s always human emotion in it. There’s always
525
00:32:47,490 –> 00:32:51,250
like these degrees of champions and like,
526
00:32:51,250 –> 00:32:54,610
like there’s always these. There’s heroism and like we’re
527
00:32:54,610 –> 00:32:58,450
elevating things to, to make it feel more real or to more. Have
528
00:32:58,450 –> 00:33:02,170
more impact. And it’s never like, if you were to ever
529
00:33:02,170 –> 00:33:05,970
read a history book that was literally just factual information, nobody would read the
530
00:33:05,970 –> 00:33:09,330
whole. Nobody would read it. No, like, no, it’s like, wait,
531
00:33:09,570 –> 00:33:13,330
okay, so World War II started in this date, ended in this date. The
532
00:33:13,330 –> 00:33:17,010
battle of this date was this date, this day, the battle. Like, like. But
533
00:33:17,650 –> 00:33:21,370
because we tie everything into some sort of
534
00:33:21,370 –> 00:33:25,210
emotion, we have to feel an emotion for anything that we’re
535
00:33:25,210 –> 00:33:28,850
involved in, whether it’s reading history, creating history,
536
00:33:29,250 –> 00:33:32,924
looking at the future, etc. Etc. So to your point about the 2
537
00:33:32,976 –> 00:33:36,540
+2 is always going to equal 4. Yes, because there’s no
538
00:33:36,540 –> 00:33:40,380
emotion in that that you cannot, you can’t invoke emotion into
539
00:33:40,460 –> 00:33:41,980
math. Right.
540
00:33:44,380 –> 00:33:48,220
Like, I remember, I read recently, I read
541
00:33:48,220 –> 00:33:51,900
recently, somebody, somebody was arguing some political.
542
00:33:52,060 –> 00:33:55,780
Blah, blah, blah, whatever. And the person, literally, the, the argument, what
543
00:33:55,780 –> 00:33:58,620
they were talking about was like, listen, you cannot
544
00:33:58,780 –> 00:34:02,590
distort math. And they literally said
545
00:34:02,590 –> 00:34:06,310
exactly what you said. 2/2 is always going to equal 4. So if you’re looking
546
00:34:06,310 –> 00:34:09,510
at statistical data that is verified
547
00:34:09,990 –> 00:34:12,870
statistical data, I’m not talking about stuff that we just make up.
548
00:34:13,750 –> 00:34:16,990
First time I ever heard this line, it made me laugh so hard. It’s like,
549
00:34:16,990 –> 00:34:20,310
hey, Tom, did you know that 86 of all statistics are made up? And I
550
00:34:20,310 –> 00:34:23,270
went, there you go, what? Like, I had to stop. And I was like,
551
00:34:23,990 –> 00:34:26,550
that’s made up. Exactly. Right. So, like,
552
00:34:28,630 –> 00:34:32,370
so but if you. But if it’s verified statistical data and
553
00:34:32,370 –> 00:34:35,890
you can say that X number of people voted this way,
554
00:34:35,890 –> 00:34:39,370
X number of people voted that way. So 28 of people
555
00:34:39,370 –> 00:34:43,050
voted this Way based on the information they were given that is statistical.
556
00:34:43,050 –> 00:34:46,690
You cannot change that. That. It just is what it is,
557
00:34:46,770 –> 00:34:50,010
right? They were trying to make that argument and the guy was still fighting, like,
558
00:34:50,010 –> 00:34:53,810
pushing back, going, no, you can, you can, you can manipulate statistics.
559
00:34:54,130 –> 00:34:57,719
Well, yes, but then they’ve not. Then they’re not validated. They’re not
560
00:34:57,719 –> 00:35:01,359
verifiable statistics at that point. They’re 86% of
561
00:35:01,359 –> 00:35:04,879
statistics are made up. Like, you have to be able to validate it. If you
562
00:35:04,879 –> 00:35:08,479
can’t validate it, then it’s not, it’s not reasonable to see. So, so
563
00:35:08,479 –> 00:35:10,759
my, my point to all that is that,
564
00:35:11,879 –> 00:35:15,319
yes, mathematical information, things that you can
565
00:35:15,319 –> 00:35:19,039
literally see, two plus two equals four. I, if anybody
566
00:35:19,039 –> 00:35:22,479
tries to argue that, I just think they’re being silly, and all of us are
567
00:35:22,479 –> 00:35:26,110
going to think they’re being silly. But when they start debating on the
568
00:35:26,110 –> 00:35:29,870
merits of historical information that is written by people, that are
569
00:35:29,870 –> 00:35:32,910
written by the winners, so to speak, that are written by the
570
00:35:33,310 –> 00:35:37,030
majority, all of that stuff when
571
00:35:37,030 –> 00:35:40,790
you read that, you feel something when you read that. That’s
572
00:35:40,790 –> 00:35:44,230
another bias that we probably don’t talk about is the emotional
573
00:35:44,230 –> 00:35:47,710
bias that you get when you read some of those things. So therefore,
574
00:35:48,110 –> 00:35:51,830
if you see that the United States and its allies are going to
575
00:35:51,830 –> 00:35:55,390
win World War II based on what you’re reading, and you say, okay, so
576
00:35:55,630 –> 00:35:59,350
why, why do we have the. Why are we trying to prevent World War
577
00:35:59,350 –> 00:36:02,550
iii? We know the new the. The allies are going to win. That’s what happens,
578
00:36:02,550 –> 00:36:05,990
right? So we look throughout history, but, like, can we stop for a second? And
579
00:36:05,990 –> 00:36:08,790
let’s, like, not lose the millions of lives that we’re going to lose because of
580
00:36:08,790 –> 00:36:12,550
it. Like, is there a way to circumvent? Right. Just go from point A to
581
00:36:12,550 –> 00:36:15,470
point C instead of, like, can we skip point B,
582
00:36:16,750 –> 00:36:20,570
skip point D. Yeah, don’t learn. That’s what we don’t learn. Right. Well,
583
00:36:20,570 –> 00:36:23,970
and this is, this is. I mean, again, I watched Oppenheimer last night, right?
584
00:36:24,290 –> 00:36:28,010
And you know, it’s interesting. Like, there’s a movie coming out, and we always
585
00:36:28,010 –> 00:36:31,410
end up talking about movies. This is the moment to do this. Right? Now there’s
586
00:36:31,410 –> 00:36:35,090
a movie coming out from Catherine Bigelow, who directed A Hurt Locker
587
00:36:35,249 –> 00:36:37,810
and a couple other movies on Netflix
588
00:36:38,770 –> 00:36:42,490
about, like, Five Minutes to Nuclear Impact or
589
00:36:42,490 –> 00:36:45,570
something like that. Anyway, it reminds. It’s on. It’s going on Netflix. It’s coming out
590
00:36:45,570 –> 00:36:47,930
in December. I think it’s going to be in the theaters for, like, a brief
591
00:36:47,930 –> 00:36:50,250
minute, and then it’s going to. It’s called House of Fire, I think, or House
592
00:36:50,250 –> 00:36:54,030
of Dynamite or I can’t remember. Anyway, it doesn’t matter what the name
593
00:36:54,030 –> 00:36:57,550
of the movie is. It’s the latest entry into a
594
00:36:57,550 –> 00:37:01,070
genre of films that started, gosh,
595
00:37:01,150 –> 00:37:04,910
back in the 1950s with the movie directed by Stanley Kubrick,
596
00:37:04,910 –> 00:37:08,590
Dr. Strangelove, or how I Learned to Stop Worrying and Love the Bomb,
597
00:37:09,390 –> 00:37:13,230
where artists who are
598
00:37:13,230 –> 00:37:16,430
attempting to, to your point about emotion, who are attempting to
599
00:37:16,510 –> 00:37:19,480
emote into history, because that’s what artists do do
600
00:37:21,640 –> 00:37:25,080
and are attempting to get the viewer or the observer to engage
601
00:37:25,080 –> 00:37:28,040
emotionally with something, an idea that they have.
602
00:37:29,640 –> 00:37:33,280
There’s a long genre of films that began there and
603
00:37:33,280 –> 00:37:37,080
continues on through the Katherine Bigalow film, of which Oppenheimer is one of these films
604
00:37:39,000 –> 00:37:42,360
where to your point about World War 3,
605
00:37:42,840 –> 00:37:46,590
it’s that constant drum beat of warning that nuclear
606
00:37:46,590 –> 00:37:49,790
thermonuclear destruction is the worst thing that can possibly ever happen.
607
00:37:50,190 –> 00:37:53,390
It will destroy all of humanity, all civilization, everywhere,
608
00:37:53,950 –> 00:37:57,310
period, full stop. And it is, it is a
609
00:37:57,470 –> 00:38:00,750
drumbeat that runs through history.
610
00:38:01,710 –> 00:38:04,750
And to your point about forgetting, and this is sort of the last thing I
611
00:38:04,750 –> 00:38:06,710
want to say, and then we can go back to the book. But your point
612
00:38:06,710 –> 00:38:10,350
about forgetting, we’ve forgotten how bad World War
613
00:38:10,350 –> 00:38:13,230
II really was because we are distanced from it.
614
00:38:14,160 –> 00:38:18,000
And my concern also is that we have forgotten how
615
00:38:18,000 –> 00:38:20,800
bad a nuclear bomb can be.
616
00:38:22,240 –> 00:38:26,080
Thus you get crazy talk like in the last
617
00:38:26,080 –> 00:38:29,760
administration from certain people about
618
00:38:30,000 –> 00:38:33,840
arming, arming
619
00:38:33,840 –> 00:38:37,520
bases in Europe with nuclear tipped
620
00:38:37,920 –> 00:38:41,640
whatever, and if Russia does this thing
621
00:38:41,640 –> 00:38:45,080
or that thing in the Ukraine, we’re going to use those
622
00:38:45,080 –> 00:38:48,880
nukes. People in the last administration running around saying, saying
623
00:38:49,120 –> 00:38:52,080
nonsense like this. And of course the Russians are responding with,
624
00:38:52,960 –> 00:38:56,800
listen, we lost 60 some odd million people in
625
00:38:56,800 –> 00:39:00,480
what we call the Great Patriotic War and we didn’t miss
626
00:39:00,480 –> 00:39:03,200
a beat. You really want to go ahead, you really want to go ahead and
627
00:39:03,200 –> 00:39:06,880
pull that smoke wagon, you go right on ahead. We got nukes
628
00:39:06,880 –> 00:39:10,680
too. Yeah, exactly. And this is, this is the hubris to your point. This
629
00:39:10,680 –> 00:39:14,240
is reflect, even, just even allowing that to be said, number one,
630
00:39:14,240 –> 00:39:18,080
moves the Overton window on talking about nuclear war. But it also,
631
00:39:19,280 –> 00:39:22,720
it also reflects a level of hubris and arrogance
632
00:39:22,960 –> 00:39:26,360
that can only come from being disassociated in time and
633
00:39:26,360 –> 00:39:30,080
forgetting the lessons of history. Right. Yeah. And by the
634
00:39:30,080 –> 00:39:33,390
way, but just to circle this back into the whole purpose of this
635
00:39:33,550 –> 00:39:36,430
podcast in general, this
636
00:39:37,070 –> 00:39:40,830
nothing that we’re talking about excludes businesses, by the way. No,
637
00:39:41,150 –> 00:39:44,270
we, we’ve seen the same. Think again,
638
00:39:44,910 –> 00:39:48,710
current Environment not excluded. If you think
639
00:39:48,710 –> 00:39:52,110
about, like, I was actually, I literally just read something
640
00:39:52,270 –> 00:39:55,790
earlier, earlier today about this, which I found fascinating because
641
00:39:56,190 –> 00:39:59,790
those of us in the sales and marketing world have been talking recently about
642
00:40:00,390 –> 00:40:04,030
AI kind of plateauing a little bit. Like there’s, there’s been, like
643
00:40:04,030 –> 00:40:07,630
there was this major rush and there was all this talk about
644
00:40:07,630 –> 00:40:10,950
AI replacing people and taking jobs and all this other stuff,
645
00:40:11,270 –> 00:40:15,070
and yet I just saw a statistic, statistic. I just saw
646
00:40:15,070 –> 00:40:18,510
a statistic this morning that said of the
647
00:40:18,510 –> 00:40:22,030
Fortune 500 companies, the average company is only
648
00:40:22,030 –> 00:40:25,030
deploying AI at about a 2% efficiency rate.
649
00:40:26,000 –> 00:40:29,360
How many people you think they’re replacing at IBM because of this?
650
00:40:29,680 –> 00:40:33,240
How many people do you think at Amazon do you think they’re replacing because of
651
00:40:33,240 –> 00:40:36,520
it? It’s just not happening the way we expected it. Now here’s. Let me back
652
00:40:36,520 –> 00:40:39,840
up to my point a second ago. We saw this once before,
653
00:40:40,160 –> 00:40:43,840
folks, the dot com in the late 90s where
654
00:40:44,400 –> 00:40:48,080
investors were just throwing money upon money upon money at
655
00:40:48,480 –> 00:40:51,680
anything that had dot com at the end of it. And some of them lost
656
00:40:51,680 –> 00:40:54,830
their shirts. Some of them did okay. Some of them did all right. Yeah. The
657
00:40:54,830 –> 00:40:58,630
dot com boom didn’t. It didn’t. Just
658
00:40:58,630 –> 00:41:01,750
because there were a few winners did not mean that a,
659
00:41:02,550 –> 00:41:06,190
an intense amount of businesses failed. Yeah. Oh,
660
00:41:06,190 –> 00:41:09,910
yeah. So. So what did that teach us? Apparently
661
00:41:09,910 –> 00:41:13,630
nothing. Nothing. Because, because, because investors are
662
00:41:13,630 –> 00:41:17,030
starting to throw so much money at AI right now, I guarantee you
663
00:41:18,000 –> 00:41:20,880
a majority of them are going to lose their shirts and there are going to
664
00:41:20,880 –> 00:41:24,640
be a few, a select few that come out on top. And for some reason,
665
00:41:24,720 –> 00:41:28,360
we’re going to remember history as the dot com boom and the AI boom as
666
00:41:28,360 –> 00:41:32,120
being successful. How that happens to me is
667
00:41:32,120 –> 00:41:35,920
beyond. Like, again, business does the same stuff.
668
00:41:36,080 –> 00:41:39,800
Like, I feel, I, I’ve been telling people lately, I feel like
669
00:41:39,800 –> 00:41:43,280
I’m a veteran who’s been through like the fourth World War. Like, I feel like
670
00:41:43,280 –> 00:41:45,400
I’ve been, you know, I feel like I’ve been through four. I was going through
671
00:41:45,400 –> 00:41:48,400
the search revolution. Well, no, the first one was the Internet revolution to your point,
672
00:41:48,400 –> 00:41:51,500
with the dot com bubble. Oh, go back, go back one step further. The email.
673
00:41:51,580 –> 00:41:54,700
When email. Oh, God. Oh, yeah, email. Oh my God.
674
00:41:55,260 –> 00:41:58,020
Email’s gonna change the world. And we’re never gonna, we’re not gonna need a postal
675
00:41:58,020 –> 00:42:01,740
service anymore, right, Exactly. We have Amazon that delivers packages literally
676
00:42:01,740 –> 00:42:05,580
daily, like a day to people’s. Anyway, go ahead. Sorry. Yeah,
677
00:42:05,580 –> 00:42:08,860
yeah, sure. Amazon knows exactly where I live
678
00:42:08,940 –> 00:42:12,740
anyway. They can find me. They’ll have no Problem
679
00:42:12,740 –> 00:42:16,100
finding me, by the way. So does ups. Amazon knows where I live,
680
00:42:16,100 –> 00:42:19,780
UPS knows where I live, the postal service, everybody knows where I live now. Like,
681
00:42:19,780 –> 00:42:23,580
I can’t even get away from it at this point anyway. Good. But yeah, you
682
00:42:23,580 –> 00:42:26,540
got email. Take all that away. Email, Internet.
683
00:42:27,100 –> 00:42:30,780
Oh, God. Social media was going to revolutionize how we were going. Remember
684
00:42:30,780 –> 00:42:34,460
that. Like, we were all just going to live in these, like, these like Facebook
685
00:42:34,540 –> 00:42:37,980
pods and the Instagram Pod and the LinkedIn Pod, and we were never going to.
686
00:42:38,220 –> 00:42:41,900
And then, and then VR ar, which was a brief.
687
00:42:42,140 –> 00:42:45,930
Before that, it was, it was a 3D printing. Oh, yeah. Oh, I haven’t even
688
00:42:45,930 –> 00:42:49,570
gotten to that one yet. 3D printing. Crypto. Oh, my God.
689
00:42:49,570 –> 00:42:53,410
Yeah, crypto. And, and now, now here
690
00:42:53,410 –> 00:42:56,490
we are with AI. And I got to admit
691
00:42:57,290 –> 00:43:00,850
to your point, Tom, I get a little
692
00:43:00,850 –> 00:43:04,570
grizzled and gray when I, when I hear
693
00:43:04,570 –> 00:43:08,250
about it because I’m like. On the one hand, it’s very exciting
694
00:43:08,250 –> 00:43:11,850
because it’s a, it’s, it is a gold rush, exciting kind of, kind of thing.
695
00:43:11,850 –> 00:43:15,310
You can, you could fall into the emotion of it. Right. Because it gets to
696
00:43:15,310 –> 00:43:18,910
be emotive. It’s the bright shiny object thing. It’s the bright shiny object thing.
697
00:43:18,910 –> 00:43:22,110
Exactly. You just talked about. That’s exactly what they were. Now
698
00:43:22,670 –> 00:43:26,190
are some of them. Yeah, I’m not suggesting that they quote, unquote, failed.
699
00:43:26,350 –> 00:43:30,110
We won’t know that for 15 years, though. We won’t know for 15 years. Business
700
00:43:30,110 –> 00:43:33,830
lessons. And then, and then being more calculated about how we project, how
701
00:43:33,830 –> 00:43:37,510
we go forward, business. We learned nothing from all of those things you just said.
702
00:43:37,510 –> 00:43:41,070
The email, the, the dot com boom, the search engine, the,
703
00:43:41,230 –> 00:43:44,830
the, the, the 3D printing, the crypto, the
704
00:43:44,990 –> 00:43:48,750
social media. At all of those. There were
705
00:43:49,310 –> 00:43:52,910
thousands of companies involved in those things, most of which failed.
706
00:43:53,150 –> 00:43:56,989
Right. Like, but there will.
707
00:43:56,990 –> 00:44:00,350
And there will be an AI apocalypse. There will. Yeah, there will be an LLM
708
00:44:00,350 –> 00:44:04,070
apocalypse everywhere. Like, I was reading something of the other day in one of the
709
00:44:04,070 –> 00:44:07,830
startup. One of the startup newsletters that I read for the other project that you
710
00:44:07,830 –> 00:44:09,790
and I are on. And,
711
00:44:11,550 –> 00:44:15,190
and something like all of
712
00:44:15,190 –> 00:44:17,910
the. Oh yeah, I know what it was. All of the investor money, or the
713
00:44:17,910 –> 00:44:21,510
vast majority of investor money in Silicon Valley. If you don’t have something
714
00:44:21,510 –> 00:44:25,150
AI, you can’t get, you can’t even get into the door now.
715
00:44:25,470 –> 00:44:28,830
And I think about the event that we ran this week
716
00:44:29,070 –> 00:44:32,670
with the folks that we ran that with on that other
717
00:44:32,670 –> 00:44:36,080
project. And, and I mean,
718
00:44:36,800 –> 00:44:39,680
there’s way more things happening in the world. I Think we only had like what,
719
00:44:39,680 –> 00:44:43,400
two, maybe three out of the, out of the companies, out of the ones that
720
00:44:43,400 –> 00:44:47,160
we looked at that were. And I’m being on purposely oblique
721
00:44:47,160 –> 00:44:50,400
about this folks, but like three companies that we looked at
722
00:44:50,720 –> 00:44:54,240
that even had an AI play. The vast majority of everybody else
723
00:44:54,560 –> 00:44:58,160
is still trying to do a business the way you do a
724
00:44:58,240 –> 00:45:01,860
business. Now is there going to be an AI play built into that?
725
00:45:01,860 –> 00:45:05,460
Yeah, maybe. Probably because you got to get investors attention. But
726
00:45:05,540 –> 00:45:08,780
that’s insane to me because there’s just so many other businesses that are, that could
727
00:45:08,780 –> 00:45:12,460
operate in the world without, without an LLM. Yeah. And there’s so many other
728
00:45:12,460 –> 00:45:16,220
problems that we solve that LLMs can’t solve. So anyway, yeah, no,
729
00:45:16,220 –> 00:45:19,940
we haven’t learned. And there’ll be another bubble in 10 years. There’ll be another
730
00:45:19,940 –> 00:45:23,460
bubble, I guarantee. And it might be, honestly, it might be the humanoid robot bubble
731
00:45:23,460 –> 00:45:27,020
that I think might be the hardware bubble, that might be the hardware version, that
732
00:45:27,020 –> 00:45:30,660
might be the 3D printing version. The next bubble you.
733
00:45:31,040 –> 00:45:34,840
And so Elon, you know why, Jason? Because the more things change, the more.
734
00:45:34,840 –> 00:45:38,560
Things and the more they stick. There we go. There it
735
00:45:38,560 –> 00:45:42,320
is. There it is. Back to the book. Back
736
00:45:42,320 –> 00:45:45,920
to why don’t we learn from History by B.H. liddell Hart.
737
00:45:46,320 –> 00:45:50,040
This is an open source book, by the way. You can get it online for
738
00:45:50,040 –> 00:45:53,840
free. So. But the version that I’m reading has a
739
00:45:53,840 –> 00:45:57,520
yellow cover and it was edited and with an introduction by Gills
740
00:45:57,520 –> 00:46:01,070
Lauren. But you can grab this book anywhere
741
00:46:01,310 –> 00:46:04,430
online. This is, this is definitely an open source book and I would encourage you
742
00:46:04,510 –> 00:46:07,870
if you are in business or you’re in leadership
743
00:46:08,110 –> 00:46:09,790
or you are in tech,
744
00:46:11,790 –> 00:46:15,310
especially if you’re part of one of those LLM
745
00:46:15,310 –> 00:46:18,750
driven AI startups. I strongly recommend
746
00:46:18,990 –> 00:46:22,270
reading this book. All right, back to the book. So
747
00:46:22,670 –> 00:46:26,510
let’s look at the importance of keeping promises and the importance of care about making
748
00:46:26,510 –> 00:46:30,080
promises. And, and I want to talk about, with Tom, about something that he
749
00:46:30,080 –> 00:46:33,280
mentioned that ties into how we teach history.
750
00:46:34,880 –> 00:46:38,240
Back to the book on the importance of keeping promises.
751
00:46:38,480 –> 00:46:41,440
Civilization is built on the practice of keeping promises.
752
00:46:42,080 –> 00:46:45,840
It may not sound a high attainment, but if trust in its observance
753
00:46:45,840 –> 00:46:49,360
should be shaken, the whole structure cracks and sinks.
754
00:46:49,920 –> 00:46:52,960
Any constructive effort in all human relations, personal,
755
00:46:53,570 –> 00:46:57,370
political and commercial depend on being able
756
00:46:57,370 –> 00:47:01,010
to depend on promises. By the way,
757
00:47:01,010 –> 00:47:04,570
pause for just a minute. That’s genius. I’ve never heard
758
00:47:04,570 –> 00:47:08,010
that. I’ve never heard an argument for high trust, a high trust
759
00:47:08,010 –> 00:47:11,850
society laid out as succinctly as is laid out in those
760
00:47:11,850 –> 00:47:15,170
three Sentences. That’s brilliant. That’s brilliant writing.
761
00:47:15,810 –> 00:47:19,210
Back to the book. This truth has a reflection on the question of
762
00:47:19,210 –> 00:47:22,890
collective security among nations and on the lessons of history in regards to that
763
00:47:22,890 –> 00:47:26,640
subject. In the years before the war, the charge was constantly brought. And
764
00:47:26,640 –> 00:47:30,440
by the way, the war he’s talking about, just pause again, is World War II.
765
00:47:30,440 –> 00:47:33,440
But sometimes he’s also talking about World War I. So just you have to think
766
00:47:33,440 –> 00:47:36,880
about those both in concert with each other. All right. In the years before the
767
00:47:36,880 –> 00:47:40,360
war, the charge was constantly brought that its supporters were courting the risk of war
768
00:47:41,400 –> 00:47:45,160
by their exaggerated respect for covenants. Although
769
00:47:45,160 –> 00:47:48,680
they may have been fools in disregarding the conditions necessary for the
770
00:47:48,680 –> 00:47:52,430
effective fulfillment of pledges, they at least show themselves men
771
00:47:52,430 –> 00:47:56,150
of honor. I have that double underlined, by the way, and in the long
772
00:47:56,150 –> 00:47:59,910
view of more fundamental common sense than those who
773
00:47:59,910 –> 00:48:03,750
argued that we should give aggressors a free hand so long as they left us
774
00:48:03,750 –> 00:48:07,590
alone. History has shown repeatedly
775
00:48:07,750 –> 00:48:11,510
that the hope of buying safety in this way is the greatest of
776
00:48:11,510 –> 00:48:15,350
delusions. The importance of care about
777
00:48:15,350 –> 00:48:19,170
making promises. It is immoral to make promises
778
00:48:19,170 –> 00:48:22,610
that one cannot in practice fulfill in the sense that the
779
00:48:22,610 –> 00:48:26,130
recipient expects on the ground in
780
00:48:26,130 –> 00:48:29,850
1939. This is ahead of World War II. I question the
781
00:48:29,850 –> 00:48:33,570
underlying morality of the Polish guarantee as well as its practicality. If
782
00:48:33,570 –> 00:48:37,250
the Poles had realized the military inability of Britain and France to save them from
783
00:48:37,250 –> 00:48:40,690
defeat and of what such a defeat would mean to them individually and
784
00:48:40,690 –> 00:48:44,320
collectively, it is unlikely that they would have shown such a stubborn opposition to
785
00:48:44,320 –> 00:48:48,120
Germany’s originally modest demands for Danzig and a
786
00:48:48,120 –> 00:48:51,680
passage through the Corridor, since it was obvious to me that they were bound to
787
00:48:51,680 –> 00:48:55,320
lose those points and even much more in the event of a conflict.
788
00:48:55,480 –> 00:48:59,320
It seemed to me wrong on our part to make promises that we
789
00:48:59,320 –> 00:49:03,040
that were bound to encourage false hopes. It
790
00:49:03,040 –> 00:49:05,920
also seemed to me that any such promises were the most certain way to produce
791
00:49:05,920 –> 00:49:09,690
war. Because the inevitable provocativeness of guaranteeing at such
792
00:49:09,690 –> 00:49:13,530
a moment of tension an area which we had hitherto treated as outside our
793
00:49:13,530 –> 00:49:17,210
sphere of interest, because of the manifest temptation which the
794
00:49:17,210 –> 00:49:20,850
guarantee offered to a military minded people like the Germans, to show how
795
00:49:20,850 –> 00:49:24,490
fatuously impractical our guarantee was, and because of its
796
00:49:24,490 –> 00:49:27,850
natural effect on stiffening the attitude of a people, the
797
00:49:27,850 –> 00:49:31,450
Poles who had always shown themselves exceptionally intractable in
798
00:49:31,450 –> 00:49:34,450
negotiating a reasonable settlement of any issue
799
00:49:38,220 –> 00:49:42,020
and historian could not help seeing certain parallels between the long standing aspect of
800
00:49:42,020 –> 00:49:45,780
the Polish German situation and that between Britain and the Boer Republics 40
801
00:49:45,780 –> 00:49:49,460
years earlier, and remembering the effects on us of the attempts of the other European
802
00:49:49,460 –> 00:49:52,860
powers to induce or coerce us into negotiating a settlement with the Boers.
803
00:49:53,340 –> 00:49:57,020
If our own reaction then had been so violent, it could hardly be expected that
804
00:49:57,020 –> 00:50:00,460
the reaction of a nation filled with an even more bellicose spirit would be less
805
00:50:00,460 –> 00:50:04,200
violent, especially as the attempt to compel negotiation was backed by
806
00:50:04,200 –> 00:50:08,000
an actual promise of making war if Poland felt moved to
807
00:50:08,000 –> 00:50:11,040
resist the German conditions.
808
00:50:15,120 –> 00:50:18,600
That is a brilliant piece of analysis. That’s why I’m reading this. That is a
809
00:50:18,600 –> 00:50:22,080
brilliant piece of analysis of the psychology
810
00:50:22,160 –> 00:50:26,000
of nation states, which we don’t often talk about. Like, we try
811
00:50:26,000 –> 00:50:29,790
to pretend that nations are these, and
812
00:50:29,790 –> 00:50:33,390
maybe we do it less so now than we have in the past in history.
813
00:50:33,390 –> 00:50:37,070
We try to pretend that nation states are somehow this amorphous collection of
814
00:50:37,070 –> 00:50:40,790
ideas. But nation states actually do have their own psychology and
815
00:50:40,790 –> 00:50:44,510
their own character. And history reveals that. And
816
00:50:44,510 –> 00:50:47,190
Liddell Hart there is brilliant
817
00:50:48,150 –> 00:50:51,910
in basically saying, if you’re going to make a promise,
818
00:50:51,910 –> 00:50:55,620
keep it, but understand the character as a nation
819
00:50:55,620 –> 00:50:58,980
state of the nation state
820
00:50:59,220 –> 00:51:02,540
that you are making the promise to understand their
821
00:51:02,540 –> 00:51:06,020
character, study them, examine them,
822
00:51:06,660 –> 00:51:10,100
don’t just. And then this gets to our current geopolitical
823
00:51:10,580 –> 00:51:14,420
climate. Don’t just, like in the case of NATO, sign a piece
824
00:51:14,420 –> 00:51:18,180
of paper 80 years ago and then just sort of pretend like everything’s
825
00:51:18,180 –> 00:51:22,030
the same as it was 80 years ago. And nation
826
00:51:22,030 –> 00:51:25,150
states, characters change just like people.
827
00:51:25,550 –> 00:51:28,710
France now is
828
00:51:28,710 –> 00:51:32,390
significantly different than they were at
829
00:51:32,390 –> 00:51:35,230
the back end of World War II after being
830
00:51:35,470 –> 00:51:38,670
humiliated by the Germans. They’re significantly different.
831
00:51:39,950 –> 00:51:42,990
They can pay for the protection of their own
832
00:51:43,150 –> 00:51:46,110
continent, by the way, this is all that Trump is saying, by the way. He’s
833
00:51:46,110 –> 00:51:49,890
been saying this since his first administration. Maybe, maybe you
834
00:51:49,890 –> 00:51:53,210
could pay for the protection of your own continent,
835
00:51:53,370 –> 00:51:57,010
because maybe the character of France as a nation state
836
00:51:57,010 –> 00:52:00,690
has changed. Maybe the character of Germany as a nation state has changed. Maybe the
837
00:52:00,690 –> 00:52:04,370
character of Sweden and Finland and at all has
838
00:52:04,370 –> 00:52:07,530
changed since World War II.
839
00:52:08,090 –> 00:52:11,850
Maybe we don’t just have to keep honoring these guarantees
840
00:52:11,850 –> 00:52:15,500
in perpetuity. Now, the other
841
00:52:15,500 –> 00:52:19,140
idea in there, which I find to be interesting, is this idea of being a
842
00:52:19,140 –> 00:52:22,900
person of honor. And this gets to diplomacy, right? And
843
00:52:22,900 –> 00:52:26,580
so if civilization is built on promises, and
844
00:52:27,140 –> 00:52:30,820
one of the big lessons that Liddell Hart learned from World War I
845
00:52:31,300 –> 00:52:32,260
was how,
846
00:52:35,060 –> 00:52:38,860
how fatuous, to a certain degree, diplomacy
847
00:52:38,860 –> 00:52:42,640
really was in the run up to that war. And so
848
00:52:42,640 –> 00:52:46,400
that mistake where, interestingly enough, has been corrected, we now have more
849
00:52:46,400 –> 00:52:50,160
diplomatic venues to get more people to talk as leaders of
850
00:52:50,160 –> 00:52:53,520
nation states to talk than ever before in the History of the world. It’s kind
851
00:52:53,520 –> 00:52:56,760
of insane that we have a un. That’s actually kind of nuts in the history
852
00:52:56,760 –> 00:53:00,600
of the world, and we don’t appreciate how. Bananas in pajamas, that is.
853
00:53:00,600 –> 00:53:04,440
We just don’t. We just don’t. And the first place that we don’t appreciate
854
00:53:04,440 –> 00:53:08,190
that, I think, is in how we teach history to
855
00:53:08,190 –> 00:53:11,190
the next generation. So Tom hit on this,
856
00:53:12,630 –> 00:53:16,390
and this is an interesting point. The biggest challenge in teaching
857
00:53:16,390 –> 00:53:20,110
history is getting people to care
858
00:53:20,110 –> 00:53:23,030
about it who were born after all of that was done.
859
00:53:24,950 –> 00:53:28,550
And so that’s the question to Tom. How do we get people to care?
860
00:53:29,270 –> 00:53:32,870
How do we teach history to people who were born after
861
00:53:32,870 –> 00:53:36,530
all of that was over? Well, and before
862
00:53:36,530 –> 00:53:40,250
you. Before you get to that challenge, I think what happens even
863
00:53:40,250 –> 00:53:44,010
before that is even if you. Let’s say. Let’s say you don’t get
864
00:53:44,010 –> 00:53:47,290
them excited about it, but you at least get them to read it, there’s another
865
00:53:47,290 –> 00:53:51,090
bias that happens that psychology proves that there’s a. An
866
00:53:52,290 –> 00:53:56,130
in. What’s it. What the hell would they. I forget how they worded it. I
867
00:53:56,130 –> 00:53:59,730
don’t remember the title they gave it, but it’s like a. It’s like an obvious.
868
00:54:00,530 –> 00:54:04,050
An obvious ton ability factor. My point is,
869
00:54:04,450 –> 00:54:07,370
so when you. When you read. When you read the history of World War II
870
00:54:07,370 –> 00:54:11,090
and you see, oh, like, oh, you watched this guy come into power and
871
00:54:11,090 –> 00:54:14,770
nobody really wanted. And then it’s almost like you see the inevitability
872
00:54:15,170 –> 00:54:19,010
without somebody actually teaching it to you. Well, we’re not speaking
873
00:54:19,010 –> 00:54:22,170
German right now, so obviously we won, right? Like, so then, like, there’s like an
874
00:54:22,170 –> 00:54:26,010
obvious factor that happens in reading history that. That
875
00:54:26,010 –> 00:54:29,690
gives you a bias that you already know the outcome. So
876
00:54:29,920 –> 00:54:33,760
another reason why today that. That whole history repeats itself kind of
877
00:54:33,760 –> 00:54:37,400
thing is because we. We have this expectation of
878
00:54:37,400 –> 00:54:40,760
inevitability, right? That because history showed us this
879
00:54:40,760 –> 00:54:44,520
inevitability, now we’re going to expect this inevitability. And whether it happens or
880
00:54:44,520 –> 00:54:47,520
not, which, by the way, it usually does.
881
00:54:48,160 –> 00:54:51,360
That’s why we keep saying the more things stay the same,
882
00:54:51,760 –> 00:54:55,520
but. But we keep making the same mistakes over and over again
883
00:54:55,600 –> 00:54:58,560
because we take the inevitability factor into it
884
00:54:59,130 –> 00:55:02,570
subconsciously. So to your point about getting the next
885
00:55:02,570 –> 00:55:06,250
generation to learn to care
886
00:55:06,250 –> 00:55:09,610
about it, that’s another reason why you and I get
887
00:55:09,690 –> 00:55:13,370
involved so heavily in conversations about film. It’s a
888
00:55:13,370 –> 00:55:17,090
media in which we can get the next generation to understand and learn from some
889
00:55:17,090 –> 00:55:20,930
of those historical events. You produce Band
890
00:55:20,930 –> 00:55:24,330
of Brothers or Saving Private Ryan or.
891
00:55:24,720 –> 00:55:27,920
And you Get a the next generation to watch that movie. And they go, oh
892
00:55:27,920 –> 00:55:31,080
my God. And they think it’s just cinematography. And then they realize it’s a real
893
00:55:31,080 –> 00:55:34,880
thing, it actually happened. They’re like, oh my God, I should go
894
00:55:34,880 –> 00:55:38,680
learn about World War. Like we have mechanisms that we can use
895
00:55:38,680 –> 00:55:41,920
and pull the lever on to get the next generation to care.
896
00:55:43,200 –> 00:55:47,000
Well, we did. I’m not sure they care anymore. They don’t watch it. They don’t
897
00:55:47,000 –> 00:55:50,160
watch all that many more. Maybe we should put them in TikTok videos. Maybe that’s
898
00:55:50,160 –> 00:55:53,830
what it is. Just short 30 second clips of what happened with
899
00:55:53,830 –> 00:55:57,390
World War II. And maybe then we’ll get some the next generation to care about
900
00:55:57,390 –> 00:55:59,710
history the. Way that we do. But
901
00:56:00,910 –> 00:56:04,630
kidding aside though, but that, that’s. We, we’ve got to
902
00:56:04,630 –> 00:56:08,390
stop. We’ve got to get it out of the. We, we have to. We. If
903
00:56:08,390 –> 00:56:11,790
we can change the way the mechanisms in which we teach,
904
00:56:12,270 –> 00:56:16,110
we will be able to touch the hearts of the next generation. Because
905
00:56:16,770 –> 00:56:20,130
when you and I were growing up, it was books, it was
906
00:56:20,450 –> 00:56:24,050
our imagination that reading these books and, and how
907
00:56:24,050 –> 00:56:27,850
they impacted us. This generation doesn’t really care
908
00:56:27,850 –> 00:56:31,570
so much about books, but they care about media. And so like some of these,
909
00:56:31,570 –> 00:56:35,090
like video media and all that. So if we can use that,
910
00:56:36,210 –> 00:56:39,170
all we need to do is get their attention. Once we get their attention and
911
00:56:39,170 –> 00:56:42,520
they care about it, then they’ll go back and, and figure out the rest. Right.
912
00:56:42,520 –> 00:56:46,120
So there’s a lot of talk about teachers in this country, particularly the five years
913
00:56:46,120 –> 00:56:49,680
after Covid. Right. Because you know, we all went on lockdown and then,
914
00:56:50,320 –> 00:56:53,760
you know, everybody who’s a parent who had their kid in the public school system
915
00:56:54,480 –> 00:56:57,919
kind of looked over the shoulder. This is kind of what happened. Actually, not kind
916
00:56:57,919 –> 00:57:01,120
of. This is what happened. Everybody who had a kid in public school
917
00:57:01,680 –> 00:57:05,040
all of a sudden had their kids at home learning off of a laptop
918
00:57:05,360 –> 00:57:09,170
and looking over the shoulder and actually for the
919
00:57:09,170 –> 00:57:10,170
first time in
920
00:57:13,290 –> 00:57:16,890
a long time in America, actually seeing for
921
00:57:17,610 –> 00:57:21,450
four, six, eight hours a day what a teacher is actually teaching
922
00:57:21,450 –> 00:57:25,050
their kids. This is why things have started to crack apart with the K through
923
00:57:25,050 –> 00:57:28,850
12 system, which by the way, the unions were
924
00:57:28,850 –> 00:57:32,610
the ones that insisted on lockdowns and worked
925
00:57:32,610 –> 00:57:36,170
in concert with the government and insisted on and still insist on lock
926
00:57:36,170 –> 00:57:39,820
on, on. What are you teaching students
927
00:57:39,820 –> 00:57:43,660
from, from home in many areas, particularly urban areas of
928
00:57:43,660 –> 00:57:47,340
our country. Even though we know three point about statistical
929
00:57:47,340 –> 00:57:50,220
data, we have good research now that
930
00:57:51,340 –> 00:57:54,460
kids lose a step when they are taught virtually
931
00:57:55,820 –> 00:57:59,340
particularly if they are being switched from being taught in person to
932
00:58:00,060 –> 00:58:03,840
being taught virtually, we know this. This is a fact. Okay?
933
00:58:04,240 –> 00:58:08,080
It’s like that two plus two thing. It just is. This is
934
00:58:08,080 –> 00:58:11,840
the thing. Okay? We could argue about why. We could argue
935
00:58:11,840 –> 00:58:14,840
about what the inputs are. All that, that’s fine. But you can’t argue with the
936
00:58:14,840 –> 00:58:18,680
fact, okay? We talk a lot about
937
00:58:18,680 –> 00:58:21,880
teachers in this country and how much teachers get paid and we lament and we
938
00:58:21,880 –> 00:58:24,400
wring our hands and all of that. I don’t want to get into any of
939
00:58:24,400 –> 00:58:27,120
that. Instead, I want to talk about something that’s a little bit more
940
00:58:28,320 –> 00:58:31,960
egregious, I think, which is the
941
00:58:31,960 –> 00:58:35,600
fact that. And this goes to your point about students and TikTok,
942
00:58:36,320 –> 00:58:40,120
I would agree. However, I
943
00:58:40,120 –> 00:58:43,760
think students. I agree. And I think that students
944
00:58:43,760 –> 00:58:47,520
care about ideas the same way students always
945
00:58:47,600 –> 00:58:51,280
have. And what we lack are
946
00:58:51,280 –> 00:58:54,560
teachers willing to engage
947
00:58:54,640 –> 00:58:56,560
passionately with ideas
948
00:58:58,980 –> 00:59:02,780
about the subject matter they are hired to teach. And
949
00:59:02,780 –> 00:59:06,420
this is hugely important with history, right? So, for instance,
950
00:59:07,380 –> 00:59:11,220
if I am going to teach. No. If
951
00:59:11,220 –> 00:59:15,020
I’m going to going to attend a class on an area that
952
00:59:15,020 –> 00:59:18,780
you know a lot about in history, Native American history, it
953
00:59:18,780 –> 00:59:21,620
doesn’t matter whether I agree with your conclusions or not about that history.
954
00:59:22,420 –> 00:59:26,140
None of that matters. None of that matters. I’m
955
00:59:26,140 –> 00:59:29,700
going there to tie into your passion about
956
00:59:29,700 –> 00:59:33,300
that because your passion is going to
957
00:59:33,300 –> 00:59:37,060
either make me care more or it’s going to make
958
00:59:37,060 –> 00:59:40,580
me disagree. But either way, I will give you attention
959
00:59:40,820 –> 00:59:44,300
because of the passion that I see coming from you. This is the whole plot
960
00:59:44,300 –> 00:59:48,020
of the movie Dead Poets Society. How do you make poetry and Shakespeare
961
00:59:48,020 –> 00:59:51,460
interesting? Right? Well, you make poetry and Shakespeare interesting by having a
962
00:59:51,460 –> 00:59:55,290
dynamic person teaching poetry and Shakespeare. That’s how you make it interesting.
963
00:59:56,160 –> 00:59:58,800
Same thing with history. And yet
964
01:00:00,240 –> 01:00:04,000
when you hear teachers talk about teaching history, it’s
965
01:00:04,000 –> 01:00:07,360
either one of two poles. You either have a person who’s got a personal,
966
01:00:07,920 –> 01:00:11,360
as a teacher, got a personal bugaboo of a thing that they’re drilling in on
967
01:00:11,680 –> 01:00:15,240
that’s interesting to them, but it’s not interesting maybe to 99 of the people that
968
01:00:15,240 –> 01:00:19,080
they’re teaching. You see this in colleges a lot. Or you get
969
01:00:19,080 –> 01:00:21,440
a person who is in a K through 12 space where
970
01:00:23,190 –> 01:00:26,790
they would love to be passionate about history, but they’re fighting uphill
971
01:00:26,790 –> 01:00:30,590
against a lack of comprehension, a lack of preparation, kids
972
01:00:30,590 –> 01:00:34,230
falling behind the policies of the system of the blah, blah, blah, blah,
973
01:00:34,230 –> 01:00:37,829
blah. And they would love to be passionate, but they’re not going to try to
974
01:00:37,829 –> 01:00:41,190
break the system because they’re getting paid $30,000 a year. They just want their pension
975
01:00:41,190 –> 01:00:43,430
at the end of it. They’re just not going to try to break the system.
976
01:00:43,510 –> 01:00:46,830
And by the way, they know that the teacher. Because I know teachers in K
977
01:00:46,830 –> 01:00:50,110
through 12 system over the course of my life. They know that the teacher down
978
01:00:50,110 –> 01:00:53,340
the hall who’s supposed to be teaching English comprehension or
979
01:00:53,580 –> 01:00:57,220
spelling or whatever, is it doing their job. They know
980
01:00:57,220 –> 01:01:00,900
this. They know, they know. And so by the time that
981
01:01:00,900 –> 01:01:03,900
kid shows up in sixth period for history,
982
01:01:05,260 –> 01:01:08,860
they’re fighting uphill against the previous, you know, the other previous five
983
01:01:08,860 –> 01:01:12,340
periods of nonsense. And so they’re just trying to get through the period and get
984
01:01:12,340 –> 01:01:16,020
the kid out. And those are the two poles
985
01:01:16,020 –> 01:01:17,660
the teachers are fighting a pill against.
986
01:01:19,790 –> 01:01:23,390
And history is so critical for our time. Now, I would argue that it’s
987
01:01:23,390 –> 01:01:26,910
probably as critical or more critical even than English than
988
01:01:26,910 –> 01:01:29,630
English, right? Even being able to speak and write well.
989
01:01:31,550 –> 01:01:34,269
Because history is the place where, for
990
01:01:35,790 –> 01:01:39,310
it’s the platform where our political battles are being fought.
991
01:01:39,870 –> 01:01:43,710
It’s the place where ideologies are being
992
01:01:43,710 –> 01:01:46,920
made and are being rendered. It’s even a place where, oddly enough,
993
01:01:47,000 –> 01:01:50,360
identities are being formed of all kinds.
994
01:01:51,160 –> 01:01:55,000
And teachers have a huge, huge responsibility, particularly
995
01:01:55,000 –> 01:01:57,960
in the K12 system. Have a huge responsibility. And I don’t know how you fix
996
01:01:57,960 –> 01:01:58,840
the two poles problem.
997
01:02:02,520 –> 01:02:05,000
Well, and to your point, I think,
998
01:02:07,960 –> 01:02:11,690
I think history is too big of a subject for
999
01:02:11,690 –> 01:02:15,210
that to be the, the thing that they use to try to break the polls.
1000
01:02:15,210 –> 01:02:17,890
Right? So. Because yeah, like
1001
01:02:19,170 –> 01:02:22,610
if you just think about. So I remember, I remember going through The K through
1002
01:02:22,610 –> 01:02:25,650
12 system, whatever in public schools. And I remember,
1003
01:02:26,690 –> 01:02:30,050
I remember like in. So we had to take four years when I got to
1004
01:02:30,050 –> 01:02:33,570
high school. There were four years of high school, right? 9, 10, 11, 12 Y.
1005
01:02:33,570 –> 01:02:37,370
And the, and they, they. This is not a, a
1006
01:02:37,370 –> 01:02:40,960
college environment where you get to select your, you know, your
1007
01:02:40,960 –> 01:02:44,080
course, your courses, right? So they. You dictated
1008
01:02:44,960 –> 01:02:48,800
year one. Year one was world history, year two was US
1009
01:02:48,800 –> 01:02:51,360
history, year three was.
1010
01:02:54,320 –> 01:02:57,960
US. History post World War I, I think it was. So it was
1011
01:02:57,960 –> 01:03:01,440
US history one which was like from the beginning of the country,
1012
01:03:02,400 –> 01:03:05,720
you know, 1500s, uh, up until, you know,
1013
01:03:05,720 –> 01:03:09,160
1918, World War I. And then it was
1014
01:03:09,560 –> 01:03:12,760
history from World War u. S history. Two
1015
01:03:13,240 –> 01:03:16,880
was World War I to present day. And then the, the
1016
01:03:16,880 –> 01:03:20,720
four. The fourth year was, was some sort of like there
1017
01:03:20,720 –> 01:03:24,280
was some nonsense. I say nonsense, but it was because it was like, it was
1018
01:03:24,280 –> 01:03:28,040
like theoretical history. Like it was like a future thing. Like we were, we were
1019
01:03:28,040 –> 01:03:31,830
trying to predict like his history predictions or something. Some craziness like that.
1020
01:03:31,830 –> 01:03:34,630
Which by the way, not a single one of us will write about anything. But
1021
01:03:34,630 –> 01:03:38,310
whatever. But because we’re not driving
1022
01:03:38,310 –> 01:03:42,110
flying cars right now. Yeah. We literally just talked
1023
01:03:42,110 –> 01:03:45,590
about this before we hit the record button. We’re not flying cars. We’re not teleporting
1024
01:03:45,590 –> 01:03:49,430
anywhere. We don’t have the Galactic Empire, the Galactic Federation
1025
01:03:49,430 –> 01:03:53,070
of Planets. We don’t have any of that yet. And of my graduating class in,
1026
01:03:53,310 –> 01:03:56,910
in whatever night, early night, whatever, the 19, whatever,
1027
01:03:56,910 –> 01:04:00,610
whatever. We all thought that was happening because we watched
1028
01:04:00,610 –> 01:04:04,290
the Jetsons growing up. Right. Whatever. Anyway, but, but
1029
01:04:04,290 –> 01:04:08,010
anyway, to your, to your point, the. When they, when they
1030
01:04:08,010 –> 01:04:11,810
said to us, like, you know, freshman year is
1031
01:04:11,810 –> 01:04:15,370
going to be world history. And we were like, okay, cool. Whose world?
1032
01:04:15,690 –> 01:04:18,890
What world? Like, because. And they, they went through these,
1033
01:04:19,370 –> 01:04:23,010
they went through these timelines so fast. Yeah. You
1034
01:04:23,010 –> 01:04:26,480
couldn’t get passionate about any of them even if you wanted to. Right. Like
1035
01:04:26,640 –> 01:04:30,320
I. Most of what I learned about in world history my freshman year, I had
1036
01:04:30,320 –> 01:04:33,160
to go reback and I had to go back and revisit things that I thought
1037
01:04:33,160 –> 01:04:36,280
were even a little bit interesting to see if I was curious about him. The
1038
01:04:36,280 –> 01:04:40,040
main dynasty in China, the Samurai. All the like,
1039
01:04:40,040 –> 01:04:43,120
we, and we talked about some of these things on your podcast, but I had
1040
01:04:43,120 –> 01:04:46,880
to go backwards and after the fact to see, like, was that
1041
01:04:46,880 –> 01:04:50,560
worth my time and effort to dig a little deeper because they had.
1042
01:04:50,640 –> 01:04:54,310
It’s so surface level and to your point, if I’m a history
1043
01:04:54,310 –> 01:04:57,630
teacher, for example. To your point. Exactly.
1044
01:04:58,510 –> 01:05:02,310
And I would make the argument that being a teacher in this country
1045
01:05:02,310 –> 01:05:06,070
is the easiest job on the planet. It is the easiest job on the planet
1046
01:05:06,070 –> 01:05:09,110
because number one, you’re not doing it at all unless you
1047
01:05:09,110 –> 01:05:12,750
absolutely love it and you really want to be there
1048
01:05:12,830 –> 01:05:16,550
because as you know and me being on this. But I love history and there’s
1049
01:05:16,550 –> 01:05:20,280
certain parts of history that I love way more than others. I could never
1050
01:05:20,280 –> 01:05:23,720
be a history teacher because I would want to teach just that. And they won’t
1051
01:05:23,720 –> 01:05:27,200
let me. Right. Right. Because I have to follow some syllabus and.
1052
01:05:27,200 –> 01:05:30,800
Etc. So I have to love teaching more than
1053
01:05:30,800 –> 01:05:33,960
anything else in my being in order to be a history teacher.
1054
01:05:34,200 –> 01:05:38,040
Yep. Which I don’t. Which is why I’m not a history teacher. So
1055
01:05:38,040 –> 01:05:41,080
if I, if I were, that’s why I say, and by the way, any teacher
1056
01:05:41,080 –> 01:05:43,400
out there, you don’t have to come find me to kill me. I’m not suggesting
1057
01:05:43,400 –> 01:05:47,010
you have an easy job. That is not what I said. I didn’t say you’re
1058
01:05:47,330 –> 01:05:50,250
the function of your job is easy. I’m just saying it’s the easiest job in
1059
01:05:50,250 –> 01:05:53,730
the planet because you’re not doing it unless you love it. And they always
1060
01:05:53,810 –> 01:05:56,450
say if you love what you do, you never work a day in your life.
1061
01:05:56,450 –> 01:06:00,290
So there you go, Ex. So you have the easiest job
1062
01:06:00,290 –> 01:06:03,050
if you, you’re doing it because you love it. And if you love it, then
1063
01:06:03,050 –> 01:06:06,650
it’s not really work. And if you, if it’s not really work, then stop complaining
1064
01:06:06,650 –> 01:06:10,010
about it anyway. But no, I’m kidding. I’m totally, totally
1065
01:06:10,010 –> 01:06:13,730
kidding. No, no, it’s fine. I, I already nuked. I already nuked teachers already.
1066
01:06:13,730 –> 01:06:17,390
I nuked the unions. I’m in far more trouble than you are. Oh,
1067
01:06:17,390 –> 01:06:19,870
yeah, because you went after the union. I’m just going after the individual. Oh, yeah,
1068
01:06:20,030 –> 01:06:23,790
yeah, yeah, yeah, right. More than me. They’re gonna. Yeah, Randy Baumgarten’s gonna
1069
01:06:23,790 –> 01:06:27,590
find me in about 10 seconds. Anyway, sorry, go ahead, keep going. Yeah, no, but,
1070
01:06:27,590 –> 01:06:30,750
but anyway. But the point, like, to your point, it’s, it’s part of the problem
1071
01:06:30,750 –> 01:06:34,190
that we have with, with, especially with subject matter like this book,
1072
01:06:34,590 –> 01:06:37,150
is that the subject matter is so
1073
01:06:38,190 –> 01:06:41,270
intense, so wide, so varied, so
1074
01:06:41,270 –> 01:06:45,040
expansive, that there’s no way that a single individual is
1075
01:06:45,040 –> 01:06:48,160
going to be able to teach all of it. And if we focus on one
1076
01:06:48,160 –> 01:06:52,000
thing or another, it’s, it’s not going to be. It’s, it’s so isolating
1077
01:06:52,000 –> 01:06:55,360
that there. You’re not understanding all of the other factors involved. For
1078
01:06:55,360 –> 01:06:59,200
example, and to your, to your point about, like, if you really
1079
01:06:59,200 –> 01:07:02,280
go deep, if you could literally be a teacher about
1080
01:07:02,760 –> 01:07:06,600
the preamble to World War II, never mind the actual war, but
1081
01:07:06,600 –> 01:07:10,200
the, the events that led up to World War II, you could teach an entire
1082
01:07:10,280 –> 01:07:13,040
semester college syllabus on just that
1083
01:07:14,080 –> 01:07:17,840
if somebody was willing to take it. But they don’t
1084
01:07:17,840 –> 01:07:21,600
find that valuable enough to take that course, like that one course about
1085
01:07:21,920 –> 01:07:25,600
the, you know, the, the. The era between World War I and World
1086
01:07:25,600 –> 01:07:29,040
War II that basically caused World War II to happen in the first place.
1087
01:07:29,360 –> 01:07:33,120
All the geopolitical landscaping that happened, all the political power shifting that
1088
01:07:33,120 –> 01:07:36,800
happened, all that stuff in one instance. And by the
1089
01:07:36,800 –> 01:07:40,410
way, not just the time frame, forget about just the time frame,
1090
01:07:40,490 –> 01:07:44,250
but what was happening in Germany was different than what was happening
1091
01:07:44,250 –> 01:07:47,730
in the rest of Europe that was different. That was happening in China. Did China
1092
01:07:47,730 –> 01:07:50,490
have anything to do with this? Did the United States being so,
1093
01:07:51,690 –> 01:07:55,490
so arrogant and egotistical, thinking that they were too strong to even
1094
01:07:55,490 –> 01:07:59,210
bother with it in the first place. What was that geopolitical landscape?
1095
01:07:59,210 –> 01:08:02,970
How was that economy working for you, by the way? Black. Black Friday was in
1096
01:08:02,970 –> 01:08:06,680
there. In. Involved in that as well. Like there. There. There’s so many different
1097
01:08:06,680 –> 01:08:10,480
components to what just brought us to war. One that. Sorry.
1098
01:08:10,480 –> 01:08:14,040
To. To World War II, that you could not be an expert at all of
1099
01:08:14,040 –> 01:08:17,840
it. You just can’t. So. Right. That’s another reason why we don’t learn from the
1100
01:08:17,840 –> 01:08:21,480
mistakes of history. It’s too vast and we just don’t
1101
01:08:21,480 –> 01:08:25,200
get deeply involved in enough to, like, actually pluck out
1102
01:08:25,200 –> 01:08:28,920
the real lessons. So this goes. But this goes
1103
01:08:28,920 –> 01:08:31,840
to. Okay, so this goes to a basic problem with. With the. With the structure
1104
01:08:31,840 –> 01:08:34,650
of the public school system, which we never actually talked about that on this podcast.
1105
01:08:34,800 –> 01:08:38,640
It’s been a long time coming. The structure of K through 12
1106
01:08:38,720 –> 01:08:42,160
schooling is built, of course, on the
1107
01:08:42,160 –> 01:08:45,840
Henry Ford industrialization model,
1108
01:08:46,240 –> 01:08:50,080
which, by the way, takes its. Talk about history, takes
1109
01:08:50,080 –> 01:08:53,920
its cue, or took its cue from a Prussian model,
1110
01:08:53,920 –> 01:08:57,760
a German Prussian model of education, which was focused on
1111
01:08:58,160 –> 01:09:01,760
getting people just enough information to become
1112
01:09:02,000 –> 01:09:05,600
rote soldiers. Right. To be able
1113
01:09:05,600 –> 01:09:09,120
to. Because this is what. This is what not only Kaiser Wilhelm.
1114
01:09:09,680 –> 01:09:13,440
Well, every German leader from Kaiser Wilhelm all the way to Adolf
1115
01:09:13,440 –> 01:09:16,960
Hitler wanted. Right. Was
1116
01:09:17,040 –> 01:09:20,840
compliant troops that would do what they
1117
01:09:20,840 –> 01:09:24,560
were told. And Henry Ford and
1118
01:09:24,560 –> 01:09:28,280
John D. Rockefeller. And John. John Dewey. Rockefeller, sorry,
1119
01:09:28,280 –> 01:09:31,390
Dewey and Ford and all those other folks got together
1120
01:09:32,110 –> 01:09:34,870
and a bunch of other folks got together and they said, we will build an
1121
01:09:34,870 –> 01:09:37,470
American public education system to take these people from the farm
1122
01:09:38,350 –> 01:09:40,030
who were used to
1123
01:09:41,950 –> 01:09:45,750
walking behind a plow and doing whatever it is that they wanted. Right? And we
1124
01:09:45,750 –> 01:09:49,390
have to turn them into industrial cogs in a room
1125
01:09:49,630 –> 01:09:53,350
so that they don’t get up and leave to turn a person into an
1126
01:09:53,350 –> 01:09:57,140
industrial cog, starting at the age of 4 or
1127
01:09:57,140 –> 01:10:00,940
5 or 6 until they are 17
1128
01:10:00,940 –> 01:10:04,620
or 18. You don’t need to teach
1129
01:10:04,620 –> 01:10:08,380
that person history. So, of course history is taught. As to your
1130
01:10:08,380 –> 01:10:12,100
point, this dizzying array of names
1131
01:10:12,100 –> 01:10:15,020
and dates and nonsense that you just blast through
1132
01:10:15,660 –> 01:10:19,500
because you’re not actually teaching people ideas or how to think, because
1133
01:10:19,500 –> 01:10:22,620
God forbid you do that. You don’t want them to think. You want them to
1134
01:10:22,620 –> 01:10:25,880
shut up and go to the factory. And this is my
1135
01:10:25,960 –> 01:10:29,640
opposition, by the way, one of my core oppositions to the entire public
1136
01:10:29,640 –> 01:10:33,160
school system in America. We need to reform the
1137
01:10:33,160 –> 01:10:36,920
whole thing. I don’t know if that’s possible.
1138
01:10:37,080 –> 01:10:39,959
My wife is way more radical on this than I am, which is why we
1139
01:10:39,959 –> 01:10:43,600
homeschool our kids. Yes, I did just say. But we homeschool our kids. My wife
1140
01:10:43,600 –> 01:10:47,280
is way more radical on this than I am. I think the
1141
01:10:47,280 –> 01:10:50,920
system probably has to be, not probably has to be reformed
1142
01:10:52,040 –> 01:10:55,400
because we no longer live in a world where
1143
01:10:55,400 –> 01:10:58,520
rote factory work,
1144
01:10:59,000 –> 01:11:02,320
even if we bring all the factories back to America, it still won’t be rope
1145
01:11:02,320 –> 01:11:05,400
factory work. It will not be the rope factory work of
1146
01:11:06,120 –> 01:11:09,400
what Henry Ford was wanting his workers to be between
1147
01:11:09,960 –> 01:11:13,720
1910 and 1930.
1148
01:11:14,360 –> 01:11:17,850
It’s not going to be that. That worked really well in the 20th century.
1149
01:11:18,010 –> 01:11:20,650
Mass industrialization worked really well. I’m not knocking it.
1150
01:11:21,690 –> 01:11:23,930
And we don’t live in that world anymore. We don’t live in a world of
1151
01:11:23,930 –> 01:11:27,530
mass industrialization. We need people to think in terms of ideas,
1152
01:11:28,650 –> 01:11:32,250
not in terms of rote responses to the test.
1153
01:11:35,210 –> 01:11:38,890
And so history, which, to your point about
1154
01:11:39,050 –> 01:11:41,450
being an expert in what happened in the world between,
1155
01:11:42,780 –> 01:11:46,340
let’s say, let’s conserve, let’s take, let’s pick a 25 year period between like, you
1156
01:11:46,340 –> 01:11:50,140
know, 1915 and, and you know, our 15 year period,
1157
01:11:50,140 –> 01:11:52,860
1950. Oh, no, no, let’s put 25, 1915 to
1158
01:11:52,860 –> 01:11:56,060
1939. Being an expert just on that
1159
01:11:56,380 –> 01:11:59,980
requires you to be as, as someone once said,
1160
01:11:59,980 –> 01:12:02,980
Richard Dawkins, I think when he was talking to Jordan Peterson, used this phrase, which
1161
01:12:02,980 –> 01:12:06,340
I really like, be drunk on ideas. You have to be drunk on the
1162
01:12:06,340 –> 01:12:10,110
ideas of that period. Because to your point,
1163
01:12:10,110 –> 01:12:13,870
there’s a lot of ideas of that period, economic, social, cultural, and on
1164
01:12:13,870 –> 01:12:17,630
and on and on. I don’t know that you have anybody in
1165
01:12:17,630 –> 01:12:21,070
K through 12 who
1166
01:12:21,070 –> 01:12:24,470
approaches because they’re still doing the rope thing.
1167
01:12:24,710 –> 01:12:28,510
Right, the rope preparation thing. And are there some systems that are changed, some
1168
01:12:28,510 –> 01:12:31,590
parts of the school system, some parts of the country that are changing? Yes. I’m
1169
01:12:31,590 –> 01:12:33,470
sure that if you’re in the sound of my voice, you’re going to think of
1170
01:12:33,470 –> 01:12:37,280
Your K through 12 public education building or your
1171
01:12:37,280 –> 01:12:40,600
student and you’re going to say, well, not my kid. That’s not what’s happening. Blah,
1172
01:12:40,600 –> 01:12:43,880
blah, blah, blah, blah. It’s great. Okay, sure. No one raindrop blames itself for the
1173
01:12:43,880 –> 01:12:47,000
flood. Okay, cool. Yeah, you’re special. Got it. All right.
1174
01:12:48,120 –> 01:12:51,800
But you see the outcomes, by the way, when those
1175
01:12:51,800 –> 01:12:54,800
kids who have gone through the K through 12 system and some of them have
1176
01:12:54,800 –> 01:12:58,520
matriculated to college now come out and
1177
01:12:58,600 –> 01:13:01,720
not only do they not know history, but they know tick tock really well.
1178
01:13:02,930 –> 01:13:06,770
They don’t know how to write a sentence. And now I, as an employer,
1179
01:13:06,770 –> 01:13:10,450
as a leader, have to lead these people around who, who have
1180
01:13:10,450 –> 01:13:14,170
never been given the who’s. Who’s whose thirst has
1181
01:13:14,170 –> 01:13:17,570
never been activated for even the history
1182
01:13:17,809 –> 01:13:21,450
of the business that they are in because their thirst was never
1183
01:13:21,450 –> 01:13:25,250
activated properly for the history of the world and
1184
01:13:25,250 –> 01:13:28,900
country that they live in. That’s a
1185
01:13:28,900 –> 01:13:32,540
tragedy. I said all of that. To say that that is an
1186
01:13:32,860 –> 01:13:35,500
absolute tragedy on how we teach history.
1187
01:13:36,620 –> 01:13:40,420
And as an amateur historian and a person who gets interested
1188
01:13:40,420 –> 01:13:44,140
in all of this stuff, it’s an absolute bugaboo for me. Like,
1189
01:13:45,020 –> 01:13:48,380
we do. We homeschool our kids and so, like, when my wife has a difficult
1190
01:13:48,380 –> 01:13:52,180
question in history that she does not know how to answer or how to propose
1191
01:13:52,180 –> 01:13:55,370
or difficult idea that pops up in history, go ask.
1192
01:13:56,090 –> 01:13:59,730
She told me to go ask. Go ask your father. Go ask him. He’ll tell
1193
01:13:59,730 –> 01:14:02,690
you. And I’ll talk to you for four hours about it. Because I just, I
1194
01:14:02,690 –> 01:14:05,930
just know. I just know the stuff. Like, sure, ask me about the collapse of
1195
01:14:05,930 –> 01:14:09,370
the rise and fall the Ottoman Empire. Like, why do I know about that?
1196
01:14:10,890 –> 01:14:13,530
Why do I need to know about that? Like the sultan here,
1197
01:14:14,490 –> 01:14:16,970
like, why do I need to know about that? Or why do I need to
1198
01:14:16,970 –> 01:14:20,290
know about the monetary system of, you know, the Greek
1199
01:14:20,290 –> 01:14:23,990
islands as described by Heraclitus?
1200
01:14:24,470 –> 01:14:27,710
Why do I need. Why, why do I know this? Because I’m passionate about the
1201
01:14:27,710 –> 01:14:31,270
ideas behind those kinds of things. And there’s always a core idea.
1202
01:14:31,350 –> 01:14:35,110
And of course the idea is of humanity advancing and building civilization
1203
01:14:35,670 –> 01:14:38,550
built on to Lidell Hart’s idea, built on promises.
1204
01:14:39,430 –> 01:14:42,550
But at the end of the day, I’m drunk on ideas. And you need a
1205
01:14:42,550 –> 01:14:46,390
history teacher to be drunk on ideas, particularly at the K12 level, because
1206
01:14:46,390 –> 01:14:50,070
my, the kids will pick up on that passion and then they’ll lock it.
1207
01:14:53,110 –> 01:14:55,270
And by the way, you never want to ask a kid to predict the future.
1208
01:14:56,790 –> 01:15:00,590
Like, that’s ridiculous. That’s. That’s some nonsense. Did you,
1209
01:15:00,590 –> 01:15:03,069
when you were in high school, did you go to, like, not go to. But
1210
01:15:03,069 –> 01:15:05,310
were you like, in that? Did you. You don’t have to say the year you
1211
01:15:05,310 –> 01:15:09,030
graduated, but like, in the four years of high school, looking back,
1212
01:15:09,110 –> 01:15:12,910
was that part of like the years in this country where we were doing experimental
1213
01:15:12,910 –> 01:15:16,270
learning? Oh, God, yes. Yeah, yeah, yeah. Okay. All right, so you, you.
1214
01:15:16,510 –> 01:15:19,390
I’m sorry, I’m sorry.
1215
01:15:20,030 –> 01:15:23,710
Yeah, it was just. And again, it was like. But
1216
01:15:24,430 –> 01:15:27,910
there was. They were trying to teach us like,
1217
01:15:27,910 –> 01:15:31,750
predictive analytics. I. They shouldn’t have called it a history class, put it that
1218
01:15:31,750 –> 01:15:35,150
way. But, but it fell under, under the history department
1219
01:15:35,790 –> 01:15:39,360
because it was supposed to be historical data that was supposed to be predictive analytics.
1220
01:15:41,030 –> 01:15:43,990
I always thought that that class should have been under the math department, but that’s
1221
01:15:45,110 –> 01:15:48,950
whatever. So, but anyway, either way, nothing,
1222
01:15:48,950 –> 01:15:51,990
nothing that we had. Not a single one of us got a damn thing right.
1223
01:15:52,870 –> 01:15:56,630
But, but okay, so. Well, one last question.
1224
01:15:56,710 –> 01:15:59,590
So you, you brought up something that’s very interesting. So
1225
01:16:00,390 –> 01:16:04,070
I’ll ask, I’ll frame my observation the four of a question without any
1226
01:16:04,070 –> 01:16:07,760
run up. How’s your factual history? Why
1227
01:16:07,760 –> 01:16:08,880
is that so popular?
1228
01:16:12,080 –> 01:16:13,680
Well, I, I, I think,
1229
01:16:18,320 –> 01:16:22,160
I, I think it gives you retrospect and, and I think that
1230
01:16:22,160 –> 01:16:25,760
the, we talked a lot about history invoking emotions here.
1231
01:16:26,240 –> 01:16:29,960
I think sometimes I, I think sometimes
1232
01:16:29,960 –> 01:16:33,670
that counterfactual information you get a different sense,
1233
01:16:33,670 –> 01:16:37,430
you get a different emotion coming from it and then it’s up to you to
1234
01:16:37,430 –> 01:16:40,270
kind of balance that that, that too. Again,
1235
01:16:41,150 –> 01:16:43,950
this is, this is a very, this to me is a very clear
1236
01:16:45,070 –> 01:16:48,510
example of. There’s always three sides to every story.
1237
01:16:48,670 –> 01:16:51,630
His, hers and the truth. That’s why to me,
1238
01:16:51,870 –> 01:16:55,510
counterfactual information that can be very, very important when you
1239
01:16:55,510 –> 01:16:59,340
start seeing. Again, I, I’m not talking about
1240
01:16:59,340 –> 01:17:02,980
statistical, verifiable statistical data. I’m talking about
1241
01:17:03,700 –> 01:17:07,540
like, I’ll just take our, an example
1242
01:17:07,540 –> 01:17:10,980
from, from my, my own, my own
1243
01:17:11,140 –> 01:17:14,580
repertoire. Here, right here in Massachusetts, out in western part of
1244
01:17:14,580 –> 01:17:18,100
Massachusetts, there’s a small town that originally was a native community.
1245
01:17:18,820 –> 01:17:21,620
And the, the,
1246
01:17:23,710 –> 01:17:27,310
the facts of the case, the statistical data behind it
1247
01:17:27,310 –> 01:17:31,150
was a hundred armed men walked
1248
01:17:31,150 –> 01:17:34,590
into that village and killed 400 unarmed people.
1249
01:17:34,990 –> 01:17:38,750
Period. Okay. Okay. The factual versus
1250
01:17:38,750 –> 01:17:41,950
counterfactual details to it were
1251
01:17:42,830 –> 01:17:46,470
that the armed men were colonialists. The
1252
01:17:46,470 –> 01:17:50,160
native community was comprised at the time of the
1253
01:17:50,160 –> 01:17:53,840
massacre or at the time of the, the incursion. The, the time of the
1254
01:17:53,840 –> 01:17:56,120
event was mostly women and children.
1255
01:17:57,640 –> 01:18:01,480
So you had a hundred armed men going in and killing 400
1256
01:18:01,720 –> 01:18:05,120
unarmed women and children. Okay, the
1257
01:18:05,120 –> 01:18:08,960
counter. So that’s, that’s all factual information that’s supposed to invoke
1258
01:18:08,960 –> 01:18:12,680
a certain kind of feeling to you. And I’m, as I’m even saying it, I
1259
01:18:12,680 –> 01:18:16,090
get riled up a little bit inside. But so then the alternate, the
1260
01:18:16,090 –> 01:18:19,850
counterfactual information of that was the armed men had
1261
01:18:19,850 –> 01:18:23,690
intel that told them that warriors were going to be there and
1262
01:18:23,690 –> 01:18:27,210
that they should shoot on site. And it didn’t matter whether they were
1263
01:18:27,370 –> 01:18:30,810
just shoot the people and we’ll, it’s like basically drop the bodies and we’ll figure
1264
01:18:30,810 –> 01:18:34,410
it out later. Most of those armed men,
1265
01:18:35,210 –> 01:18:39,050
years later had tremendous
1266
01:18:39,050 –> 01:18:42,810
mental breakdowns. Over it. These guys did not live
1267
01:18:42,810 –> 01:18:46,530
normal, healthy lives after the fact. So they did. They were
1268
01:18:46,530 –> 01:18:49,850
basically just taking. Let me, Let me ask you if you heard this story before.
1269
01:18:50,730 –> 01:18:54,170
We killed innocent people, but we were just taking orders. We’re a military
1270
01:18:54,890 –> 01:18:58,450
faction that we’re just, we’re just doing our job. We’re just taking orders. So you
1271
01:18:58,450 –> 01:19:01,370
shouldn’t vilify us over this. Right.
1272
01:19:02,330 –> 01:19:06,170
So again, when you talk about counterfactual information or,
1273
01:19:06,170 –> 01:19:09,670
or. And again, it’s not because the.
1274
01:19:09,750 –> 01:19:13,550
Again, facts are facts, you can’t really change them, but sometimes when
1275
01:19:13,550 –> 01:19:17,110
you hear only one side of the facts and you don’t hear both
1276
01:19:17,190 –> 01:19:20,950
sides of, again, factual information, albeit. And again,
1277
01:19:21,750 –> 01:19:24,990
the point is now you have his, hers in the truth. Right? So now you
1278
01:19:24,990 –> 01:19:28,590
have the, the. The. We also don’t
1279
01:19:28,590 –> 01:19:32,110
know, the warriors that were expected to be there, were they
1280
01:19:32,110 –> 01:19:35,720
deemed, you know, criminals
1281
01:19:35,720 –> 01:19:39,480
or some sort of. What is the
1282
01:19:39,480 –> 01:19:43,280
word that I’m thinking of here? Somebody that you’re
1283
01:19:43,280 –> 01:19:46,960
targeting very specifically in warfare. I can’t remember there’s a particular name for it.
1284
01:19:47,280 –> 01:19:51,080
Oh, insurgents. Yeah. Yeah. So. So though. And
1285
01:19:51,080 –> 01:19:54,000
they were known. They. We don’t know. There were certain facts that we don’t know.
1286
01:19:54,000 –> 01:19:57,440
Now after the fact, after the deed was done, after, you know, and this came
1287
01:19:57,440 –> 01:20:00,970
out probably 100 years later,
1288
01:20:01,450 –> 01:20:05,050
that those particular warriors were in fact involved
1289
01:20:05,050 –> 01:20:08,890
in raids on, you know, on the colonial. And this, this was
1290
01:20:09,290 –> 01:20:12,970
pre. Before the time of the state of Massachusetts even existed.
1291
01:20:13,370 –> 01:20:17,090
So this is like in the 1600s. And. But so that
1292
01:20:17,090 –> 01:20:20,610
you find out after the fact that those warriors were indeed, in fact part
1293
01:20:20,610 –> 01:20:24,330
of. Of raids that were on, you know, colonial
1294
01:20:24,330 –> 01:20:28,050
towns and they, they did do some stuff. They. That could have warranted them
1295
01:20:28,530 –> 01:20:32,330
being singled out and, and going after. So
1296
01:20:32,330 –> 01:20:36,010
does that justify the hundred armed men going in there and killing 400
1297
01:20:36,010 –> 01:20:39,570
unarmed and innocent women and children? Probably not that, but
1298
01:20:39,730 –> 01:20:41,810
let me say. Let me rephrase that. No.
1299
01:20:44,610 –> 01:20:48,050
Does that negate them from. From responsibility of quote,
1300
01:20:48,050 –> 01:20:51,770
unquote, just taking orders? No. But if you
1301
01:20:51,770 –> 01:20:55,170
were in that moment and you were one of those soldiers, would you have done
1302
01:20:55,170 –> 01:20:57,800
something different? My guess is no.
1303
01:20:58,760 –> 01:21:02,280
Oh, does something different. No. Everybody’s a hero after the fact
1304
01:21:03,800 –> 01:21:07,600
or. Or a villain after the fact. Everybody’s a hero or a villain. Right. But
1305
01:21:07,600 –> 01:21:11,320
in that moment, can you truly crucify that per those. Those
1306
01:21:11,320 –> 01:21:15,120
hundred guys? Some could say yes to a
1307
01:21:15,120 –> 01:21:18,760
degree. Because again, again, walking into a village of
1308
01:21:19,000 –> 01:21:22,560
400 innocent women and children, by the way, I keep saying the word innocent because
1309
01:21:22,560 –> 01:21:26,290
that’s the way it’s Depicted, you can simply say 400 women
1310
01:21:26,290 –> 01:21:29,810
and children. Right. Yeah. You don’t. Yeah, you don’t need another adjective on there.
1311
01:21:30,770 –> 01:21:34,330
But. But again, but the word innocent is. Is purposeful.
1312
01:21:34,330 –> 01:21:38,050
It’s intended to invoke emotion to you. Right, Right.
1313
01:21:38,050 –> 01:21:41,650
So that. That’s. Again, that’s kind of what I’m. What I’m getting at here. And
1314
01:21:41,650 –> 01:21:44,850
when you’re describing the 100 men, it’s 100 armed
1315
01:21:45,090 –> 01:21:48,450
military men. Like, again, you’re supposed to get a feeling
1316
01:21:48,710 –> 01:21:52,070
of what you’re. What you’re seeing here. And you’re supposed to develop this
1317
01:21:52,230 –> 01:21:55,430
vision in your brain as to what was happening at the moment. Now, there are
1318
01:21:55,430 –> 01:21:59,230
people that will argue that at some point One of those
1319
01:21:59,230 –> 01:22:03,070
100 should have realized that they weren’t armed insurgents, that they were actually
1320
01:22:03,070 –> 01:22:06,710
women and children. Stopped the massacre. Maybe you killed
1321
01:22:06,710 –> 01:22:10,150
100 of them, but you’re not going to kill all 400.
1322
01:22:10,790 –> 01:22:14,510
Right. But that’s. That’s the argument that’s made in most of these cases
1323
01:22:14,510 –> 01:22:18,360
where you’re no longer just following orders. You are now just a
1324
01:22:18,360 –> 01:22:21,760
murderer because you’ve. You’ve walked into this village, you started shooting.
1325
01:22:21,920 –> 01:22:25,280
Nobody stopped to actually think or do or whatever. But again,
1326
01:22:25,840 –> 01:22:29,680
that’s the whole. Like, to your point about. Or the question about car. Counterfactual
1327
01:22:29,680 –> 01:22:33,520
information is. To me, it’s always that.
1328
01:22:33,840 –> 01:22:37,600
That version of his, hers and the truth. Always. There’s always some semblance
1329
01:22:37,600 –> 01:22:41,360
of. You’re gonna get a story from the victor, a story from the loser.
1330
01:22:41,360 –> 01:22:45,130
And the. The actual story is probably somewhere in the middle. Somewhere in the
1331
01:22:45,130 –> 01:22:48,330
middle. Yeah. Yeah, yeah, yeah. Well. And
1332
01:22:52,650 –> 01:22:56,250
Liddell Hart has an answer for this. Back to the book. There’s a thought about
1333
01:22:56,250 –> 01:23:00,010
this. Back to the book. Back to. Why don’t we learn from history?
1334
01:23:00,970 –> 01:23:04,330
So this is in the War and Peace section. The
1335
01:23:04,330 –> 01:23:06,650
dilemma of the Intellectual.
1336
01:23:10,900 –> 01:23:13,060
We did not plan this, folks, so this is good.
1337
01:23:15,940 –> 01:23:19,420
Neither intellectuals nor their critics appear to recognize the
1338
01:23:19,420 –> 01:23:23,060
inherent dilemma of the thinking man and its inevitability.
1339
01:23:24,020 –> 01:23:27,100
The dilemma should be faced, for it is a natural part of the growth of
1340
01:23:27,100 –> 01:23:30,860
any human mind. An intellectual ought to realize the extent to
1341
01:23:30,860 –> 01:23:34,580
which the world is shaped by human emotions, emotions uncontrolled
1342
01:23:34,580 –> 01:23:37,900
by reason. His thinking must have been shallow and his
1343
01:23:37,900 –> 01:23:41,620
observations narrow. If he fails to realize that having once
1344
01:23:41,620 –> 01:23:45,460
learned to think and to use reason as a guide, however, he cannot possibly
1345
01:23:45,460 –> 01:23:49,139
float with the current of popular emotion and fluctuate with its violent
1346
01:23:49,139 –> 01:23:52,980
changes unless he himself ceases to think or is deliberately false in
1347
01:23:52,980 –> 01:23:56,780
his own thought. And in the Latter case, it
1348
01:23:56,780 –> 01:24:00,520
is likely that he will commit intellectual suicide gradually, quote,
1349
01:24:00,520 –> 01:24:04,040
unquote, by the death of a thousand cuts. A deeper
1350
01:24:04,040 –> 01:24:07,840
diagnosis of the malady from which left wing intellectuals have suffered in the
1351
01:24:07,840 –> 01:24:11,520
past might suggest that their troubles have come not from following reason too far,
1352
01:24:11,920 –> 01:24:14,960
but from not following it far enough to realize the general power of
1353
01:24:14,960 –> 01:24:18,760
unreason. That’s an interesting point. Many
1354
01:24:18,760 –> 01:24:22,600
of them also seem to have suffered from failing to apply reason internally as
1355
01:24:22,600 –> 01:24:26,290
well as externally, through not using it with the control of their own
1356
01:24:26,290 –> 01:24:30,090
emotions. In that way, they unwittingly helped to get this country into the mess
1357
01:24:30,090 –> 01:24:33,770
of the last war and then found themselves in an intellectual mess as a
1358
01:24:33,770 –> 01:24:37,450
result. In one of the more penetrating criticisms
1359
01:24:37,450 –> 01:24:39,930
written on this subject, George Orwell
1360
01:24:41,690 –> 01:24:45,330
exposed a profound truth in saying that, quote, the energy that
1361
01:24:45,330 –> 01:24:49,170
actually shapes the world springs from emotions. He referred to the deep
1362
01:24:49,170 –> 01:24:53,010
seated and dynamic power of racial pride, leader worship, religious belief, love
1363
01:24:53,010 –> 01:24:56,580
of war. There are powerful emotions beyond these, however,
1364
01:24:56,740 –> 01:25:00,540
the energy of the intellectual himself springs from emotion, the love of
1365
01:25:00,540 –> 01:25:04,260
truth, the desire for wider knowledge and understanding. That emotion
1366
01:25:04,260 –> 01:25:06,980
has done quite a lot to shape the world, as a study of world history
1367
01:25:07,220 –> 01:25:10,900
amply shows. In the thinking man, that
1368
01:25:10,900 –> 01:25:14,540
source of energy dries up only when he ceases to believe
1369
01:25:14,540 –> 01:25:18,100
in the guiding power of thought and allows himself to become merely a
1370
01:25:18,100 –> 01:25:21,390
vehicle for the prevailing popular emotions of
1371
01:25:21,790 –> 01:25:25,590
the moment. I’m going to skip
1372
01:25:25,590 –> 01:25:28,990
down because he’s going to talk about Bertrand Russell, he says, and I’m going to
1373
01:25:28,990 –> 01:25:32,270
read this paragraph. History bears witness to the vital part that
1374
01:25:32,270 –> 01:25:36,110
prophets, quote, unquote, have played in human progress, which is evidence of
1375
01:25:36,110 –> 01:25:39,950
the ultimate practical value of expressing unreservedly the truth as
1376
01:25:39,950 –> 01:25:43,790
one sees it. Yet it also becomes clear that the acceptance and spreading
1377
01:25:43,790 –> 01:25:46,840
of their vision has always depended on another class of men, leaders
1378
01:25:47,560 –> 01:25:51,040
who had to be philosophical strategists striking a
1379
01:25:51,040 –> 01:25:54,280
compromise between truth. I think you’ve said this actually,
1380
01:25:54,760 –> 01:25:58,440
Tom, striking a compromise between truth and
1381
01:25:58,600 –> 01:26:02,199
men’s receptivity to it. Their effect
1382
01:26:02,199 –> 01:26:05,840
is often depended as much on their own limitations in perceiving the
1383
01:26:05,840 –> 01:26:09,160
truth as on their practical wisdom in proclaiming it.
1384
01:26:09,960 –> 01:26:13,800
The prophet must be stoned. That is their lot and the test of their self
1385
01:26:13,800 –> 01:26:17,440
fulfillment. A leader who is stoned, however, may merely prove that he has failed in
1386
01:26:17,440 –> 01:26:21,080
his function through a deficiency of wisdom or through confusion, confusing his
1387
01:26:21,080 –> 01:26:24,880
function with that of a prophet. Time alone can tell whether the effect of
1388
01:26:24,880 –> 01:26:28,680
such a sacrifice redeems the apparent failure as a leader that does honor to
1389
01:26:28,680 –> 01:26:32,480
him as a man. And then, finally, I’m going to Skip
1390
01:26:32,480 –> 01:26:36,080
a couple paragraphs to go down. Opposition to the truth is inevitable,
1391
01:26:36,160 –> 01:26:39,400
especially if it takes the form of a new idea. But the degree of resistance
1392
01:26:39,400 –> 01:26:42,440
can be diminished by giving thought not only to the aim but to the method
1393
01:26:42,440 –> 01:26:45,470
of approach. This is an important tactic.
1394
01:26:46,190 –> 01:26:48,830
Avoid a frontal attack on a long established position.
1395
01:26:49,630 –> 01:26:53,470
Instead, seek to turn it by a flank movement so that a more
1396
01:26:53,470 –> 01:26:56,430
penetrable side is exposed to the thrust of truth.
1397
01:26:57,390 –> 01:27:01,070
But any such indirect approach. But in any such indirect
1398
01:27:01,070 –> 01:27:04,830
approach, take care not to diverge from the truth. For nothing is
1399
01:27:04,830 –> 01:27:08,430
more fatal to its real advancement than to lapse
1400
01:27:09,060 –> 01:27:10,980
into untruth.
1401
01:27:16,420 –> 01:27:19,860
I wish this guy was still alive. I’d love to have him on the show.
1402
01:27:26,180 –> 01:27:29,700
The dilemma of the intellectual. Or,
1403
01:27:29,860 –> 01:27:33,140
you know, I mean, I. There’s a clip floating around from one of our episodes
1404
01:27:33,140 –> 01:27:36,950
of our show. We were talking about Sam Altman. If we’re so smart,
1405
01:27:36,950 –> 01:27:40,590
why aren’t we rich? Tom, you know. Yeah, I watch that a lot,
1406
01:27:41,630 –> 01:27:42,430
that clip.
1407
01:27:50,350 –> 01:27:54,190
I mean, the big
1408
01:27:54,190 –> 01:27:57,790
separation in life is between the thinker and the doer, right? And,
1409
01:27:59,230 –> 01:28:02,910
and the intellect is hobbled by his mind and the working of his faculties.
1410
01:28:03,380 –> 01:28:06,580
He’s hobbled into not doing or not acting.
1411
01:28:08,980 –> 01:28:12,420
However, the theory of how people operate can only take you so far.
1412
01:28:12,820 –> 01:28:16,620
You actually have to go into practice. You actually have to. You actually
1413
01:28:16,620 –> 01:28:20,460
have to actually interact with people, right? And the dilemma of the intellectual,
1414
01:28:20,460 –> 01:28:24,020
the, the practical dilemma that Liddell Hart is talking about
1415
01:28:24,020 –> 01:28:27,780
here relates exactly to what we’ve been talking about here on,
1416
01:28:27,860 –> 01:28:30,520
on the show. When we over
1417
01:28:30,520 –> 01:28:34,240
intellectualize history and we lack
1418
01:28:34,240 –> 01:28:37,600
the emotive where we. Or we deny the emotive power of it, or even worse.
1419
01:28:37,600 –> 01:28:40,320
And I see a lot of this happening particularly in the last, I would say
1420
01:28:40,320 –> 01:28:43,720
10 to 15 years in our country when we are. When we allow
1421
01:28:43,800 –> 01:28:47,440
history to fall into political or ideological
1422
01:28:47,440 –> 01:28:51,200
capture. Now we’re tying emotions
1423
01:28:51,200 –> 01:28:54,920
to ideology to wind people up because we don’t know
1424
01:28:55,770 –> 01:28:59,530
another way to get them to act like.
1425
01:28:59,530 –> 01:29:03,170
One of the knocks in our time is how few people, particularly
1426
01:29:03,170 –> 01:29:06,010
young people, protest
1427
01:29:06,410 –> 01:29:09,850
injustices like previous generations
1428
01:29:10,170 –> 01:29:13,930
did in the 60s and 70s. And there’s a lot of
1429
01:29:13,930 –> 01:29:17,770
different factors that go into why young people don’t protest injustices at
1430
01:29:17,770 –> 01:29:21,500
as high a rate as was
1431
01:29:21,500 –> 01:29:25,300
done by previous generations. There’s a lot of different factors that go into this.
1432
01:29:26,260 –> 01:29:30,020
But the biggest thing that I see is the protests from young people
1433
01:29:30,420 –> 01:29:34,100
that are about injustices that may or may not be
1434
01:29:34,100 –> 01:29:37,340
happening are probably some of the most ill
1435
01:29:37,340 –> 01:29:39,060
informed protests
1436
01:29:41,460 –> 01:29:44,260
in recent history because
1437
01:29:46,590 –> 01:29:50,350
they have not been given the intellectual foundation
1438
01:29:51,790 –> 01:29:54,990
They’ve just the intellectual foundation to think through their positions.
1439
01:29:55,870 –> 01:29:58,310
So I’ll give you a case in point. I can pro, this is a big
1440
01:29:58,310 –> 01:30:01,990
example and I can point to this when I see a protest on
1441
01:30:01,990 –> 01:30:05,830
an Ivy League college campus. Columbia is
1442
01:30:05,830 –> 01:30:07,630
too easy to pick on, so I’ll pick on Stanford.
1443
01:30:11,790 –> 01:30:15,620
And, and, and just for the record. Stanford is not an Ivy League school,
1444
01:30:15,620 –> 01:30:19,380
but go ahead. Well, they, they think they are, but anyway, it’s fine. I
1445
01:30:19,380 –> 01:30:23,220
know, I know they’re not, but they think they are anyway, so. Yeah,
1446
01:30:25,140 –> 01:30:28,460
and, and, and students unroll a banner in
1447
01:30:28,460 –> 01:30:32,260
protest and the banner on
1448
01:30:33,060 –> 01:30:35,700
says Gays for G.
1449
01:30:39,950 –> 01:30:43,710
There’s a fundamental, there’s several fundamental. And I’m
1450
01:30:43,710 –> 01:30:47,310
not the only person to point this out. This is why it’s so easy. There’s
1451
01:30:47,310 –> 01:30:51,030
several fundamental steps in
1452
01:30:51,030 –> 01:30:54,749
the ladder to that thought that you just
1453
01:30:54,749 –> 01:30:58,510
haven’t been educated on gay people
1454
01:30:58,510 –> 01:31:01,550
in Gaza? Well, there are none
1455
01:31:03,390 –> 01:31:07,060
because in that part of the world, Islam has a
1456
01:31:07,380 –> 01:31:10,660
tradition as a religion
1457
01:31:12,420 –> 01:31:16,020
that is not as, shall we say, tolerant as
1458
01:31:16,020 –> 01:31:19,220
Western Christianity has been. And Western
1459
01:31:19,220 –> 01:31:21,460
secularism has been around
1460
01:31:23,300 –> 01:31:26,820
the public open practice out of the closet of
1461
01:31:27,140 –> 01:31:30,700
homosexuality. And so if you
1462
01:31:30,700 –> 01:31:34,510
are unrolling a banner that proclaims that you are
1463
01:31:34,510 –> 01:31:38,230
gay and you, you are protesting against a
1464
01:31:38,230 –> 01:31:41,990
perceived injustice happening during a war in Gaza,
1465
01:31:42,470 –> 01:31:45,830
so thus you are pro the people in Gaza, you have to understand
1466
01:31:46,310 –> 01:31:50,150
that the history of the people that you claim to be in, in solidarity
1467
01:31:50,150 –> 01:31:53,990
with, those people don’t want to be in solidarity
1468
01:31:53,990 –> 01:31:57,070
with you. And by the way, this is not something that they’ve recently like come
1469
01:31:57,070 –> 01:32:00,590
to a conclusion about. They don’t want to be in
1470
01:32:00,590 –> 01:32:02,270
solidarity with you going back to like 600,
1471
01:32:04,510 –> 01:32:08,230
the 6th century. And by the way, they’ve been very clear about
1472
01:32:08,230 –> 01:32:10,110
this. They’re not hiding it.
1473
01:32:12,030 –> 01:32:15,470
So that reveals an intellectual poverty
1474
01:32:16,110 –> 01:32:19,870
that is incredibly troubling. That could be an intellectual poverty that
1475
01:32:19,870 –> 01:32:23,610
again could be solved through an actual understanding
1476
01:32:23,680 –> 01:32:27,360
understanding of the facts of history. So would you, would you
1477
01:32:27,360 –> 01:32:30,960
been, would you have been more understanding of the
1478
01:32:30,960 –> 01:32:34,520
rolling out of the banner if it said gays for Gaza and then in
1479
01:32:34,520 –> 01:32:36,960
parentheses it said we understand the irony?
1480
01:32:38,880 –> 01:32:41,960
Yes, I would have been. Yes, that, that would actually, that would actually have worked
1481
01:32:41,960 –> 01:32:45,240
for me. Yes, actually, that, that would, that would have worked for me. Absolutely. And
1482
01:32:45,240 –> 01:32:47,440
I picked that one because it’s, it’s a hot bit one, but you can see
1483
01:32:47,440 –> 01:32:50,130
it everywhere on social media. I picked that. So I picked that one. One. It’s
1484
01:32:50,130 –> 01:32:53,930
also very blatant. Like, I knew exactly where you were going with
1485
01:32:53,930 –> 01:32:57,610
this as soon as you said, the. What the banner said. Yeah, it’s very blatant.
1486
01:32:57,610 –> 01:33:00,450
I do, I do agree with you in, in a sense, like, so
1487
01:33:01,730 –> 01:33:05,570
there’s. Just a lack of, a lack of historical understanding behind that. Right, right.
1488
01:33:05,970 –> 01:33:09,730
And, and to my point, it’s not whether you are gay or
1489
01:33:09,730 –> 01:33:13,330
not. Shouldn’t have, like, who cares? I, I don’t. It’s not
1490
01:33:13,890 –> 01:33:16,530
if they’re gay or not. But if you’re going to write something like that,
1491
01:33:17,680 –> 01:33:21,480
to your point, or to my point a second ago, be clever
1492
01:33:21,480 –> 01:33:25,240
about it and make sure you let the world know you understand the irony
1493
01:33:25,240 –> 01:33:29,040
behind it. Like, we, we know that the people in Gaza don’t give a
1494
01:33:29,040 –> 01:33:32,880
crap about us, but we, we care about them. We care about the
1495
01:33:32,880 –> 01:33:36,680
right, we care about the X, Y, whatever, whatever. If you feel
1496
01:33:36,680 –> 01:33:40,400
it’s an atrocity or if you feel like it’s a humanitarian thing,
1497
01:33:40,640 –> 01:33:44,470
whatever, whatever you’re. But, but to your point, and I, I don’t disagree
1498
01:33:44,470 –> 01:33:47,990
with you, by the way, that there is a fundamental disconnect
1499
01:33:48,230 –> 01:33:51,430
with some of the, the ways in which people are
1500
01:33:52,070 –> 01:33:55,670
project that they think versus what they actually think or what
1501
01:33:55,670 –> 01:33:59,430
or, or why they think that way. There’s, there’s disconnect there. And I, I think,
1502
01:33:59,510 –> 01:34:02,390
yeah, when I saw, I, I saw something similar
1503
01:34:03,270 –> 01:34:06,590
several times. And it’s not just Stanford that has done it. There’s been. The
1504
01:34:06,590 –> 01:34:10,160
LGBTQ community has done several things like that.
1505
01:34:10,640 –> 01:34:14,400
And, and Right. I just go, but do you,
1506
01:34:15,040 –> 01:34:18,720
like, do you understand who you’re supporting? And by the way,
1507
01:34:18,800 –> 01:34:22,000
I’m not suggesting that we shouldn’t stop the, the,
1508
01:34:22,640 –> 01:34:26,280
this craziness that’s happening in the Middle East. I. No, no, no. I am for
1509
01:34:26,280 –> 01:34:30,080
stopping killing people. Believe me. I don’t care what country you’re in.
1510
01:34:30,080 –> 01:34:33,400
I had. Stop killing people. Like, why are we doing this? This doesn’t make any
1511
01:34:33,400 –> 01:34:36,940
sense. You, you referenced a few minutes ago or earlier in this
1512
01:34:36,940 –> 01:34:40,740
podcast, this, this podcast episode that we today have more
1513
01:34:40,820 –> 01:34:44,580
avenues for diplomacy than we ever have yet. Ever.
1514
01:34:44,900 –> 01:34:48,100
Yet we still have. These things happen. Like,
1515
01:34:48,660 –> 01:34:52,380
absolutely. Because. Because human nature will. Human nature will. Will out.
1516
01:34:52,380 –> 01:34:55,980
Even Ladell Hart would say this human nature will out. Like, it just, it just
1517
01:34:55,980 –> 01:34:59,780
will. People will go to war. I’m reading John Keegan’s, which we’re
1518
01:34:59,780 –> 01:35:03,220
going to cover that on the podcast too. Coming up here in an episode, John
1519
01:35:03,220 –> 01:35:06,230
Keegan’s History of the First World War War. And it is
1520
01:35:07,510 –> 01:35:11,270
the thing, the reasons he talks about for going to war are
1521
01:35:11,270 –> 01:35:15,030
the same reasons that people go to war now. It has not changed.
1522
01:35:15,110 –> 01:35:18,790
National pride. A leader’s personal hurt feelings
1523
01:35:18,950 –> 01:35:22,750
and arrogance and ego. Right. They think they can get away with it or
1524
01:35:22,750 –> 01:35:26,590
they have hurt pride as a nation or as a leader. Arrogance,
1525
01:35:26,590 –> 01:35:30,310
ego, pride, power. Right, exactly.
1526
01:35:30,310 –> 01:35:33,950
And, and so they’re just going to. There’s going to do it. So I, I
1527
01:35:33,950 –> 01:35:37,470
would. And yes, if there were a, there were a banner that were rolled out
1528
01:35:37,470 –> 01:35:41,310
underneath there, absolutely. I would, I would be
1529
01:35:41,310 –> 01:35:44,470
like, oh, okay, so you, you know what the game is. Okay, cool. I can
1530
01:35:44,470 –> 01:35:47,670
leave that. But, but because there’s not a banner there
1531
01:35:48,150 –> 01:35:51,910
now, I have to treat you or I have to treat that message. I have
1532
01:35:51,910 –> 01:35:55,190
to deal with that message. I have to examine that message critically
1533
01:35:55,750 –> 01:35:59,600
in the realm of ideas and in the realm of ideas,
1534
01:36:00,080 –> 01:36:01,120
that message.
1535
01:36:04,240 –> 01:36:07,120
It may take a brain surgeon. Normally you say it doesn’t take a brain surgeon.
1536
01:36:07,440 –> 01:36:10,960
No, no. It may take a brain surgeon to, to
1537
01:36:11,280 –> 01:36:15,120
parse critically what’s happening inside of that
1538
01:36:15,120 –> 01:36:18,800
message and to realize, again, the poverty of idea that’s at the
1539
01:36:18,800 –> 01:36:21,200
bottom of it. And by the way, there’s a lot of these messages that are
1540
01:36:21,200 –> 01:36:24,680
floating around now, like, okay, you want to be in solidarity with the Palestinians for
1541
01:36:24,680 –> 01:36:28,480
whatever identity group you come from in America. Cool beings, that’s fine. You have the
1542
01:36:28,480 –> 01:36:32,320
freedom to do that. You live in America. You should be thanking whatever deity
1543
01:36:32,320 –> 01:36:35,920
you pray to that you live in America. Cool. Fine. Understand,
1544
01:36:36,480 –> 01:36:39,280
though, that your particular affinity group,
1545
01:36:41,840 –> 01:36:44,640
the history of your particular affinity group in this country
1546
01:36:45,920 –> 01:36:49,520
is a history that is, that is uniquely situated
1547
01:36:50,160 –> 01:36:52,900
in this historical context of this, this country.
1548
01:36:54,100 –> 01:36:56,980
So now I’m going to pick on the Columbia students. Now I’m going to go
1549
01:36:56,980 –> 01:37:00,660
for them. So when you’re talking about
1550
01:37:05,540 –> 01:37:08,100
bringing the Intifada to Columbia University,
1551
01:37:09,460 –> 01:37:13,260
you don’t know what you’re talking about. You
1552
01:37:13,260 –> 01:37:16,900
just don’t know what you’re talking about. You have no idea. You, you, you’ve now,
1553
01:37:17,300 –> 01:37:21,020
you’ve now taken a historical event that you may feel emotion
1554
01:37:21,020 –> 01:37:24,780
about or a series of historical events that you may feel emotion about
1555
01:37:24,780 –> 01:37:27,220
because of the way they were taught to you by a person who’s an expert
1556
01:37:27,220 –> 01:37:30,940
on those historical events. You’ve taken that emotion, you’ve
1557
01:37:30,940 –> 01:37:34,100
channeled it into political ideology, and you’ve come up with a
1558
01:37:34,100 –> 01:37:36,820
sloganeering idea that sounds good on paper.
1559
01:37:37,700 –> 01:37:41,140
Right. We’re going to internationalize or globalize the Intifada. Right.
1560
01:37:42,260 –> 01:37:45,910
And you don’t understand what
1561
01:37:45,910 –> 01:37:48,750
you’re asking for because someone like me
1562
01:37:49,630 –> 01:37:52,750
looks at that message and goes, okay.
1563
01:37:53,390 –> 01:37:56,830
And by the way, there’s no irony underneath that you’re deadly Sincere.
1564
01:37:57,070 –> 01:38:00,790
Okay, that’s cool. I’m deadly sincere about not having the
1565
01:38:00,790 –> 01:38:04,470
Intifada here because I know exactly what that word means. And the
1566
01:38:04,470 –> 01:38:08,030
history that that word represents inside of that historical
1567
01:38:08,030 –> 01:38:11,070
context over there. That has, by the way,
1568
01:38:11,950 –> 01:38:15,350
some of it does have something to do with us as America, but it has
1569
01:38:15,350 –> 01:38:19,190
a vast majority, vast majority of that has more
1570
01:38:19,190 –> 01:38:22,750
to do with how the European colonial powers dealt with the
1571
01:38:22,750 –> 01:38:26,270
breakup of the Ottoman Empire and dealt with Turkey
1572
01:38:26,270 –> 01:38:29,830
going all the way back to the rise of the Ottoman empire in like
1573
01:38:29,830 –> 01:38:31,310
1086 or something.
1574
01:38:33,870 –> 01:38:37,110
That’s history. That’s what I’m talking about. That poverty of
1575
01:38:37,110 –> 01:38:40,830
understanding, that inability to link ideas together and have second thoughts.
1576
01:38:41,250 –> 01:38:45,090
That’s the part that’s most troubling to me around why we don’t learn
1577
01:38:45,090 –> 01:38:47,970
from history. That’s the most troubling thing to me. And of course I’m going to
1578
01:38:47,970 –> 01:38:51,050
have people who are going to email me and say, well, hey, son, yes, we
1579
01:38:51,050 –> 01:38:54,690
absolutely do understand that. It goes to 1086. And you have to understand that colonialism,
1580
01:38:54,770 –> 01:38:58,250
daba dabba, dabba, systemic oppression. And I’m going to say, well, that’s one
1581
01:38:58,250 –> 01:39:01,770
interpretation. That’s just like your interpretation,
1582
01:39:01,770 –> 01:39:03,890
man. I have a different interpretation.
1583
01:39:05,970 –> 01:39:09,290
I want to see that you have an actual depth to your
1584
01:39:09,290 –> 01:39:13,090
interpretation. And I don’t think you do, because I don’t think you studied
1585
01:39:13,090 –> 01:39:13,370
history.
1586
01:39:17,210 –> 01:39:21,010
Yeah. Oh, I’m going all in on this one. I’m going
1587
01:39:21,010 –> 01:39:24,690
all in on this episode. I got nothing on this one. I’m going all in
1588
01:39:24,690 –> 01:39:28,290
on this one. I am, because I’m passionate about history and I’m
1589
01:39:28,290 –> 01:39:31,770
passionate about people actually learning these things and actually caring
1590
01:39:32,170 –> 01:39:35,610
in sincere ways about what the words are that they say
1591
01:39:35,970 –> 01:39:39,570
and understanding the historical context that, that they exist and that they exist
1592
01:39:39,570 –> 01:39:41,570
in. And also understanding that
1593
01:39:43,730 –> 01:39:47,530
the history of this country is not the history of everywhere else. And so
1594
01:39:47,530 –> 01:39:51,370
the kinds of quirks and things that you get to get away with here, if
1595
01:39:51,370 –> 01:39:54,890
I put you anywhere else, I mean, we’re talking about the Middle east right now,
1596
01:39:54,890 –> 01:39:58,130
but let’s talk about any, any country on the African continent.
1597
01:39:58,610 –> 01:40:00,530
If I put you there, you’re going to have, you’re going to have, you’re going
1598
01:40:00,530 –> 01:40:04,180
to have a problem because it’s a different thing there. And
1599
01:40:04,180 –> 01:40:07,860
so outsourcing that or
1600
01:40:07,860 –> 01:40:11,460
insourcing those ideas to here doesn’t. Doesn’t work. And
1601
01:40:11,460 –> 01:40:15,260
then the other, the other way doesn’t work either. Outsourcing those ideas to other
1602
01:40:15,260 –> 01:40:19,020
places. Like the conversations that we have in America about social justice
1603
01:40:19,100 –> 01:40:22,940
based on our particular historical understanding of what that means don’t work
1604
01:40:22,940 –> 01:40:26,740
in France and they don’t work in England because they come
1605
01:40:26,740 –> 01:40:30,300
from a different lineage and a different heritage and a different way of thinking about
1606
01:40:30,300 –> 01:40:34,100
both society and justice which gets
1607
01:40:34,100 –> 01:40:36,940
back to historical consciousness anyway.
1608
01:40:37,900 –> 01:40:41,700
And, and you know, so, so listen, this is, this is summed up very
1609
01:40:41,700 –> 01:40:44,900
easily, right? This is subbed up very easily. Yeah, go ahead. We just did two
1610
01:40:44,900 –> 01:40:48,540
hours. We did. And, and the idea, the concept
1611
01:40:48,540 –> 01:40:52,340
here is very simply we’re never going to learn
1612
01:40:52,340 –> 01:40:55,500
from history. It’s never going to happen because
1613
01:40:55,980 –> 01:40:59,830
of the, the, the emotions behind pride, power,
1614
01:41:00,070 –> 01:41:03,670
ego. That, that’s part of, it’s,
1615
01:41:03,670 –> 01:41:06,750
it’s human nature. It’s who we are as a people, as, as a, as a,
1616
01:41:06,750 –> 01:41:10,590
as an entity, as a being. We, we, we get our, we get
1617
01:41:10,590 –> 01:41:14,350
our pride hurt, we want to fight, we get, we, we, we,
1618
01:41:14,350 –> 01:41:17,950
we see something that we don’t have that somebody else has, we want it, we’re
1619
01:41:17,950 –> 01:41:21,270
gonna. Fight. Like it, it,
1620
01:41:21,670 –> 01:41:23,510
we’re never. Now that being said,
1621
01:41:25,520 –> 01:41:29,280
if we ever get to the point where, you know, there’s
1622
01:41:29,280 –> 01:41:32,080
been lots of talks about how, you know, over
1623
01:41:32,480 –> 01:41:36,120
evolution where like you know, we’re losing our pinky, we lost our
1624
01:41:36,120 –> 01:41:39,440
appendix because certain things happen and there’s some sci fi
1625
01:41:39,679 –> 01:41:43,240
theory out there that, that says that as we, as our
1626
01:41:43,240 –> 01:41:47,080
brains get bigger and we are urged to actually that that
1627
01:41:47,080 –> 01:41:50,650
fight or flight mechanism will go away and
1628
01:41:50,810 –> 01:41:54,650
the urge to fight for power will go away because we’ll have the intellect
1629
01:41:54,970 –> 01:41:57,850
to not have to do that anymore. Until that happens,
1630
01:41:59,130 –> 01:42:02,170
we’re never going to learn from history because we are going to fight over things
1631
01:42:02,170 –> 01:42:05,810
that are prideful for us. We’re going to fight over things that we want power
1632
01:42:05,810 –> 01:42:09,650
over. We’re going to fight over things that you know, our leadership
1633
01:42:09,650 –> 01:42:13,290
has a ego, ego trip over
1634
01:42:14,090 –> 01:42:17,480
because of their own, their own self interest or wor. Like
1635
01:42:17,640 –> 01:42:21,400
that’s going to continue to happen. So and, and
1636
01:42:21,400 –> 01:42:24,840
I don’t think any, there, there was a reference to,
1637
01:42:25,400 –> 01:42:29,160
there was a reference to the, the, the, the, the, the
1638
01:42:29,160 –> 01:42:32,600
military strategy room that goes a little beyond
1639
01:42:32,840 –> 01:42:36,040
the Joint Chiefs of Staffs in, in World War II in thinking about
1640
01:42:37,320 –> 01:42:41,000
Japan would never in a million years bomb Pearl harbor because they just know how
1641
01:42:41,330 –> 01:42:43,970
big and strong the United States is. And there was like a,
1642
01:42:45,010 –> 01:42:48,850
a believability bias that happened in there where even if one person said
1643
01:42:48,930 –> 01:42:52,730
but wait, the data shows this and they
1644
01:42:52,730 –> 01:42:56,370
go yeah, I don’t care what the data says, Japan will never do it. Right?
1645
01:42:57,169 –> 01:42:59,930
Japan’s never going to do that. We’re too big, we’re too big, we’re too big,
1646
01:42:59,930 –> 01:43:03,610
etc. Etc. We’re too strong, we’re too powerful, etc. Whatever. And
1647
01:43:03,610 –> 01:43:07,110
there’s a, there’s a report even of saying that when Japan’s,
1648
01:43:08,220 –> 01:43:12,060
when Japan’s planes were in, in route,
1649
01:43:12,700 –> 01:43:16,420
a radar radar technician went
1650
01:43:16,420 –> 01:43:20,140
to his superior officer and said hey, there’s a blip on this map. It looks
1651
01:43:20,140 –> 01:43:23,980
a little funny. And the radar got the, the, the superior officer said,
1652
01:43:23,980 –> 01:43:26,700
oh well, we’re, I, I think we’re expecting our
1653
01:43:26,700 –> 01:43:29,500
B17s coming back. Don’t worry about it.
1654
01:43:31,020 –> 01:43:34,750
An hour later, Pearl harbor doesn’t exist or you know, the rest is history.
1655
01:43:34,750 –> 01:43:38,590
So yeah, so there there. So even when there is a
1656
01:43:38,590 –> 01:43:40,670
singular voice of reason
1657
01:43:42,350 –> 01:43:46,150
because of like that consensus bias that happened
1658
01:43:46,150 –> 01:43:49,990
in that we still don’t listen to it. So even if there is an
1659
01:43:49,990 –> 01:43:53,790
intellect that comes above and beyond or goes beyond or our thinking and
1660
01:43:54,350 –> 01:43:57,390
again all the things we talked about today, whether it’s military,
1661
01:43:57,390 –> 01:44:01,160
geopolitical, whether it’s business related because you don’t think that the,
1662
01:44:01,240 –> 01:44:04,880
that somebody out there is going to an like the, the
1663
01:44:04,880 –> 01:44:08,160
VC world going hey can we put a little bit of a break on this,
1664
01:44:08,160 –> 01:44:11,680
on spending on these AI companies? Because there’s going to be a bubble somewhere. And
1665
01:44:11,680 –> 01:44:14,600
then the, the consensus bias still happens and goes
1666
01:44:15,080 –> 01:44:17,880
nah, you don’t know. What you’re talking about. Nah, you don’t know what you’re talking
1667
01:44:17,880 –> 01:44:19,400
about. You don’t know what you’re talking about.
1668
01:44:22,040 –> 01:44:25,850
So foreign. We’re
1669
01:44:25,850 –> 01:44:29,570
not solving this problem today, Hasan. No, we’re not solving this problem today.
1670
01:44:29,570 –> 01:44:33,090
But maybe, maybe, maybe that we can. Identify it
1671
01:44:33,170 –> 01:44:36,890
is probably tells me that there is much smarter people than us that
1672
01:44:36,890 –> 01:44:39,650
have already identified it and they couldn’t do anything about it.
1673
01:44:41,890 –> 01:44:45,690
Well so at least, at least what we can say is this. I think we
1674
01:44:45,690 –> 01:44:49,490
can say this. I think we could sum up the our two hour long conversation
1675
01:44:49,650 –> 01:44:53,290
in this, this, this two hour long episode around why We Don’t Learn
1676
01:44:53,290 –> 01:44:55,810
from History by BH Lart with this
1677
01:44:57,570 –> 01:45:01,290
I am a proponent and I think probably you are Tom, as
1678
01:45:01,290 –> 01:45:04,930
you Tom, are as well. I am a proponent of a. More,
1679
01:45:06,050 –> 01:45:09,650
some would say tragic but maybe that’s not it. A more
1680
01:45:10,210 –> 01:45:13,810
small C. Small C conservative view of history
1681
01:45:14,530 –> 01:45:18,330
rather than a large capital P progressive view of history
1682
01:45:18,730 –> 01:45:22,410
which is the kind of view of history that quite frankly
1683
01:45:22,730 –> 01:45:26,370
folks in the west have had since the Enlightenment going back
1684
01:45:26,370 –> 01:45:30,130
to the 17th century. This idea that everything will
1685
01:45:30,130 –> 01:45:33,970
just somehow consistently get better and that
1686
01:45:33,970 –> 01:45:36,250
history is merely the Tale of
1687
01:45:37,610 –> 01:45:41,460
endless improvement going forward. I mean, you see this with the techno future optimistic
1688
01:45:41,690 –> 01:45:45,370
right now, like Mark Andreessen and others that are. To your point, they’re sitting around
1689
01:45:45,370 –> 01:45:49,210
in all those rooms dumping all this money into AI because like, the future
1690
01:45:49,210 –> 01:45:52,490
will just be better. Maybe it will, maybe it won’t.
1691
01:45:53,370 –> 01:45:56,570
The tragic view of human nature. They also serve another
1692
01:45:57,450 –> 01:46:01,210
very vital problem with the. The whole theory of history teaching us
1693
01:46:01,290 –> 01:46:04,490
is because they feel if they throw enough money at it, it will eventually get
1694
01:46:04,490 –> 01:46:08,340
better regardless. Right? You’re right. Exactly. No matter how tragic human nature is,
1695
01:46:08,490 –> 01:46:11,930
is right. Right. And. And the thing is you can’t.
1696
01:46:13,850 –> 01:46:17,530
In taking a small c. Conservative view of. Of. Of human nature.
1697
01:46:17,530 –> 01:46:21,370
Not of human nature, sorry, of history, which I, I believe
1698
01:46:21,370 –> 01:46:25,130
I’m a proponent of. You.
1699
01:46:25,370 –> 01:46:29,210
You understand that no matter how many trillions of dollars you throw at human
1700
01:46:29,210 –> 01:46:32,410
nature. And yes, I did use trillions.
1701
01:46:34,180 –> 01:46:37,660
It doesn’t matter. Human nature is impervious to your
1702
01:46:37,660 –> 01:46:41,180
trillions. And human nature is a whack a
1703
01:46:41,180 –> 01:46:44,700
mole. It pops up in all kinds of odd
1704
01:46:44,700 –> 01:46:48,380
places and it. It pushes
1705
01:46:48,380 –> 01:46:52,220
and expands boundaries, usually against your
1706
01:46:52,220 –> 01:46:55,380
will and the will of your dollars. And by the way, the will of your
1707
01:46:55,380 –> 01:46:59,220
power. You know, you could say that
1708
01:46:59,870 –> 01:47:03,670
the trillions of dollars will do it, or rules and regulations will
1709
01:47:03,670 –> 01:47:07,110
do it, or laws will do it, or punishments and
1710
01:47:07,110 –> 01:47:10,030
consequences will do it. And the fact of the matter is,
1711
01:47:11,310 –> 01:47:14,990
a small c conservative reading of history
1712
01:47:16,350 –> 01:47:18,510
says that you’re actually incorrect.
1713
01:47:19,950 –> 01:47:23,790
You’re wrong. And you’re wrong with
1714
01:47:24,510 –> 01:47:27,720
a. You’re wrong. And you’re. And you’re one more.
1715
01:47:29,240 –> 01:47:32,920
You’re one more wealthy person or wealthy system
1716
01:47:33,960 –> 01:47:35,560
or wealthy set of institutions
1717
01:47:37,720 –> 01:47:41,000
that will wind up on the ash sheep of history
1718
01:47:41,800 –> 01:47:43,720
being read about by a bunch of people
1719
01:47:45,720 –> 01:47:49,320
four or five hundred years from now going, why those people didn’t get it.
1720
01:47:50,200 –> 01:47:50,680
Yeah.
1721
01:47:54,290 –> 01:47:58,130
And so with that, there’s no upper.
1722
01:47:58,290 –> 01:48:01,650
Upper on this episode of the podcast. There’s a way to end with this. I
1723
01:48:01,650 –> 01:48:04,210
think we’re done. I think we’re just. We’re just done. Wait, hold on. I think
1724
01:48:04,210 –> 01:48:04,770
we’re done here
1725
01:48:10,930 –> 01:48:14,210
with that. We’re. We’re out.










