Welcome to the home of AMSEcast, STEM of History and the AMSE Science Report
Aug. 28, 2024

Evolving Toward a Better Future with David Sloan Wilson

Evolving Toward a Better Future with David Sloan Wilson

Alan talks with David Sloan Wilson, renowned biologist and author, to explore the broader applications of Darwin's theory beyond genetics to cultural and personal evolution. Wilson argues against conflating evolution with Social Darwinism and...

Alan talks with David Sloan Wilson, renowned biologist and author, to explore the broader applications of Darwin's theory beyond genetics to cultural and personal evolution. Wilson argues against conflating evolution with Social Darwinism and highlights cooperation as a crucial trait for societal progress. He emphasizes the need for experimental and inclusive decision-making and discusses how failure drives improvement, the impact of cultural interventions, and the role of religion in fostering community. Wilson also critiques traditional economic models and explains his aim to integrate evolutionary science into global cooperation.
 
 
Guest Bio
David Sloan Wilson is a distinguished evolutionary biologist with a doctorate from Michigan State University. His impressive academic career spans institutions such as Harvard University, the University of Washington, and the State University of New York Binghamton, where he is now Distinguished Professor Emeritus. David founded the Evolution Institute and co-founded the nonprofit ProSocial World, including the New Paradigm Coalition Initiative. He is an award-winning author known for his influential works, including This View of Life, Evolution for Everyone, The Neighborhood Project, and his novel Atlas Hugged. David’s research and writing explore the applications of evolutionary theory to society and culture.
 
 
Show Notes
  • (2:21) - What the evolution paradigm is
  • (4:22) - How the evolution paradigm is seen in cultures and how it differs from Social Darwinism
  • (6:56) - The special conditions necessary for the evolution paradigm to be effective
  • (11:51) - The importance of a common goal for cooperation to work when people have conflicting opinions
  • (14:11) - How failure is handled under the evolution paradigm
  • (16:16) - Applying the evolution paradigm to education
  • (26:17) - How the evolution paradigm applies to faith and religion
  • (37:13) - How the cooperative approach works when it comes to national economics
  • (39:20) - How individuals express themselves when they don’t agree with the larger group
  • (44:07) - Wilson’s novel, Atlas Hugged

 

Links Referenced
 
Transcript
1
00:00:04,040 --> 00:00:07,840
Welcome to AMSEcast, coming to you from Oak Ridge, Tennessee,

2
00:00:08,170 --> 00:00:11,500
a global leader in science, technology, and innovation.

3
00:00:12,240 --> 00:00:15,319
My name is Alan Lowe, director of the American Museum

4
00:00:15,339 --> 00:00:18,750
of Science and Energy, and the K-25 History Center.

5
00:00:19,700 --> 00:00:23,089
Each episode of AMSEcast presents world-renowned authors,

6
00:00:23,530 --> 00:00:27,509
scientists, historians, policymakers, and everyone in between,

7
00:00:27,939 --> 00:00:31,150
sharing their insights on a variety of fascinating topics.

8
00:00:36,379 --> 00:00:37,730
Welcome to AMSEcast.

9
00:00:37,880 --> 00:00:40,449
I’m honored today to be joined by David Sloan Wilson.

10
00:00:41,050 --> 00:00:43,989
With a doctorate from Michigan State University, David has had an

11
00:00:43,990 --> 00:00:48,539
incredible career at a variety of prestigious institutions, including—among

12
00:00:48,540 --> 00:00:52,549
others—Harvard University, University of Washington, Michigan State,

13
00:00:52,820 --> 00:00:56,029
and the State University of New York Binghamton, where he was a

14
00:00:56,030 --> 00:00:59,850
professor, now Distinguished Professor Emeritus of biological sciences

15
00:00:59,860 --> 00:01:03,470
and anthropology, and where he began their evolutionary studies program.

16
00:01:04,030 --> 00:01:07,520
David is the founder of the Evolution Institute, and co-founder of the

17
00:01:07,520 --> 00:01:12,640
nonprofit ProSocial World, which includes the New Paradigm Coalition Initiative.

18
00:01:13,189 --> 00:01:16,660
David is an award-winning prolific author, and today we’re going to touch

19
00:01:16,660 --> 00:01:21,520
upon topics in several of his books, including This View of Life: Completing

20
00:01:21,520 --> 00:01:25,910
the Darwinian Revolution, Evolution for Everyone: How Darwin’s Theory Can

21
00:01:25,910 --> 00:01:30,270
Change the Way we Think About Our Lives, The Neighborhood Project: Using

22
00:01:30,290 --> 00:01:35,820
Evolution to Improve my City, One Block at a Time, and his novel, Atlas Hugged.

23
00:01:36,440 --> 00:01:38,420
David, welcome to AMSEcast.

24
00:01:38,420 --> 00:01:39,400
Thank you so much, Alan.

25
00:01:39,670 --> 00:01:40,500
Pleasure to be here.

26
00:01:40,830 --> 00:01:44,729
And I know we talked before we started today, our mutual friend, Guru

27
00:01:44,730 --> 00:01:47,790
Madhavan, brought us together, and our thanks to him for doing that.

28
00:01:48,100 --> 00:01:49,280
Yep, absolutely.

29
00:01:49,280 --> 00:01:54,679
In his new book, Wicked Problems, I’ll be bringing that in because we are—this

30
00:01:54,680 --> 00:01:58,039
is our age of wicked problems that we’re living in right now, isn’t it?

31
00:01:58,280 --> 00:01:59,410
We have a few of those, David.

32
00:01:59,420 --> 00:02:00,126
Yes, we do [laugh]

33
00:02:00,126 --> 00:02:00,183
.
[laugh]

34
00:02:00,240 --> 00:02:04,220
.
So, I think that you and Guru together can solve them.

35
00:02:04,670 --> 00:02:05,559
I have great faith.

36
00:02:05,570 --> 00:02:07,079
So, [laugh] he’s—

37
00:02:07,150 --> 00:02:07,880
With many others.

38
00:02:08,280 --> 00:02:09,620
Only with the help of many others.

39
00:02:09,800 --> 00:02:10,820
That is very true.

40
00:02:10,850 --> 00:02:11,610
Very true.

41
00:02:11,940 --> 00:02:16,579
So, I so enjoyed learning about your work, reading your work, and

42
00:02:16,800 --> 00:02:20,920
something very central to that is something called the evolution paradigm.

43
00:02:20,920 --> 00:02:24,809
Can you tell us this as just as we start, what is the evolution paradigm?

44
00:02:26,080 --> 00:02:26,530
Right.

45
00:02:26,560 --> 00:02:29,720
Well, I mean, a way to begin is to note that

46
00:02:29,720 --> 00:02:32,980
the word evolution has a very general meaning.

47
00:02:32,990 --> 00:02:35,429
People use the word all the time for any

48
00:02:35,429 --> 00:02:38,820
kind of change or any kind of development.

49
00:02:39,000 --> 00:02:41,610
But Darwin had a very specific meaning of

50
00:02:41,610 --> 00:02:45,560
evolution as the engine of change in nature.

51
00:02:45,560 --> 00:02:49,630
And so, he was pointing out three vital ingredients: first, that

52
00:02:49,910 --> 00:02:53,409
organisms vary in just about anything that could be measured about

53
00:02:53,410 --> 00:02:57,649
them; two, that those differences make a difference; and three,

54
00:02:58,279 --> 00:03:01,570
that traits are heritable, basically passed on to their offspring.

55
00:03:01,570 --> 00:03:05,570
And so, you put those three ingredients together, and then you get

56
00:03:05,570 --> 00:03:09,700
populations changing over time and adapting to their environment.

57
00:03:09,700 --> 00:03:13,969
So, there’s Darwin’s theory of evolution in a nutshell.

58
00:03:13,969 --> 00:03:14,960
There’s point one.

59
00:03:15,400 --> 00:03:19,770
And point two is, although Darwin knew nothing about genes, once genes

60
00:03:19,770 --> 00:03:23,859
were discovered and the science of genetics was born, then the study

61
00:03:23,860 --> 00:03:28,919
of evolution became constricted to the study of genetic evolution.

62
00:03:29,349 --> 00:03:32,530
So, around the world, when one person says the word

63
00:03:32,530 --> 00:03:35,400
‘evolution,’ other people hear the word ‘genes.’

64
00:03:36,039 --> 00:03:39,100
And that’s actually not as general as it needs

65
00:03:39,100 --> 00:03:41,430
to be because there’s other forms of evolution.

66
00:03:41,780 --> 00:03:45,720
There’s cultural evolution, there’s even our personal evolution.

67
00:03:45,730 --> 00:03:48,829
And so, what’s happened recently that’s new and worth talking

68
00:03:48,830 --> 00:03:54,209
about is this expansion of evolutionary thinking to explain all

69
00:03:54,209 --> 00:03:59,090
of the fast-paced changes taking place outside and within us.

70
00:03:59,560 --> 00:04:04,660
The applicability of Darwin’s theory of evolution is what my

71
00:04:04,660 --> 00:04:08,310
career is about and what’s so exciting to report to others.

72
00:04:08,990 --> 00:04:12,594
I know you mentioned this—I’m forgetting now, David, forgive me, in which of

73
00:04:12,730 --> 00:04:18,440
the books—but you talk about the evolution of cultures and how what you’re not

74
00:04:18,440 --> 00:04:22,649
talking about, essentially, is the very discredited theory of Social Darwinism.

75
00:04:22,670 --> 00:04:27,279
So, can you tell us, in terms of cultures, how this paradigm is implemented

76
00:04:27,280 --> 00:04:32,230
or seen, and how that’s different from those old ideas of Social Darwinism?

77
00:04:32,310 --> 00:04:38,390
Yeah, so Social Darwinism is the moral justification of inequality, and

78
00:04:38,390 --> 00:04:43,099
the idea that basically, the fit—first of all, the successful people

79
00:04:43,099 --> 00:04:47,880
in the world are the most fit, and they deserve to replace the unfit.

80
00:04:48,109 --> 00:04:52,719
One point to make historically, Alan, is that it’s not the case that

81
00:04:52,720 --> 00:04:57,460
Darwin’s theory led to Social Darwinism, and all these pernicious things.

82
00:04:57,460 --> 00:05:03,419
Darwin’s theory was interpreted in so many different ways from the beginning.

83
00:05:03,429 --> 00:05:07,900
I mean, many people interpreted Darwin’s theory as saying, “Oh, my heavens, this

84
00:05:07,920 --> 00:05:13,450
means that the existing social order is not fixed.” Societies were different

85
00:05:13,450 --> 00:05:17,320
in the past than in the present, and they can be different in the future.

86
00:05:17,340 --> 00:05:23,800
So, in other words, people with a socialistic bent embraced Darwin’s theory, and

87
00:05:23,800 --> 00:05:29,010
then others did embrace this idea that the successful deserves to replace the.

88
00:05:29,010 --> 00:05:33,500
So, Social Darwinism is a thing for sure, but it wasn’t the only thing.

89
00:05:33,500 --> 00:05:36,119
And it’s not the case that Darwin’s theory

90
00:05:36,120 --> 00:05:39,820
is inherently prone to this kind of a policy.

91
00:05:39,820 --> 00:05:45,109
But in any case, what’s proven to be the case is almost the very opposite.

92
00:05:45,190 --> 00:05:50,270
Because what’s so distinctive about our species is how cooperative we are.

93
00:05:50,340 --> 00:05:53,620
Cooperation is the signature adaptation of our

94
00:05:53,620 --> 00:05:57,229
species, and cooperation is what’s needed, basically.

95
00:05:57,320 --> 00:06:01,670
And what Darwin’s theory tells us is that cooperation—and everything

96
00:06:01,670 --> 00:06:06,960
associated with it: love, altruism, sympathy, empathy, all of the positive

97
00:06:07,280 --> 00:06:12,699
emotions, everything that we regard as morally good, basically—can

98
00:06:12,710 --> 00:06:17,770
evolve by a Darwinian process, but only if special conditions are met.

99
00:06:17,859 --> 00:06:23,280
And when those conditions are not met, then we get dysfunction, selfishness,

100
00:06:23,280 --> 00:06:30,410
and conflict, and lack of coordination, and disruption in all of their forms.

101
00:06:30,410 --> 00:06:34,869
And in both nature and in human life, what we find is both.

102
00:06:34,990 --> 00:06:37,579
I mean, you don’t need to be a scientist to know that.

103
00:06:37,690 --> 00:06:42,510
We find spectacular examples of cooperation, and spectacular failures.

104
00:06:43,060 --> 00:06:46,340
And so, I think that what Darwin’s theory of evolution, properly

105
00:06:46,340 --> 00:06:51,379
understood, does is enables us to diagnose that, and then

106
00:06:51,389 --> 00:06:54,919
work, in the future, towards more cooperative arrangements.

107
00:06:55,010 --> 00:06:57,969
I want to dig into some more of these specifics,

108
00:06:57,970 --> 00:07:00,049
but you mentioned those special conditions.

109
00:07:00,060 --> 00:07:03,270
So, what are some of those special conditions

110
00:07:03,270 --> 00:07:05,670
that need to be in place for that to be effective?

111
00:07:06,410 --> 00:07:08,979
They’re embodied in a meme that I’ve coined with the

112
00:07:08,980 --> 00:07:12,490
other Wilson, the evolutionary biologist, Edward O.

113
00:07:12,490 --> 00:07:13,030
Wilson.

114
00:07:13,070 --> 00:07:16,760
And the meme is, selfishness beats altruism within groups;

115
00:07:16,960 --> 00:07:21,070
altruistic groups beat selfish groups; everything else is commentary.

116
00:07:21,090 --> 00:07:24,589
So, to unpack that meme, what we need to know—and I think what’s often

117
00:07:24,590 --> 00:07:30,030
overlooked because we’re always encouraging people to be good in one way or

118
00:07:30,240 --> 00:07:35,789
another—and what that fails to appreciate is the vulnerability of goodness.

119
00:07:35,790 --> 00:07:41,020
When we act on behalf of others or on behalf of our groups as a whole, we are

120
00:07:41,020 --> 00:07:48,330
inherently exposing ourselves to exploitation and free-riding by others who are

121
00:07:48,330 --> 00:07:53,490
not producing social benefits, and are accepting them without producing them.

122
00:07:53,990 --> 00:07:56,869
And so, the vulnerability, everything we call

123
00:07:57,259 --> 00:07:59,440
prosocial is something that you have to know.

124
00:07:59,440 --> 00:08:03,600
There is an inherent disadvantage, and that disadvantage needs to

125
00:08:03,600 --> 00:08:09,239
be countered by the advantage of when cooperators work together.

126
00:08:09,260 --> 00:08:12,390
Then, as a group, they function better than other

127
00:08:12,390 --> 00:08:15,500
groups where cooperation is not taking place.

128
00:08:15,500 --> 00:08:20,150
And so cooperation, those special conditions require a process of

129
00:08:20,469 --> 00:08:26,880
selection among groups to offset their process of selection within groups.

130
00:08:27,360 --> 00:08:30,909
So, what we’re talking about as evolution as a multi-layered

131
00:08:30,950 --> 00:08:34,260
thing, that is what’s important, and the need for higher

132
00:08:34,260 --> 00:08:37,669
levels of selection in order to evolve cooperation.

133
00:08:38,589 --> 00:08:39,970
Very, very interesting.

134
00:08:40,000 --> 00:08:44,410
Let’s dig into some of the specifics that you mentioned in your writings.

135
00:08:44,460 --> 00:08:47,240
Decision-making and the creation of public policy.

136
00:08:47,300 --> 00:08:50,660
At one point in my life, I was very fortunate to be head of the Howard

137
00:08:50,660 --> 00:08:53,510
Baker Center for Public Policy at the University of Tennessee—now the

138
00:08:53,510 --> 00:08:58,239
Howard Baker School, I’m glad to say—so I thought a lot about these topics.

139
00:08:58,299 --> 00:09:00,790
How should decision-making and the creation of public

140
00:09:00,790 --> 00:09:04,610
policy, how should or could they be informed by evolution?

141
00:09:05,340 --> 00:09:09,000
Well, one point to make is that the world is so complex for the

142
00:09:09,000 --> 00:09:12,700
decisions we need to make—back to Guru Madhavan with our wicked

143
00:09:12,700 --> 00:09:17,669
problems—that there is no group of experts that can make wise decisions.

144
00:09:17,799 --> 00:09:19,960
The world is too complex for that.

145
00:09:20,350 --> 00:09:22,500
Therefore, you have to experiment.

146
00:09:22,510 --> 00:09:26,880
And so, decision-making fundamentally has to be a form of experimentation.

147
00:09:27,150 --> 00:09:29,940
And by the way, experimentation is nothing more

148
00:09:29,940 --> 00:09:34,150
than a very conscious process of cultural evolution.

149
00:09:34,559 --> 00:09:37,370
When we make a decision, we have some goal—there’s our target of

150
00:09:37,370 --> 00:09:40,700
selection—we orient variation around the goal, we’ve heard what

151
00:09:40,710 --> 00:09:44,070
might work what might not work, and then we try it out and we

152
00:09:44,070 --> 00:09:48,560
identify what works best among those alternatives, then we replicate

153
00:09:48,560 --> 00:09:52,109
it, and then we repeat that process again, and again, and again.

154
00:09:52,119 --> 00:09:57,740
So, a decision-making process is an evolutionary, a Darwinian process.

155
00:09:57,750 --> 00:09:57,996
So, what we can is that experimentation is required.

156
00:09:57,996 --> 00:09:58,880
And the second thing we can say is

157
00:10:02,770 --> 00:10:07,430
that that needs to include all the stakeholders because you don’t have a

158
00:10:07,440 --> 00:10:12,959
special class of people who do the experimentation, so the decision-making

159
00:10:12,960 --> 00:10:16,780
process should be inclusive, it should include everyone that’s involved.

160
00:10:16,920 --> 00:10:20,750
First of all, everyone that might contribute to the solution—in other words,

161
00:10:20,750 --> 00:10:25,649
who knows about it—and anyone who might be influenced by the decision.

162
00:10:25,660 --> 00:10:30,780
So decision-making, we can say, must be experimental and inclusive.

163
00:10:31,059 --> 00:10:34,270
And let me give an example, Alan, from industry.

164
00:10:34,670 --> 00:10:38,229
One point to make is that sometimes cultural evolution works well, kind of,

165
00:10:38,230 --> 00:10:43,250
all by itself, without us explicitly thinking of it as an evolutionary process.

166
00:10:43,469 --> 00:10:49,860
And Toyota, the Toyota manufacturing method is famous, at least in the

167
00:10:49,860 --> 00:10:54,710
business world, but what it is an implementation of exactly what I described.

168
00:10:54,770 --> 00:10:58,069
So, you have this big assembly plant which has a very, very,

169
00:10:58,090 --> 00:11:02,849
very complex system, and if anything goes wrong on the assembly

170
00:11:02,860 --> 00:11:07,064
plant, then the worker there who notices it, signals it.

171
00:11:07,150 --> 00:11:09,069
It used to be that there was ropes hanging from

172
00:11:09,070 --> 00:11:11,610
the ceiling, that you’d actually pull the rope.

173
00:11:11,840 --> 00:11:16,680
And then there’s a swarm of activity, just around the solution to that problem.

174
00:11:16,870 --> 00:11:21,000
And the group that forms around it includes the workers that are there, plus

175
00:11:21,000 --> 00:11:25,830
the management whose offices are not on the top floor, but on the shop floor.

176
00:11:25,830 --> 00:11:28,680
So, the managers are working directly with the workers.

177
00:11:28,760 --> 00:11:32,200
And when they come up with a candidate solution, they don’t implement it.

178
00:11:32,620 --> 00:11:36,150
They try it out experimentally, and then they look at all the

179
00:11:36,280 --> 00:11:38,959
indirect effects that might take place, or that percolates

180
00:11:38,969 --> 00:11:42,759
through the whole system, and only then do they implement it.

181
00:11:42,820 --> 00:11:45,110
So, that’s a process of continuous improvement,

182
00:11:45,190 --> 00:11:47,660
and it’s why the Toyota method works so well.

183
00:11:48,130 --> 00:11:49,100
It certainly does.

184
00:11:49,180 --> 00:11:52,270
I guess the question, then, in terms of decision-making, in the

185
00:11:52,270 --> 00:11:55,299
world we live in today, we see so many conflicting opinions.

186
00:11:56,120 --> 00:11:59,600
How does that cooperation work to get to that initial decision that we’re

187
00:11:59,600 --> 00:12:03,360
going to make this as our first experimental approach to this problem?

188
00:12:03,370 --> 00:12:07,740
How do you bring those parties together to get to that if

189
00:12:07,740 --> 00:12:10,839
they are so—if they’re on opposite sides of the argument?

190
00:12:11,900 --> 00:12:14,820
Yeah, well, that’s a great point, of course,

191
00:12:14,860 --> 00:12:18,869
and what is needed actually is a common goal.

192
00:12:20,010 --> 00:12:22,459
A lot of this work traces to somebody, Elinor

193
00:12:22,710 --> 00:12:25,490
Ostrom, who should be familiar to everyone.

194
00:12:26,070 --> 00:12:30,321
If you’re not already familiar with Elinor Ostrom, then please do [laugh]

195
00:12:30,321 --> 00:12:30,747
.
Yeah, right [laugh]

196
00:12:30,889 --> 00:12:35,450
.
And she outlined eight core design principles for groups to

197
00:12:35,450 --> 00:12:39,420
function well, and my work is very much based on her work.

198
00:12:39,420 --> 00:12:41,220
And the first core design principle, it’s

199
00:12:41,220 --> 00:12:43,050
a strong sense of identity and purpose.

200
00:12:43,050 --> 00:12:47,060
In other words, we must see ourselves as the members

201
00:12:47,099 --> 00:12:50,110
of a group with a common purpose, what we’re about.

202
00:12:50,110 --> 00:12:53,689
And if we’re not, well, that’s just, like… a recipe for conflict.

203
00:12:53,730 --> 00:12:57,520
So, that means that we need to find a common goal.

204
00:12:57,520 --> 00:13:00,599
At a small-scale, that happens all the time.

205
00:13:00,680 --> 00:13:03,670
You know, it might be our business or a neighborhood or something.

206
00:13:03,670 --> 00:13:06,770
So, I think that there’s many, many cases in which groups

207
00:13:06,770 --> 00:13:10,660
do have a strong sense of purpose, and then there’s more

208
00:13:10,660 --> 00:13:13,330
roles that are required, but at least they have that.

209
00:13:13,350 --> 00:13:17,690
And ultimately, there’s a lot to convey here, Alan, in a short space,

210
00:13:17,700 --> 00:13:20,180
so I hope that I’m not, you know, seeming like a firehose, here.

211
00:13:20,180 --> 00:13:23,410
But I think what that ultimately means, what we need to

212
00:13:23,410 --> 00:13:26,730
work towards is a global identity, and that, the first

213
00:13:26,730 --> 00:13:30,230
and foremost, I’m a human being and citizen of the planet.

214
00:13:30,260 --> 00:13:32,290
And then I have many other identities.

215
00:13:32,470 --> 00:13:37,310
And those identities are important—my nationality, my religion, my town, up

216
00:13:37,310 --> 00:13:43,010
and down, different identities—but those identities will always exist, in some

217
00:13:43,010 --> 00:13:47,469
form, but they need to be coordinated with the planetary good in mind, I think.

218
00:13:47,469 --> 00:13:50,310
And so, that’s very important to work towards.

219
00:13:50,310 --> 00:13:53,660
And I don’t think it’s too idealistic to say that that’s possible.

220
00:13:53,670 --> 00:13:56,550
It’s already a reality for many people, and I

221
00:13:56,570 --> 00:14:00,000
think needs to become a reality for everyone.

222
00:14:00,270 --> 00:14:02,490
Also, I know we’re spending a lot of time on

223
00:14:02,490 --> 00:14:04,710
this question, but I think it’s so fascinating.

224
00:14:04,959 --> 00:14:08,970
One more question for you, in terms of those groups coming together.

225
00:14:09,320 --> 00:14:10,910
How is failure handled?

226
00:14:10,920 --> 00:14:13,970
So, you see, in many worlds, if you think, well, this is an

227
00:14:13,980 --> 00:14:18,860
innovative solution to this problem and it fails, that experiment—the

228
00:14:18,860 --> 00:14:22,210
first one—doesn’t work, then often you’ll pay a price for

229
00:14:22,210 --> 00:14:24,920
that, right, you know, your job or a demotion or something.

230
00:14:25,030 --> 00:14:28,610
How can that be handled better within this cooperation we’re talking about?

231
00:14:29,150 --> 00:14:32,400
Yeah, from an evolutionary perspective, failure is the leading edge of

232
00:14:32,400 --> 00:14:37,020
adaptation, and so it really does require an attitude towards failure

233
00:14:37,020 --> 00:14:41,400
of a constructive attitude towards failure is where we, you know,

234
00:14:44,360 --> 00:14:47,370
we have a goal, there’s alternatives, we’re going to try

235
00:14:47,370 --> 00:14:50,360
something out, and then we see how it works, and if it

236
00:14:50,360 --> 00:14:53,130
fails, we’ve learned something from it, for heaven’s sakes.

237
00:14:53,639 --> 00:14:54,759
We’ve learned something.

238
00:14:54,759 --> 00:14:57,390
And then we embrace the alternative, and

239
00:14:57,390 --> 00:15:00,510
then we repeat again, and again, and again.

240
00:15:00,510 --> 00:15:03,280
So, I think that’s really a wonderful attitude to

241
00:15:03,280 --> 00:15:07,040
have towards failure, a positive attitude to have.

242
00:15:07,040 --> 00:15:10,300
And just to, like, you know, you know, let’s fail again, and again, and again.

243
00:15:11,130 --> 00:15:12,150
But safely, of course.

244
00:15:12,150 --> 00:15:14,159
You can’t fail in a catastrophic way.

245
00:15:14,400 --> 00:15:17,189
Just like the Toyota assembly plant, when we try something

246
00:15:17,190 --> 00:15:22,610
out, we need to do it on a pilot basis and then expand.

247
00:15:22,610 --> 00:15:26,400
And we need to look at natural variation.

248
00:15:26,490 --> 00:15:32,280
I think that’s one reason that America—or any Federalist Society—works well:

249
00:15:32,299 --> 00:15:36,150
there’s so many different experiments, there’s 50 states, there’s hundreds

250
00:15:36,150 --> 00:15:43,090
of cities, there’s—and look around, and you’ll find stuff that works.

251
00:15:44,400 --> 00:15:47,300
And then, of course, there’s what theory might tell us to do.

252
00:15:47,320 --> 00:15:51,350
So, the variation comes in a number of forms.

253
00:15:51,350 --> 00:15:55,130
One is natural variation, the other is the variation when we think

254
00:15:55,130 --> 00:15:58,589
about things carefully with our theories, and do very carefully

255
00:15:58,630 --> 00:16:02,780
controlled experiments, and they all feed into this process.

256
00:16:04,110 --> 00:16:07,089
Another topic we think a lot about it at the AMSE Foundation is

257
00:16:07,090 --> 00:16:10,770
education, and you certainly talk about education in your writings.

258
00:16:11,190 --> 00:16:15,100
How can this approach be applied to how we educate our children?

259
00:16:15,950 --> 00:16:17,839
That’s a great topic.

260
00:16:17,910 --> 00:16:21,060
And let’s just take note that we could ask that question, let’s

261
00:16:21,060 --> 00:16:25,360
say—and I hope we will—for economics, and the fact that we can do

262
00:16:25,360 --> 00:16:30,590
that, that we have a framework that’s equally relevant to any subject.

263
00:16:30,879 --> 00:16:33,120
We might talk about religion, I’m not sure, but

264
00:16:33,130 --> 00:16:36,380
whatever we talk about, will be covered by this.

265
00:16:36,870 --> 00:16:41,829
And the economics education comparison, they share something in common,

266
00:16:42,330 --> 00:16:48,109
which is the current paradigm, placing such an emphasis on individual-level

267
00:16:48,280 --> 00:16:51,730
competition, the competition, it gets us a little bit back to Social

268
00:16:52,120 --> 00:16:56,200
Darwinism, and the idea that competition is always good, and that if

269
00:16:56,250 --> 00:16:59,899
we want progress, then we just pit individuals against each other.

270
00:17:00,020 --> 00:17:04,899
And that’s as much true in school as it is in many business

271
00:17:05,099 --> 00:17:08,480
corporations, where as a student, you work as an individual.

272
00:17:08,480 --> 00:17:12,060
I know you work in teams, but you get an individual grade, and whether you get

273
00:17:12,060 --> 00:17:16,210
into college depends on what their grade might be, and so on, and so forth.

274
00:17:16,219 --> 00:17:20,960
So, it’s an individual level competition treadmill in business and in

275
00:17:20,960 --> 00:17:26,389
school, and it turns out that it’s toxic because it’s basically lower level

276
00:17:26,389 --> 00:17:30,980
of competition when in fact what we want our various forms of well-managed

277
00:17:30,980 --> 00:17:36,050
competition at the group level: people succeeding as teams and so on.

278
00:17:36,050 --> 00:17:37,000
So, there’s one.

279
00:17:37,559 --> 00:17:38,350
That’s just one.

280
00:17:38,350 --> 00:17:42,810
But another one is, naively we think that we want our

281
00:17:42,810 --> 00:17:46,370
children to develop as fast as possible, and so we push

282
00:17:46,370 --> 00:17:50,460
formal education on them sooner than then we should.

283
00:17:50,460 --> 00:17:54,009
And it turns out that just growing up, you might say, in the

284
00:17:54,009 --> 00:17:58,219
absence of education, such things as play—our parents, and

285
00:17:58,219 --> 00:18:00,740
maybe as children, we were lucky enough where your mom would

286
00:18:00,740 --> 00:18:02,899
say, “Go out and play and don’t come back until dinner.”

287
00:18:02,900 --> 00:18:03,830
[laugh] . Yes.

288
00:18:03,830 --> 00:18:05,870
And you go out, and what do you do?

289
00:18:05,880 --> 00:18:09,159
You meet up with your friends, those friends are mixed age groups.

290
00:18:09,770 --> 00:18:11,950
Playing is so interesting.

291
00:18:11,950 --> 00:18:14,849
It’s much more sophisticated than we think.

292
00:18:14,920 --> 00:18:20,120
And so, when we push our young children into academic education

293
00:18:20,120 --> 00:18:24,910
too early, there’s a lot of research on this—we are stunting them,

294
00:18:24,910 --> 00:18:29,070
my friends, and we’re stunting them socially, and intellectually.

295
00:18:29,070 --> 00:18:32,020
What these studies find is that these academic programs for

296
00:18:32,020 --> 00:18:35,880
young children, they have very short-term benefits, small

297
00:18:35,880 --> 00:18:39,410
and short term benefits that wash out almost immediately.

298
00:18:39,460 --> 00:18:45,409
And then later on, we have deficits, sometimes extreme deficits.

299
00:18:45,410 --> 00:18:51,240
And so, the whole concept of education really needs to be reformulated

300
00:18:51,360 --> 00:18:53,880
with human development, what we know about human development.

301
00:18:53,880 --> 00:18:55,090
And it’s huge.

302
00:18:55,540 --> 00:18:57,199
It’s huge, and it’s tragic.

303
00:18:57,700 --> 00:19:00,320
When we talk about the anxiety epidemic in

304
00:19:00,320 --> 00:19:03,700
children, we blame it too much on social media.

305
00:19:03,770 --> 00:19:07,600
I mean, social media, of course, is a big factor, but there’s

306
00:19:07,620 --> 00:19:10,849
other factors involved, and seeing it from a evolutionary

307
00:19:10,850 --> 00:19:16,639
perspective leads to a much, much more humane form of education.

308
00:19:16,640 --> 00:19:19,190
We’re starting at the earliest age.

309
00:19:20,160 --> 00:19:24,649
I must say, as a student, pre-social media student, I had my fair

310
00:19:24,650 --> 00:19:29,019
share of anxiety, so I did not depend upon social media for that.

311
00:19:29,320 --> 00:19:31,630
I will say, I hope you can visit us in Oak Ridge.

312
00:19:31,720 --> 00:19:34,200
When you look at our high school here, for example, I think they’re

313
00:19:34,200 --> 00:19:39,390
leading the way in that team approach, really, truly phenomenal teachers

314
00:19:39,390 --> 00:19:42,669
and students over there, looking at kind of a new way of doing things.

315
00:19:42,849 --> 00:19:45,790
Have you seen other best practices elsewhere that

316
00:19:45,790 --> 00:19:47,879
we should be looking at for our education system?

317
00:19:48,740 --> 00:19:51,830
Well, I’ll mention two: one is called the Good Behavior Game.

318
00:19:51,940 --> 00:19:54,570
And what that involves—this is played at a very young age,

319
00:19:54,570 --> 00:19:58,480
like, kindergarten and first grade—you get the class together,

320
00:19:58,480 --> 00:20:01,780
you ask them to nominate what counts as good behavior.

321
00:20:01,910 --> 00:20:05,149
And even though—even kids that are, you know,

322
00:20:05,150 --> 00:20:07,270
Johnny the Badass knows what it means to be good.

323
00:20:07,920 --> 00:20:12,640
And so, what they come up with is very, very similar to what the teacher would

324
00:20:12,640 --> 00:20:16,089
have said, but the fact that they came up with it is very, very important.

325
00:20:16,800 --> 00:20:20,540
And so, this gets listed and is put everywhere on the wall and stuff like that.

326
00:20:20,870 --> 00:20:23,500
And then the class gets broken into small groups who

327
00:20:23,529 --> 00:20:27,189
compete to be good while they’re doing their schoolwork.

328
00:20:27,190 --> 00:20:28,980
So, they’re doing the same schoolwork anyone

329
00:20:28,980 --> 00:20:31,960
else might do, but they’re doing it in groups.

330
00:20:32,440 --> 00:20:34,910
And if they can manage to do it without too

331
00:20:34,910 --> 00:20:37,280
many misbehaviors, then they get a small prize.

332
00:20:37,809 --> 00:20:42,510
And so, now you’re basically, the kids are competing to be

333
00:20:42,510 --> 00:20:46,100
good, and if anyone within one of these little groups acts

334
00:20:46,100 --> 00:20:48,800
up, then the other kids are saying, “No, no, don’t do that.”

335
00:20:48,800 --> 00:20:49,090
[laugh] . Right.

336
00:20:49,090 --> 00:20:55,319
And just it turns within-group competition into cooperation.

337
00:20:55,380 --> 00:20:59,409
Now, this has been played—very good longitudinal studies have been done

338
00:20:59,410 --> 00:21:05,360
on this, and when this is played for a single year—that’s it, a year—in

339
00:21:05,360 --> 00:21:08,610
either the first or second grade, and then these kids had been followed

340
00:21:08,720 --> 00:21:13,080
through adulthood—the study is still in progress—their following the

341
00:21:13,080 --> 00:21:19,069
results of this one-year intervention, and is just so many benefits.

342
00:21:19,190 --> 00:21:21,830
It’s like money in the bag, and so—in terms of earning

343
00:21:21,830 --> 00:21:25,435
compound—there’s, like, 70-to-1 ratio in terms of the

344
00:21:25,590 --> 00:21:29,560
social benefits compared to the cost of the program.

345
00:21:29,560 --> 00:21:30,810
And the program is spreading.

346
00:21:31,349 --> 00:21:34,889
Dennis Embry is the person who’s known for this.

347
00:21:34,889 --> 00:21:38,070
And then the other example is a study, one of the first studies

348
00:21:38,070 --> 00:21:42,270
I did in Binghamton, where I had an opportunity to help design a

349
00:21:42,270 --> 00:21:45,970
school for at-risk students, and to make a long story short, these

350
00:21:45,970 --> 00:21:49,039
students had flunked three or more of their courses the previous

351
00:21:49,040 --> 00:21:52,399
years, so they were on their way to being dropouts, for sure.

352
00:21:52,529 --> 00:21:55,919
And what we were able to do in the class was to build

353
00:21:55,920 --> 00:21:58,700
in these core design principles that I’m talking about.

354
00:21:58,700 --> 00:22:01,810
We didn’t teach them; we built an end to the social environment.

355
00:22:01,810 --> 00:22:03,370
That’s an important point.

356
00:22:03,370 --> 00:22:08,550
And so, much depends on psychological safety and other protections, basically.

357
00:22:08,559 --> 00:22:12,629
So again, to make a long story short, without many resources at

358
00:22:13,040 --> 00:22:17,530
our disposal, we created what we thought was the ideal, or the best

359
00:22:17,530 --> 00:22:21,660
social environment we could, based on these evolutionary principles.

360
00:22:21,700 --> 00:22:24,630
And not only did these students perform better than

361
00:22:24,630 --> 00:22:27,529
a comparison group in a randomized control trial, but

362
00:22:27,530 --> 00:22:30,490
they did as well as the average student in the system.

363
00:22:30,550 --> 00:22:33,800
And then not only that, the main benefit was

364
00:22:33,800 --> 00:22:36,179
achieved during the first quarter of the year.

365
00:22:36,600 --> 00:22:40,300
And you got to ask the question, isn’t this too good to be true?

366
00:22:41,219 --> 00:22:45,769
And the answer is, I mean, the way to make sense of it is, is that

367
00:22:46,059 --> 00:22:50,100
all of us are like any sensible turtle or snail; we know how to

368
00:22:50,100 --> 00:22:53,780
pull into our shells when we’re dangerous and insecure environment.

369
00:22:53,780 --> 00:22:55,629
And we also know how to come out of our shell

370
00:22:55,720 --> 00:22:58,539
when we’re in a safe and secure environment.

371
00:22:58,550 --> 00:23:00,929
We’d like to be prosocial, we’d like to be

372
00:23:01,500 --> 00:23:04,470
cooperative, as long as you don’t get hammered for it.

373
00:23:04,470 --> 00:23:08,940
And so, the students, despite very hard lives—which we couldn’t change

374
00:23:09,210 --> 00:23:14,970
outside of school—in school, by providing this environment, they looked

375
00:23:14,970 --> 00:23:19,620
around, and within the first quarter, they came out of their shells.

376
00:23:19,889 --> 00:23:20,830
That’s what happened.

377
00:23:21,469 --> 00:23:22,770
And they were very resilient.

378
00:23:23,090 --> 00:23:27,750
You said in terms of post-school, going into adulthood,

379
00:23:27,850 --> 00:23:30,860
these programs still continue to show benefits, even when

380
00:23:31,460 --> 00:23:34,050
they go out into a world that often is uncooperative?

381
00:23:34,679 --> 00:23:35,960
In many cases, yes.

382
00:23:35,960 --> 00:23:36,250
Yes.

383
00:23:36,250 --> 00:23:37,935
I mean, there’s a lot to say about that.

384
00:23:38,080 --> 00:23:41,159
In one sense, you think I mean, it does seem contradictory.

385
00:23:41,199 --> 00:23:43,100
Let me note the apparent contradiction.

386
00:23:43,100 --> 00:23:46,350
In one sense, we’re saying that there were—we’re

387
00:23:46,350 --> 00:23:48,850
like turtles, we pull in and out of our shells.

388
00:23:48,850 --> 00:23:53,315
As soon as you leave that social environment, back into our shells, we go.

389
00:23:53,660 --> 00:23:56,729
But in another sense, you can do an

390
00:23:56,730 --> 00:23:59,329
intervention and it can have permanent effects.

391
00:23:59,330 --> 00:24:00,879
Both of those things are true.

392
00:24:00,920 --> 00:24:04,100
I’m not sure how much we’ll be able to unpack it.

393
00:24:04,480 --> 00:24:09,600
In fact, there’s extraordinary studies—this is very science documented—there’s

394
00:24:09,700 --> 00:24:14,110
numerous studies that show, do a comparison pre, post, and follow up.

395
00:24:14,110 --> 00:24:16,419
So, pre and post, you hope for an improvement

396
00:24:16,440 --> 00:24:18,449
because that’s what the experiment was about.

397
00:24:19,300 --> 00:24:21,360
The post follow up is really interesting

398
00:24:21,360 --> 00:24:23,870
because nothing took place during that period.

399
00:24:23,900 --> 00:24:25,310
The experiment was over.

400
00:24:25,370 --> 00:24:28,700
And a lot of these studies, they showed not only an improvement between

401
00:24:28,700 --> 00:24:32,920
pre and post, but a further improvement between post and follow up.

402
00:24:33,009 --> 00:24:37,790
And what that means is, is that you did something [laugh] that was permanent.

403
00:24:38,490 --> 00:24:39,650
They’re different now.

404
00:24:40,100 --> 00:24:43,680
They’re thinking in a way that’s different, and that’s causing them to

405
00:24:43,680 --> 00:24:47,719
act in a way—and then they’re further improving upon it all by themselves.

406
00:24:48,030 --> 00:24:51,409
So, that’s the best possible outcome, and often that’s the outcome you get.

407
00:24:51,940 --> 00:24:55,579
You make a note that scientists, your fellow scientists

408
00:24:55,580 --> 00:24:59,190
across the spectrum of disciplines should, and really

409
00:24:59,200 --> 00:25:01,950
must, embrace the concept of evolution in the work.

410
00:25:01,960 --> 00:25:03,100
Why do you say that?

411
00:25:03,820 --> 00:25:06,710
Well, the first point to make there is that if you were to say that

412
00:25:06,710 --> 00:25:09,390
to a biologist, they’d just look at you and say, “Well, of course.”

413
00:25:09,390 --> 00:25:09,859
Of course.

414
00:25:09,860 --> 00:25:10,159
Right?

415
00:25:10,160 --> 00:25:10,880
Right [laugh]

416
00:25:10,880 --> 00:25:11,160
.
“Stupid.”

417
00:25:13,180 --> 00:25:16,400
“Nothing in biology makes sense, except in the light of evolution.” That…

418
00:25:17,600 --> 00:25:22,770
mantra, that statement was spoken in the 1970s by a geneticist named Theodosius.

419
00:25:23,700 --> 00:25:27,300
In the 1970s, you could say, “There’s something about this theory

420
00:25:27,300 --> 00:25:31,330
that just covers everything,” and so that’s commonplace in biology.

421
00:25:31,330 --> 00:25:35,330
The only thing that’s new is to extend that to all things human.

422
00:25:35,340 --> 00:25:37,210
So, nothing about religion, nothing about

423
00:25:37,230 --> 00:25:39,270
economics, nothing about education, and so on.

424
00:25:39,370 --> 00:25:43,230
And that takes us back to this constriction that I talked about, that

425
00:25:43,230 --> 00:25:47,889
evolution was confined to genetic evolution all the way up to the ’70s, so when

426
00:25:48,400 --> 00:25:52,580
Dobzhansky made that statement, he was only referring to genetic evolution.

427
00:25:52,870 --> 00:25:56,829
And now the prospect is—which is the topic of my book, This View of Life:

428
00:25:57,179 --> 00:26:02,519
Completing the Darwinian Revolution—is that way of thinking middle, they

429
00:26:02,520 --> 00:26:08,630
can be expanded to everything that matters to us, the entire human realm.

430
00:26:08,670 --> 00:26:12,150
And that’s very new, only within the last 50 years at most.

431
00:26:12,470 --> 00:26:16,740
One of those, for example, being human is the importance of faith and religion.

432
00:26:16,929 --> 00:26:21,560
So, how do you see that applying to that part of human life?

433
00:26:22,449 --> 00:26:28,520
I think that once we see that cooperation is the signature

434
00:26:28,520 --> 00:26:32,130
adaptation of our species, that we do so well to live in cooperative

435
00:26:32,130 --> 00:26:35,820
groups, those groups, by definition, are larger than ourselves.

436
00:26:35,820 --> 00:26:39,630
And so, the idea that we are part of something larger than ourselves,

437
00:26:39,630 --> 00:26:43,629
which is the heart of the religious and spiritual sentiment, all of the

438
00:26:43,630 --> 00:26:49,300
keywords associated with religion—worship, sacred, and the commandments—all

439
00:26:49,300 --> 00:26:55,830
of these can be seen as adaptations for highly cooperative groups.

440
00:26:55,830 --> 00:26:58,090
And so—and that’s not new.

441
00:26:58,160 --> 00:27:00,439
Émile Durkheim said that.

442
00:27:00,439 --> 00:27:05,230
It’s fascinating that there’s two major definitions of religion: one is centered

443
00:27:05,230 --> 00:27:10,820
on supernatural agents—you know, religion is a belief in gods, and that we call

444
00:27:11,130 --> 00:27:15,969
supernatural, they intervene in the natural order, so that’s one definition—the

445
00:27:16,209 --> 00:27:20,680
other definition of religion is a set of beliefs centered around the sacred

446
00:27:20,680 --> 00:27:25,060
that unite into one moral community called a church all who adhere to them.

447
00:27:25,410 --> 00:27:28,720
That’s close to Durkheim’s definition of religion.

448
00:27:29,070 --> 00:27:33,490
It focuses on moral communities, the concept of sacredness,

449
00:27:33,809 --> 00:27:37,259
and doesn’t say a thing about supernatural agency.

450
00:27:37,290 --> 00:27:42,330
And so, when you look at religions, what you find is—and I think that this is

451
00:27:42,390 --> 00:27:47,270
a wonderful thing to say about religions—they’re amazing at forming community.

452
00:27:47,440 --> 00:27:47,920
Yes.

453
00:27:48,349 --> 00:27:49,139
They’re amazing.

454
00:27:49,139 --> 00:27:51,159
And wouldn’t we want that to be the case?

455
00:27:52,510 --> 00:27:57,649
But then they’re also, from a scientific standpoint, a human construction.

456
00:27:58,580 --> 00:27:59,730
A human construction.

457
00:27:59,730 --> 00:28:04,559
And so, to think that all of the beliefs, that basically this great bushy

458
00:28:04,559 --> 00:28:09,429
tree of religion, and the great bushy tree of culture in general, can be

459
00:28:09,509 --> 00:28:13,250
studied in the same way as the great bushy Tree of Life, of biological

460
00:28:13,470 --> 00:28:17,939
life, and that the most successful religions are the ones that, in the

461
00:28:17,940 --> 00:28:21,320
first place, endow individuals with a very strong sense of purpose.

462
00:28:21,330 --> 00:28:25,359
So, you get up in the morning brimming with purpose, and the religion

463
00:28:25,370 --> 00:28:31,350
moves you to do the right things, to basically to exist in community.

464
00:28:31,540 --> 00:28:34,149
And historically, of course, those religious communities

465
00:28:34,170 --> 00:28:38,130
were often—not always—pitted against other communities.

466
00:28:38,130 --> 00:28:43,130
And so, the dark side of religion—and religion does have a dark side; do I need

467
00:28:43,130 --> 00:28:49,200
to tell anyone that?—can be explained along with the bright side of religion.

468
00:28:49,200 --> 00:28:54,520
And it also points the way to the future, that just as the scale of human

469
00:28:54,520 --> 00:28:58,519
society, including the religions, the great religions, has been increasing.

470
00:28:58,920 --> 00:29:02,050
So, if you look, for example, at the Axial Age, what we call the

471
00:29:02,200 --> 00:29:07,210
Axial Age Religions, the major religions today, they bind together,

472
00:29:07,210 --> 00:29:12,310
they form a social glue for tens and hundreds of millions of people.

473
00:29:12,310 --> 00:29:15,679
That only happened over the course of human history.

474
00:29:16,040 --> 00:29:19,709
And what we need to do is to extend that further so that

475
00:29:19,719 --> 00:29:24,520
religion expands to, “Our common home,” as Pope Francis put

476
00:29:24,520 --> 00:29:28,240
it, “Beyond religion,” as His Holiness the Dalai Lama put it.

477
00:29:28,520 --> 00:29:31,850
And his message is that there’s 8 billion people in this

478
00:29:31,850 --> 00:29:35,240
world, and they will never belong to a single religion.

479
00:29:35,450 --> 00:29:36,340
Never ever.

480
00:29:37,000 --> 00:29:40,250
And so therefore, we need some way of thinking

481
00:29:40,480 --> 00:29:43,810
that can be applied on a worldwide basis.

482
00:29:43,810 --> 00:29:45,600
That’s why he could be such good friends with

483
00:29:45,600 --> 00:29:48,860
Desmond Tutu, a Buddhist and a Christian.

484
00:29:49,190 --> 00:29:52,030
And David Bohm, a particle physicist.

485
00:29:52,230 --> 00:29:58,360
So, science and more than one variety of religion can somehow

486
00:29:58,369 --> 00:30:02,120
merge to the worldwide scale, and that’s back to the need to

487
00:30:02,120 --> 00:30:06,139
be, first and foremost, a human being and citizen of the earth.

488
00:30:06,410 --> 00:30:09,790
Everything that’s trending in that direction, and any religion

489
00:30:09,790 --> 00:30:13,060
that resists that, that says it’s only—you know, we’re the

490
00:30:13,060 --> 00:30:16,279
only way to go to heaven, well, you know that’s out of step.

491
00:30:16,279 --> 00:30:19,110
You know that is, and should be, antiquated.

492
00:30:19,219 --> 00:30:20,750
I mean, do you think that also?

493
00:30:20,830 --> 00:30:22,000
I mean, I do [laugh]

494
00:30:22,420 --> 00:30:25,890
.
I know there are so many—they always say, don’t discuss

495
00:30:25,900 --> 00:30:29,599
politics and religion, the bar, but, I mean, there are so

496
00:30:29,599 --> 00:30:32,840
many good things over history that have come from religion.

497
00:30:32,840 --> 00:30:35,710
I was brought up in the Disciples of Christ Christian

498
00:30:35,730 --> 00:30:37,282
church, in Millersburg, Kentucky, thank you very much.

499
00:30:37,282 --> 00:30:40,060
But they are also, obviously, as you said,

500
00:30:40,060 --> 00:30:42,560
those dark moments looking through time.

501
00:30:42,560 --> 00:30:44,949
And I think it really goes back to, as you say,

502
00:30:45,460 --> 00:30:48,350
our common humanity, how we treat one another.

503
00:30:48,780 --> 00:30:51,330
And you have to judge I think any religion

504
00:30:51,339 --> 00:30:53,399
of how you treat your fellow man and woman.

505
00:30:53,440 --> 00:30:54,239
That’s the bottom line.

506
00:30:54,539 --> 00:30:55,620
That’s the bottom line.

507
00:30:56,220 --> 00:30:59,659
I’m going to—you’ve mentioned the term prosocial a few times.

508
00:30:59,730 --> 00:31:01,370
And easy for me to say today, David.

509
00:31:02,130 --> 00:31:05,780
What is the ProSocial World Initiative and the New Paradigm Coalition?

510
00:31:05,780 --> 00:31:07,729
I know when Guru first introduced this, he

511
00:31:07,730 --> 00:31:10,190
talked about this New Paradigm Coalition.

512
00:31:10,190 --> 00:31:11,969
Can you tell us in general what those are?

513
00:31:12,410 --> 00:31:15,130
Well, prosocial is a word which I think it

514
00:31:15,130 --> 00:31:17,649
speaks for itself, just the way antisocial does.

515
00:31:17,790 --> 00:31:18,255
Yeah, yeah.

516
00:31:18,420 --> 00:31:22,049
But it’s not a commonly used word until recently.

517
00:31:22,360 --> 00:31:24,109
It’s now becoming more commonly used.

518
00:31:24,190 --> 00:31:26,290
You know, there’s Google Ngram, where you can look at

519
00:31:26,290 --> 00:31:30,900
the frequency of words from Google’s vast trove of text.

520
00:31:30,900 --> 00:31:34,220
And if you look at Google Ngram for the two words

521
00:31:34,220 --> 00:31:37,210
‘antisocial’ and ‘prosocial,’ there is a period of decades

522
00:31:37,210 --> 00:31:40,090
in which prosocial was a word that was almost never used.

523
00:31:40,160 --> 00:31:42,170
Of course, antisocial has always been a common word.

524
00:31:42,460 --> 00:31:45,180
Now, prosocial is catching up, I’m glad to say.

525
00:31:45,690 --> 00:31:45,860
And—

526
00:31:46,960 --> 00:31:47,020
[laugh]

527
00:31:47,020 --> 00:31:48,310
.
Yeah, how about that?

528
00:31:48,310 --> 00:31:48,410
Right.

529
00:31:48,670 --> 00:31:51,530
So, what it means is basically anything oriented towards

530
00:31:51,530 --> 00:31:55,090
the welfare of others or society as a whole is prosocial.

531
00:31:55,610 --> 00:31:57,510
It’s a more general word than altruism.

532
00:31:57,510 --> 00:31:58,790
Altruism is like that.

533
00:31:58,800 --> 00:32:03,250
Altruism means other, but altruism for many people has the connotation of

534
00:32:04,330 --> 00:32:09,480
self-sacrifice, that in order to benefit others requires self-sacrifice.

535
00:32:09,540 --> 00:32:13,350
And prosocial is more agnostic about that.

536
00:32:13,350 --> 00:32:15,970
Sometimes that’s true, but sometimes there’s

537
00:32:15,970 --> 00:32:19,449
win-win situations: we all can benefit from things.

538
00:32:19,450 --> 00:32:20,920
In fact, that would be better, wouldn’t it?

539
00:32:21,660 --> 00:32:26,110
And so prosocial, I think, is a lovely term for its generality.

540
00:32:26,720 --> 00:32:32,949
And ProSocial World, my nonprofit organization, is basically dedicated to

541
00:32:32,960 --> 00:32:36,980
that, and our mission is to consciously evolve a world that works for all.

542
00:32:37,020 --> 00:32:41,899
What really differentiates us is this modern evolutionary perspective.

543
00:32:41,930 --> 00:32:44,590
We might be the only organization that’s really just

544
00:32:44,590 --> 00:32:49,990
at the forefront of Darwinian, generalized Darwinism.

545
00:32:50,070 --> 00:32:52,200
Who’s heard the term generalized Darwinism?

546
00:32:52,210 --> 00:32:52,860
Nobody.

547
00:32:52,910 --> 00:32:52,960
[laugh]

548
00:32:53,040 --> 00:32:56,070
.
So, I’m sure that all your listeners—if there’s a

549
00:32:56,070 --> 00:32:58,809
listener out there that’s heard that term, write to me.

550
00:33:01,010 --> 00:33:02,150
That’s how new it is.

551
00:33:02,150 --> 00:33:05,660
And the New Paradigm Coalition applies all of this to economics.

552
00:33:06,100 --> 00:33:09,010
And when we do that—economics and business—and when

553
00:33:09,010 --> 00:33:12,210
we do that, we confront an elephant in the room.

554
00:33:12,240 --> 00:33:15,599
The elephant is neoclassical economics, the dominant

555
00:33:15,870 --> 00:33:20,700
tradition in economics is called neoclassical economics, and

556
00:33:20,700 --> 00:33:24,800
it goes all the way back to Newton and Newtonian physics.

557
00:33:24,800 --> 00:33:29,759
And back in the 19th century, of course economists, that was the new topic.

558
00:33:30,270 --> 00:33:32,870
It used to be called natural and moral philosophy.

559
00:33:32,940 --> 00:33:36,879
But people studying economics, they wanted to be scientific, and back

560
00:33:36,880 --> 00:33:41,760
then that meant emulating Newton and his planetary laws of motion.

561
00:33:41,760 --> 00:33:45,080
And so, they would say, “Oh, yeah, we want a physics of social behavior.

562
00:33:45,190 --> 00:33:47,509
That’s where we want.” And that’s whatthey set about.

563
00:33:47,800 --> 00:33:49,230
It was very mathematical.

564
00:33:49,700 --> 00:33:54,429
It had to treat individuals as this kind of atom, and it went on from there.

565
00:33:54,710 --> 00:34:00,460
The New Paradigm Coalition, so on the one hand, we have this evolution paradigm.

566
00:34:00,680 --> 00:34:03,360
We haven’t said much about complexity yet, Alan, and I’m not

567
00:34:03,360 --> 00:34:07,880
sure that we will, purely for lack of time, but of course, your

568
00:34:07,889 --> 00:34:11,960
museum is all about—and you know, science—is all about complexity.

569
00:34:11,980 --> 00:34:14,509
So yeah, just to mark it.

570
00:34:14,889 --> 00:34:17,390
Two bodies of knowledge new within the last 50 years.

571
00:34:17,390 --> 00:34:20,230
One thing I like to say is that the last 50 years

572
00:34:20,230 --> 00:34:23,009
is more than just 10% of the scientific revolution.

573
00:34:23,050 --> 00:34:26,360
Two bodies of knowledge that are foundational and new.

574
00:34:26,360 --> 00:34:29,349
One is complex systems science, and it’s new because

575
00:34:29,350 --> 00:34:31,699
it required the advent of widespread computing.

576
00:34:32,139 --> 00:34:36,000
So, before—the Santa Fe Institute, for example, was founded in

577
00:34:36,000 --> 00:34:41,730
1984—so before that, you could talk about complexity, but you couldn’t

578
00:34:41,739 --> 00:34:45,770
model it because it was too complex for math, analytical math.

579
00:34:45,790 --> 00:34:48,089
You had to have computers, and they weren’t

580
00:34:48,090 --> 00:34:50,620
available until the ’70s or the ’80s.

581
00:34:50,650 --> 00:34:54,359
So, everything we associate with complex systems science, new.

582
00:34:54,409 --> 00:34:55,319
Isn’t that amazing?

583
00:34:55,330 --> 00:34:55,650
Yeah.

584
00:34:56,020 --> 00:34:56,300
Yeah.

585
00:34:56,429 --> 00:34:57,620
It’s amazing to me.

586
00:34:57,800 --> 00:34:58,740
A real revolution.

587
00:34:58,800 --> 00:34:59,090
Yeah.

588
00:34:59,750 --> 00:35:03,429
And then everything about evolution beyond genetic evolution,

589
00:35:03,469 --> 00:35:07,201
as we have said, new [laugh] since the 1970s, and ’80s.

590
00:35:07,201 --> 00:35:07,410
And so now,

591
00:35:10,700 --> 00:35:14,160
this combination, this complexity of evolution paradigm

592
00:35:14,190 --> 00:35:19,330
is what provides an alternative to neoclassical economics.

593
00:35:19,770 --> 00:35:24,319
And so, the New Paradigm Coalition is all about evolution complexity,

594
00:35:24,340 --> 00:35:32,250
but applies it specifically to economics, business, and public policy.

595
00:35:32,250 --> 00:35:38,020
And the difference is as profound as what we talked about with education.

596
00:35:38,070 --> 00:35:41,470
I mean, in some ways, just night and day.

597
00:35:41,559 --> 00:35:43,000
And let me give you one example.

598
00:35:43,710 --> 00:35:47,380
It’s a common practice in businesses called force distribution

599
00:35:48,010 --> 00:35:54,089
ranking—otherwise known as rank and yank—where a company evaluates

600
00:35:54,400 --> 00:35:57,630
all of its staff members on their performance, rewards the

601
00:35:57,630 --> 00:36:02,700
best, fires the worst, maybe does something with the middle.

602
00:36:03,150 --> 00:36:05,860
And so, when you think of this, what does that assume?

603
00:36:05,870 --> 00:36:08,520
That assumes that a performance of the individual is

604
00:36:08,520 --> 00:36:11,070
based purely on the properties of that individual.

605
00:36:11,219 --> 00:36:14,659
It accords nothing to social interactions at all, doesn’t it?

606
00:36:14,980 --> 00:36:18,710
It that some people are gold nuggets, other people are gravel, and all you have

607
00:36:18,710 --> 00:36:22,050
to do is sift them, you get the gold nuggets and then your organization works.

608
00:36:22,390 --> 00:36:25,600
Well, it accords nothing to social interactions.

609
00:36:25,600 --> 00:36:30,320
And what it causes is a complete shutdown of cooperation because in that

610
00:36:30,330 --> 00:36:33,860
system, if that’s the system that you’re locked into, and if you want to get

611
00:36:33,860 --> 00:36:38,140
ahead in that system, you never do anything to make anyone else look good.

612
00:36:39,450 --> 00:36:41,100
You’ve turned it into a hunger game.

613
00:36:41,400 --> 00:36:43,595
And that’s the way some corporations are run,

614
00:36:43,820 --> 00:36:45,960
and that’s the way some of our schools are run.

615
00:36:45,960 --> 00:36:50,509
And so, just to make the most elementary points, is that for any group to work

616
00:36:50,509 --> 00:36:54,400
well, including the business group, it has to foster cooperation, friends.

617
00:36:55,100 --> 00:36:58,009
And in order to do that—and then it just goes on from there.

618
00:36:58,070 --> 00:36:59,490
So, it’s watershed.

619
00:36:59,530 --> 00:37:00,670
I mean, it’s truly watershed.

620
00:37:00,980 --> 00:37:06,340
And that’s what the New Paradigm Coalition is intended to catalyze.

621
00:37:06,849 --> 00:37:11,639
So, on a broader scale, let’s say on a governmental scale, the New Paradigm

622
00:37:11,639 --> 00:37:16,379
Coalition, it’s certainly not founded on laissez-faire principles, but

623
00:37:16,380 --> 00:37:21,410
also you don’t call for a centralized or command structure in our economy.

624
00:37:21,730 --> 00:37:26,260
So, how does that cooperative approach work on a governmental

625
00:37:26,260 --> 00:37:30,169
level at that point, in terms of the economic future of a nation?

626
00:37:30,990 --> 00:37:33,490
Well, we already covered, I think earlier in this

627
00:37:33,490 --> 00:37:36,810
conversation, the impossibility of centralized planning.

628
00:37:37,070 --> 00:37:39,640
And so, no group of experts is capable.

629
00:37:39,660 --> 00:37:43,470
So basically, we’ve already covered the necessity of experimentation.

630
00:37:43,630 --> 00:37:46,380
To broaden that, we can say two things don’t work.

631
00:37:46,629 --> 00:37:47,890
Only one thing can work.

632
00:37:48,219 --> 00:37:50,520
One thing, centralized planning, we’ve covered that.

633
00:37:50,840 --> 00:37:56,289
The second is laissez-faire because it’s just not true that the

634
00:37:56,290 --> 00:38:00,030
lower level pursuit of self-interest is led by an invisible hand.

635
00:38:00,040 --> 00:38:00,640
No.

636
00:38:01,160 --> 00:38:02,230
No, forget about that.

637
00:38:02,759 --> 00:38:08,520
That’s the foundational message of multi-level evolution.

638
00:38:08,530 --> 00:38:11,010
Therefore, the only thing that works is experimentation.

639
00:38:11,010 --> 00:38:13,290
I think, in some ways, we have covered that.

640
00:38:13,290 --> 00:38:16,009
And that goes for a government agency or governmental

641
00:38:16,010 --> 00:38:18,650
process no less than any other kind of process.

642
00:38:18,650 --> 00:38:22,780
I actually have a series of interviews on all different topics:

643
00:38:23,059 --> 00:38:28,215
government, business, entrepreneurship, development, innovation, and in

644
00:38:28,280 --> 00:38:32,780
all cases, we can show that the only thing that works is experimentation.

645
00:38:32,980 --> 00:38:35,550
And people are pushed into that pragmatic zone.

646
00:38:35,580 --> 00:38:39,069
We have our theories, but when people are actually on the ground, just

647
00:38:39,080 --> 00:38:43,159
trying to make things work, they just get pushed into that pragmatic zone.

648
00:38:43,170 --> 00:38:45,650
They don’t have a theory for it, but it’s what they do.

649
00:38:45,840 --> 00:38:49,720
And I think providing a theory for it is—you can

650
00:38:49,720 --> 00:38:52,399
ask the question, it’s a kind of a humble question.

651
00:38:52,650 --> 00:38:55,130
“Well, if it’s what you’re already doing, then what good is the

652
00:38:55,380 --> 00:39:00,049
theory?” I mean, the theory is kind of a late arrival, isn’t it?

653
00:39:00,119 --> 00:39:03,270
And the answer to that is no, no, the theory is pretty important.

654
00:39:03,320 --> 00:39:07,190
To be able to understand this on the benefits

655
00:39:07,190 --> 00:39:09,060
of first principles is very important.

656
00:39:09,309 --> 00:39:13,720
Even if you’re a good practitioner, [man] , you want to know about the theory.

657
00:39:14,380 --> 00:39:20,960
Let’s talk, as we draw to a close here, the role in your system of cooperation.

658
00:39:21,090 --> 00:39:24,470
We have the state, you have the group… groups

659
00:39:24,480 --> 00:39:27,330
working within that state, and you have individuals.

660
00:39:27,820 --> 00:39:32,090
So, we put, I think, for good reason, a lot of focus on individual rights in

661
00:39:32,090 --> 00:39:36,070
the United States, in the western world in particular, but elsewhere, obviously.

662
00:39:36,219 --> 00:39:40,390
What’s the role and the status of the individual in a

663
00:39:40,390 --> 00:39:44,060
world that’s then built around these groups cooperating?

664
00:39:44,080 --> 00:39:46,540
If I’m an individual and I disagree with the

665
00:39:46,540 --> 00:39:49,250
group’s end result, how do I express myself?

666
00:39:49,250 --> 00:39:50,089
How am I heard?

667
00:39:51,219 --> 00:39:52,430
That’s a great question, Alan.

668
00:39:52,430 --> 00:39:55,000
I’m really happy that you focused on that.

669
00:39:55,010 --> 00:39:59,080
And I want to answer it first in the context of a smaller group because I

670
00:39:59,120 --> 00:40:02,960
think it makes so much sense in a smaller group, but then we can expand it.

671
00:40:03,290 --> 00:40:05,670
And one of the amazing virtues of these

672
00:40:05,719 --> 00:40:08,090
ideas is that they are scale independent.

673
00:40:08,160 --> 00:40:11,009
So, whatever we might say about individuals in a village,

674
00:40:11,010 --> 00:40:13,600
you might say, it also goes for nations in the global

675
00:40:13,600 --> 00:40:17,230
village, or global corporations in the global village.

676
00:40:17,230 --> 00:40:19,300
But let’s begin with a smaller group and Dr.

677
00:40:19,300 --> 00:40:20,020
Ostrom.

678
00:40:20,559 --> 00:40:23,320
What’s needed is, first of all, the group is the focus.

679
00:40:23,340 --> 00:40:27,029
We’re trying—the group—to function as a cooperative unit.

680
00:40:27,230 --> 00:40:31,769
But the members of the group have a tremendous amount of agency, and they must.

681
00:40:31,820 --> 00:40:35,060
And we’ve talked about this: decision-making must include the whole group.

682
00:40:35,110 --> 00:40:36,290
We can’t have freeloaders.

683
00:40:36,300 --> 00:40:39,770
The benefits that you get from the group must be proportional to the cost.

684
00:40:39,770 --> 00:40:43,000
We have to monitor what we do.

685
00:40:43,000 --> 00:40:45,859
If we’re not doing what we should, we need—so with all of this stuff about what

686
00:40:45,860 --> 00:40:52,470
we need to do to govern our group, and it’s the individuals who are doing that.

687
00:40:52,470 --> 00:40:56,980
I mean, so much so that actually, the concept of leadership is

688
00:40:56,980 --> 00:41:00,259
something which needs to be questioned because so often we think of

689
00:41:00,259 --> 00:41:05,279
groups as having leaders, and those leaders having a lot of power,

690
00:41:05,469 --> 00:41:08,410
but what this is telling us is that power needs to be distributed.

691
00:41:09,150 --> 00:41:12,260
And if you’re going to—and decision-making needs to be decision.

692
00:41:12,270 --> 00:41:15,280
Yes, you do need leaders, actually, but they have to be servant-leaders.

693
00:41:15,330 --> 00:41:18,170
And the concept of a servant-leader is one of the

694
00:41:18,530 --> 00:41:21,870
religious concepts, I think, that’s worth identifying.

695
00:41:21,960 --> 00:41:23,899
When I’m leading, I’m not a boss.

696
00:41:24,699 --> 00:41:27,620
I’m helping to orchestrate all of us.

697
00:41:27,620 --> 00:41:31,010
And upon not doing a good job, and somebody else should be the leader.

698
00:41:31,020 --> 00:41:34,100
And so, there you have the ca—first of all, the group

699
00:41:34,100 --> 00:41:36,759
is the focus, but the individuals are the agents.

700
00:41:37,440 --> 00:41:41,040
And so, everything that we associate in democratic

701
00:41:41,040 --> 00:41:44,299
terms with individual rights and so on is found there.

702
00:41:44,370 --> 00:41:48,730
It’s equitable, it’s participatory, it’s just beautiful,

703
00:41:48,740 --> 00:41:52,360
basically, around the standpoint of those values.

704
00:41:52,360 --> 00:41:55,680
And just a few weeks ago, I had the privilege of doing

705
00:41:55,680 --> 00:41:59,659
my own podcast with Robert Putnam of Social Capital fame.

706
00:41:59,660 --> 00:42:03,430
He’s the author of Bowling Alone, and is the best-known

707
00:42:03,800 --> 00:42:06,770
figure for—he coined the term ‘social capital.’

708
00:42:06,870 --> 00:42:08,920
He’s written an essay which pointed out that

709
00:42:10,310 --> 00:42:13,049
in 1831, there were two voyages of discovery.

710
00:42:13,590 --> 00:42:16,399
One was Charles Darwin sailing around the world on the

711
00:42:16,410 --> 00:42:20,680
Beagle, and the other was Alexei de Tocqueville, coming

712
00:42:20,680 --> 00:42:24,100
to the United States within a few months of each other.

713
00:42:24,550 --> 00:42:28,700
And Bob Putnam does a great job of weaving Tocqueville and Darwin

714
00:42:29,040 --> 00:42:33,860
together and focusing on the phrase of Tocqueville’s, “Self-interest

715
00:42:33,860 --> 00:42:38,980
rightly understood.” And Tocqueville’s theory of democracy, it made

716
00:42:38,980 --> 00:42:43,500
a distinction between individualism—which would be, like, just pure

717
00:42:43,500 --> 00:42:48,710
greed; I’m out for myself—and self-interest rightly understood, which

718
00:42:48,719 --> 00:42:52,949
is, well, I’m out for myself, but I’m doing it in the context of

719
00:42:52,949 --> 00:42:57,490
cooperating with others, ultimately, in the creation of a great nation.

720
00:42:57,560 --> 00:43:01,660
And so, Bob and I updated the concept of self-interest rightly understood.

721
00:43:01,670 --> 00:43:03,630
He channeled Tocqueville and I channeled Darwin.

722
00:43:03,639 --> 00:43:03,669
[laugh]

723
00:43:05,620 --> 00:43:11,069
.
But I think that that, to say that what’s new about all of this can

724
00:43:11,070 --> 00:43:17,080
actually be traced back and can be true to something like Tocqueville

725
00:43:17,400 --> 00:43:22,690
and the ideals of democracy, which currently are in such tatters.

726
00:43:22,750 --> 00:43:26,170
I mean, if you read some of the work on this, and you realize

727
00:43:27,080 --> 00:43:32,440
how much America is a plutocracy run for the benefit of the

728
00:43:32,440 --> 00:43:37,190
elites, it’s sad and shocking, and ultimately self-destructive.

729
00:43:37,350 --> 00:43:39,010
There’s so much that needs to be done.

730
00:43:39,330 --> 00:43:45,330
What I hope for in our civics world is that we do a better job of teaching

731
00:43:45,410 --> 00:43:49,919
our young people about how they can be engaged in the political system.

732
00:43:49,960 --> 00:43:51,170
And that’s something, when I was director

733
00:43:51,170 --> 00:43:52,690
at the Howard Baker Center, we did a lot.

734
00:43:52,710 --> 00:43:54,290
They’re doing that now at the Baker School.

735
00:43:54,290 --> 00:43:59,130
But yeah, I share your concerns about democracy

736
00:43:59,130 --> 00:44:01,639
in the world writ large today, for sure.

737
00:44:01,940 --> 00:44:04,619
And David, before we run out of time today, I want to

738
00:44:04,619 --> 00:44:07,660
make sure I at least mentioned your novel, Atlas Hugged.

739
00:44:07,660 --> 00:44:10,180
I mean, you’re in the world of fiction now, too.

740
00:44:10,190 --> 00:44:12,129
Can you tell us just a bit about Atlas Hugged?

741
00:44:12,430 --> 00:44:15,800
Well, first, my dad was a famous novelist in the 1950s.

742
00:44:15,800 --> 00:44:17,900
He got the Man in the Gray Flannel Suit and A Summer

743
00:44:17,900 --> 00:44:21,250
Place, so the idea of writing a novel was not new to me.

744
00:44:21,250 --> 00:44:23,520
But it was in the early days of economics

745
00:44:23,549 --> 00:44:26,339
when I was beginning to read about economics.

746
00:44:26,340 --> 00:44:30,340
And then somebody said in a workshop, “You know, Ayn Rand did so much

747
00:44:30,340 --> 00:44:35,260
with her book Atlas Shrugged to popularize the neoclassical paradigm.

748
00:44:35,260 --> 00:44:36,870
Shouldn’t we be doing the same?”

749
00:44:36,980 --> 00:44:40,310
In an instant, the title Atlas Hugged floated

750
00:44:40,310 --> 00:44:42,409
into my mind, and the beginning of a plot line.

751
00:44:42,490 --> 00:44:46,170
And so, over seven years, as I was actually thinking about

752
00:44:46,170 --> 00:44:50,880
economics as a scientist, I was writing my novel on the side.

753
00:44:51,000 --> 00:44:53,319
And then ultimately it got published, and

754
00:44:53,999 --> 00:44:56,070
it has made a pretty good name for itself.

755
00:44:56,070 --> 00:44:59,040
So, if you want to learn about these ideas through a story—and

756
00:44:59,040 --> 00:45:02,060
stories are so much more captivating than—even the best-reading

757
00:45:02,070 --> 00:45:05,330
nonfiction book is no match for a story, in my opinion.

758
00:45:05,800 --> 00:45:07,480
So, check out Atlas Hugged [laugh]

759
00:45:07,480 --> 00:45:08,030
.
Definitely.

760
00:45:08,030 --> 00:45:10,149
I read both of yours, and enjoyed the both

761
00:45:10,150 --> 00:45:12,470
fiction and nonfiction very much [laugh]

762
00:45:12,540 --> 00:45:12,770
.
Oh, thank you.

763
00:45:12,770 --> 00:45:13,089
Thank you very much.

764
00:45:13,089 --> 00:45:17,009
And in Atlas Hugged, a worldwide transformation takes place in a hundred days.

765
00:45:17,750 --> 00:45:21,575
So, beyond those things, what’s—not that you haven’t done [laugh]

766
00:45:21,590 --> 00:45:24,120
a lot already, David, but what are you working on right now?

767
00:45:24,120 --> 00:45:25,170
What’s next for you?

768
00:45:25,490 --> 00:45:28,390
I think I’d like to end on the concept of catalysis.

769
00:45:29,840 --> 00:45:33,880
In chemistry—you know what catalysis means in chemistry—basically,

770
00:45:33,880 --> 00:45:37,140
it’s a molecule that you can sprinkle into a chemical

771
00:45:37,150 --> 00:45:40,890
broth, and it increases the rate of reaction by orders of

772
00:45:40,890 --> 00:45:44,360
magnitude—incredible—without being used up in the process.

773
00:45:44,360 --> 00:45:47,600
And so, the idea of cultural catalysis, the idea that there’s something

774
00:45:47,600 --> 00:45:52,819
we can do culturally, to make things go faster and better in years, not

775
00:45:52,830 --> 00:45:57,500
decades or not at all, can this really take place in months and years?

776
00:45:58,300 --> 00:45:58,770
Can it?

777
00:45:59,100 --> 00:46:02,460
And if we take the concept of catalysis seriously

778
00:46:02,480 --> 00:46:06,130
for culture, then the answer would be yes, we could.

779
00:46:06,759 --> 00:46:13,229
And then of course, then if that gets planted as a genuine possibility,

780
00:46:13,250 --> 00:46:16,300
well, it highly motivates you to do it if you possibly can.

781
00:46:16,390 --> 00:46:20,230
And so, this was what causes me to be very highly motivated and to

782
00:46:20,230 --> 00:46:25,319
work through my organization and with other organizations, and to make

783
00:46:25,319 --> 00:46:29,080
this kind of positive catalysis happen as soon as it possibly can.

784
00:46:29,090 --> 00:46:31,080
And so, that provides plenty to do.

785
00:46:31,320 --> 00:46:31,940
I’m sure.

786
00:46:31,940 --> 00:46:34,370
Where can people go to learn more about your organizations?

787
00:46:34,730 --> 00:46:38,640
Well, they can go to prosocial.world is the website, and

788
00:46:38,650 --> 00:46:42,569
they can email hello@prosocial.world in order to get in

789
00:46:42,570 --> 00:46:44,970
touch with us, and then we’re very happy to work with you.

790
00:46:45,270 --> 00:46:49,729
We work on multiple fronts: education, economics, religion, spirituality.

791
00:46:49,950 --> 00:46:53,339
“Nothing about humanity makes sense except in the light of evolution.”

792
00:46:53,339 --> 00:46:57,150
And so, whatever your interest might be, we hope to provide engagement.

793
00:46:57,620 --> 00:46:59,799
And whatever your current level of knowledge might be,

794
00:46:59,800 --> 00:47:02,830
where you are on your learning curve, it doesn’t matter

795
00:47:02,920 --> 00:47:05,330
whether you’re highly educated or not, it doesn’t matter.

796
00:47:05,440 --> 00:47:09,399
Just check out hello@prosocial.world, and we’ll

797
00:47:09,400 --> 00:47:13,210
see if we could engage you in this new paradigm.

798
00:47:13,690 --> 00:47:16,230
Well, what a thought-provoking conversation, David.

799
00:47:16,250 --> 00:47:18,480
Really enjoyed having you on AMSEcast.

800
00:47:18,480 --> 00:47:19,850
Thanks so much for joining us.

801
00:47:20,240 --> 00:47:20,589
Thanks.

802
00:47:20,589 --> 00:47:22,369
Thanks very much to you, and all that you’re doing.

803
00:47:26,009 --> 00:47:28,710
Thank you for joining us on this episode of AMSEcast.

804
00:47:29,120 --> 00:47:33,440
For more information on this topic or any others, you can always visit us at

805
00:47:34,480 --> 00:47:40,079
AMSE.org or find, like, and follow us on Facebook, Instagram, and Twitter.

806
00:47:40,580 --> 00:47:42,909
I invite you to visit the American Museum of Science

807
00:47:42,910 --> 00:47:45,710
and Energy and the K-25 History Center in person.

808
00:47:46,030 --> 00:47:50,190
You can also shop at our online store and become a member at AMSE.org.

809
00:47:50,650 --> 00:47:54,089
Thanks to our production team with Matt Mullins, plus our supportive colleagues

810
00:47:54,090 --> 00:47:57,939
at the Department of Energy’s Office of Science, Office of Environmental

811
00:47:57,940 --> 00:48:01,920
Management, and Office of Legacy Management, as well as Oak Ridge National

812
00:48:01,920 --> 00:48:07,759
Laboratory, the Y-12 National Security Complex, NNSA, and the AMSE Foundation.

813
00:48:08,270 --> 00:48:10,000
And of course, thanks to our wonderful guests

814
00:48:10,000 --> 00:48:11,810
today, and to all of you for listening.

815
00:48:12,320 --> 00:48:15,089
I hope you’ll join us for the next episode of AMSEcast.

816
00:48:17,709 --> 00:48:20,690
If you’ve enjoyed this podcast, I would like to ask that you consider

817
00:48:20,690 --> 00:48:25,270
becoming a member of the 117 Society, the newest membership opportunity

818
00:48:25,320 --> 00:48:28,779
offered by the American Museum of Science and Energy Foundation.

819
00:48:29,550 --> 00:48:32,720
By joining the 117 Society, you will help us continue

820
00:48:32,720 --> 00:48:35,540
this podcast and our other innovative programming.

821
00:48:35,920 --> 00:48:38,830
You will support the expansion of our vitally important educational

822
00:48:38,830 --> 00:48:42,710
outreach, including virtual classes, and you will help ensure that both

823
00:48:42,720 --> 00:48:47,350
the American Museum of Science and Energy and the K-25 History Center can

824
00:48:47,350 --> 00:48:51,420
continue to provide world-class exhibits to our community and to the world.

825
00:48:52,260 --> 00:48:55,130
Benefits of membership includes special access to video

826
00:48:55,130 --> 00:48:59,339
and audio content, and 117 Society merchandise, as well

827
00:48:59,340 --> 00:49:02,270
as all the benefits of our Atom Splitter Membership Level.

828
00:49:02,860 --> 00:49:05,160
To learn more, go to AMSE.org.

829
00:49:06,240 --> 00:49:11,039
The 117 Society is vital to the future of AMSE and the K-25 History Center.

830
00:49:11,430 --> 00:49:14,030
I hope you will consider joining, and thank you very much.