1
00:00:00,000 --> 00:00:09,640 Welcome to episode 31 of the Language Neuroscience Podcast.
2
00:00:09,640 --> 00:00:13,840
I'm Stephen Wilson and I'm a neuroscientist at the University of Queensland in Brisbane,
3
00:00:13,840 --> 00:00:14,840 Australia.
4
00:00:14,840 --> 00:00:19,040 I have a very special guest today, Professor Masud Husain.
5
00:00:19,040 --> 00:00:23,640
He's Professor of Neurology and Cognitive Neuroscience at the University of Oxford.
6
00:00:23,640 --> 00:00:26,960
Masud is different from a lot of the guests that I've had on the podcast because he's
7
00:00:26,960 --> 00:00:28,840 not really a language and brain guy.
8
00:00:28,840 --> 00:00:34,080
He is a neuroscientist and an neurologist, but he works on topics like attention, neglect,
9
00:00:34,080 --> 00:00:38,320
the link between attention and short-term memory, motivation and apathy.
10
00:00:38,320 --> 00:00:40,480 Not the kind of topics that we normally talk about.
11
00:00:40,480 --> 00:00:43,920 And indeed, I didn't invite him to talk about his own research today.
12
00:00:43,920 --> 00:00:48,200
I invited him to talk about an editorial that he wrote for Brain, at the Journal where he
13
00:00:48,200 --> 00:00:50,680 is the editor-in-chief.
14
00:00:50,680 --> 00:00:54,920
As you guys probably know, Brain is a very important journal for our field, some of the most important
15
00:00:54,920 --> 00:00:59,120
Language and Brain papers of all time have been published in that journal, such as Ludwig
16
00:00:59,120 --> 00:01:05,480
Lichtheim's 1885 Masterpiece on the "House" - model and Norman Geschwind's beautiful two-part
17
00:01:05,480 --> 00:01:08,920 epic from 1965, many others.
18
00:01:08,920 --> 00:01:14,320
So Masud is editor-in-chief of that journal, which is a really prestigious position.
19
00:01:14,320 --> 00:01:15,920 Earlier this year, I read his editorial.
20
00:01:15,920 --> 00:01:17,600 It's called "A Mountain of Small Things."
21
00:01:17,600 --> 00:01:22,960 It's published in volume 147, issue 3, March 2024.
22
00:01:22,960 --> 00:01:28,440
And Masud argues that we have a big problem facing science, bigger even than the reproducibility
23
00:01:28,440 --> 00:01:29,520 crisis.
24
00:01:29,520 --> 00:01:33,720
So I invited him to come on the podcast and read his editorial for us, which he's going to
25
00:01:33,720 --> 00:01:35,880 do, and then we're going to have a chat about it.
26
00:01:35,880 --> 00:01:39,200 I'm recording this on the weekend before SNL in Brisbane.
27
00:01:39,200 --> 00:01:42,480 I'm hoping to see lots of the listeners of the podcast there at SNL.
28
00:01:42,480 --> 00:01:45,200 Please come and say hi if you see me.
29
00:01:45,200 --> 00:01:47,840
And I'd love to hear what you guys think about this different kind of episode, which is
30
00:01:47,840 --> 00:01:54,000
not so much a content episode, but a sort of practice of science episode.
31
00:01:54,000 --> 00:01:55,520 Very curious what you guys think.
32
00:01:55,520 --> 00:01:56,680 So let me know.
33
00:01:56,680 --> 00:02:02,000
And if you're not at the conference, feel free to send me an email, [email protected].
34
00:02:02,000 --> 00:02:03,480 And I hope you enjoy the episode. 35
00:02:03,480 --> 00:02:05,080 All right, let's get to it.
36
00:02:05,080 --> 00:02:07,080 Hi Masud, how are you?
37
00:02:07,080 --> 00:02:08,280 I'm well, thank you.
38
00:02:08,280 --> 00:02:11,560 Yeah, it's a pleasure to meet you and thanks for taking the time.
39
00:02:11,560 --> 00:02:13,120 Yeah, great for inviting me.
40
00:02:13,120 --> 00:02:14,120 Thank you.
41
00:02:14,120 --> 00:02:17,760
Before we get into it, can you tell me a little bit about yourself and just
42
00:02:17,760 --> 00:02:20,800
let our listeners know like, who you are and what you're interested in?
43
00:02:20,800 --> 00:02:26,400
Yeah, I'm a neurologist, still active, clinically, and a neuroscientist.
44
00:02:26,400 --> 00:02:30,280 And my interest has really been in sort of cognitive neurology.
45
00:02:30,280 --> 00:02:37,600
I started off in attention and prominently in people who have inattention or the neglect
46
00:02:37,600 --> 00:02:38,600 syndrome.
47
00:02:38,600 --> 00:02:42,760 That was the area of study for about 15 years or so.
48
00:02:42,760 --> 00:02:47,200 That led to work on the link between attention and short-term memory.
49
00:02:47,200 --> 00:02:51,400
And we did quite a bit of work in trying to understand the architecture of short-term memory in
50
00:02:51,400 --> 00:02:57,200
healthy people, but also started using new techniques to measure memory in different
51
00:02:57,200 --> 00:02:58,440 patient groups.
52
00:02:58,440 --> 00:03:07,280
More recently, more recently we've got into, stumbled into motivation and loss of motivation,
53
00:03:07,280 --> 00:03:13,960
syndrome called apathy, and whether there might be a neurobiological basis for that, across,
54
00:03:13,960 --> 00:03:16,360 across different patient groups.
55
00:03:16,360 --> 00:03:17,840 But also in healthy people.
56
00:03:17,840 --> 00:03:19,640 Right, okay.
57
00:03:19,640 --> 00:03:20,640 That's fascinating.
58
00:03:20,640 --> 00:03:25,080
And do you link your clinical practice and your research or do you kind of have separate
59
00:03:25,080 --> 00:03:26,560 strands there?
60
00:03:26,560 --> 00:03:27,720 There's a strong link.
61
00:03:27,720 --> 00:03:33,000
We see people in the cognitive disorders clinic who have all sorts of disorders, including
62
00:03:33,000 --> 00:03:36,160 language disorders.
63
00:03:36,160 --> 00:03:40,600
But a lot of the patients who come to our clinics also are interested in taking part in
64
00:03:40,600 --> 00:03:41,600 the research we do.
65
00:03:41,600 --> 00:03:43,640 So yeah, there's a very strong link.
66
00:03:43,640 --> 00:03:45,360 Yeah, wonderful.
67
00:03:45,360 --> 00:03:52,300
So, about your editorial, when I read it a few months ago when it came out, I was very struck
68
00:03:52,300 --> 00:03:53,300 by it. 69
00:03:53,300 --> 00:03:54,300 It really resonated.
70
00:03:54,300 --> 00:03:58,920
And I sent it to a lot of colleagues instantly, they've all kind of felt the same way.
71
00:03:58,920 --> 00:04:02,380
So I was wondering if people haven't read it, then they're not really going to get much
72
00:04:02,380 --> 00:04:03,540 out of our discussion.
73
00:04:03,540 --> 00:04:05,720 So would you be able to read it for us?
74
00:04:05,720 --> 00:04:09,160 I think it's about six paragraphs long.
75
00:04:09,160 --> 00:04:10,480 It's a very nice piece of writing.
76
00:04:10,480 --> 00:04:12,000 How do you, would you be okay with that?
77
00:04:12,000 --> 00:04:14,360 Of course, yeah, I'd be happy to do that.
78
00:04:14,360 --> 00:04:19,440
Perhaps I could just set this in a bit of context, because I'm the Editor of Brain.
79
00:04:19,440 --> 00:04:22,480 I've been doing that for the last three years.
80
00:04:22,480 --> 00:04:29,680
And of course, most editorials that we write, I write, are also about
academic work in the
81
00:04:29,680 --> 00:04:33,840 journal or related work in other journals.
82
00:04:33,840 --> 00:04:40,080
But I also think it's a responsibility of intellectuals like us and your podcast listeners
83
00:04:40,080 --> 00:04:49,760
to flag up areas where we think are actually changing in such a way that they're having a
84
00:04:49,760 --> 00:04:56,040
devastating impact on our fundamental work, whether you do language or anything else, whether
85
00:04:56,040 --> 00:04:58,840 you do cognitive neuroscience or not.
86
00:04:58,840 --> 00:05:03,280 This is actually an all pervasive problem.
87
00:05:03,280 --> 00:05:09,360
So I've thought it's my responsibility as an editor to also write about these things.
88
00:05:09,360 --> 00:05:16,880
And of course, in order to get interest, you have to write about these in a way that is not
89
00:05:16,880 --> 00:05:20,280 the normal way of academic writing.
90
00:05:20,280 --> 00:05:24,040 But I just want to assure you readers that I'm also a real academic.
91
00:05:24,040 --> 00:05:30,880
Yes, I can assure our listeners that Masud is an extremely eminent academic.
92
00:05:30,880 --> 00:05:35,680
You don't get to be the editor-in-chief of Brain when you just wander off the street.
93
00:05:35,680 --> 00:05:40,880
But yes, this is a very unusual piece of writing.It's not like a typical Brain Editorial.
94
00:05:40,880 --> 00:05:41,880 But I loved it.
95
00:05:41,880 --> 00:05:47,440 And I think our listeners will love it too.
96
00:05:47,440 --> 00:05:50,160 Okay, so this is the editorial which came out this year.
97
00:05:50,160 --> 00:05:53,960 It's called "A Mountain of Small Things".
98
00:05:53,960 --> 00:05:56,040 I live under its shadow.
99
00:05:56,040 --> 00:05:58,880 I suspect most of you do too.
100
00:05:58,880 --> 00:06:02,600 It is the great mountain of small things.
101
00:06:02,600 --> 00:06:09,680
Every year it grows a little taller, a little more imposing, a little more daunting.
102
00:06:09,680 --> 00:06:17,080
The higher it gets, the bigger the shadow it casts, a malignant darkness that pervades
103
00:06:17,080 --> 00:06:23,120 our lives, one that becomes ever more difficult to breach.
104
00:06:23,120 --> 00:06:30,680
So much so that it has become the norm for many of us to live entirely in the gloom.
105
00:06:30,680 --> 00:06:35,000 We no longer ask why it has come to this, even
106
00:06:35,000 --> 00:06:40,040 though we barely glimpse the light that warmed us in the past.
107
00:06:40,040 --> 00:06:42,840 Here we stand trembling.
108
00:06:42,840 --> 00:06:45,640 Our energy for innovation sapped, our
109
00:06:45,640 --> 00:06:50,000 motivation to focus on research, drained.
110
00:06:50,000 --> 00:06:53,760 Tell me, have you had a great thought lately?
111
00:06:53,760 --> 00:06:55,520 I rest my case.
112
00:06:55,520 --> 00:07:02,080
The mountain of small things makes transformative research, far less likely to happen.
113
00:07:02,080 --> 00:07:06,760 Its shadow smothers our aspirations.
114
00:07:06,760 --> 00:07:10,720 How has it grown so formidably, this mountain?
115
00:07:10,720 --> 00:07:17,280
I have, from time to time, trained my old set of binoculars to inspect its substance.
116
00:07:17,280 --> 00:07:24,080
The most curious thing is that where one might expect rocky outcrops or cascading waterfalls.
117
00:07:24,080 --> 00:07:29,400 There is instead paper or its digital counterpart.
118
00:07:29,400 --> 00:07:35,000 Yes, the mountain is paperwork, sheafs and sheafs of it.
119
00:07:35,000 --> 00:07:41,040
Here there are forms to fill, reports to write, statements of compliance with policy to
120
00:07:41,040 --> 00:07:44,520
sign, and your training to perform, appraisal
121
00:07:44,520 --> 00:07:50,560
documents to upload, research protocols and ethics applications to complete,
122
00:07:50,560 --> 00:07:53,680 and much, much more.
123
00:07:53,680 --> 00:07:59,760
Once you've appreciated this, you understand why we can no longer cross the shadow's edge:
124
00:07:59,760 --> 00:08:06,000
why the mountain of small things simply gets bigger and bigger over the years.
125
00:08:06,000 --> 00:08:12,160
The pressures on our employers and funders from legislation, insurers
and lawyers has meant
126
00:08:12,160 --> 00:08:18,960
there is an irresistible urge to issue, on an annual basis yet more demands upon the people
127
00:08:18,960 --> 00:08:21,080 working in the fields.
128
00:08:21,080 --> 00:08:26,360
If like me, you have ploughed a conscientious furrow, respecting compliantly for years
129
00:08:26,360 --> 00:08:32,480
the edicts that are issued, you will be rewarded by receiving a fresh set of requests.
130
00:08:32,480 --> 00:08:38,360
There will be several more forms and reports, training modules and policies that you have
131
00:08:38,360 --> 00:08:43,960
to comply with this year, generated by an increasing number of staff who are employed
132
00:08:43,960 --> 00:08:46,840 to do just this.
133
00:08:46,840 --> 00:08:52,920
This is doubly so for those of us who are clinicians as well as scientists.
134
00:08:52,920 --> 00:08:57,600
The measures in the edicts will protect us, we're told. They will help secure our institutions
135
00:08:57,600 --> 00:08:58,600 from threats.
136
00:08:58,600 --> 00:09:02,720 They are necessary to mitigate the risks to ourselves. 137
00:09:02,720 --> 00:09:09,640
But with each year, the mountain of small things gets bigger, making it ever more likely
138
00:09:09,640 --> 00:09:14,120 that we continue to live in its growing shadow.
139
00:09:14,120 --> 00:09:16,320
Cold blooded and inert, we
140
00:09:16,320 --> 00:09:23,400
are left unable to devote barely any time to the things we are actually employed to do: research,
141
00:09:23,400 --> 00:09:26,400 teaching and clinical work.
142
00:09:26,400 --> 00:09:32,080 In the darkness we have no hope of growing anything useful.
143
00:09:32,080 --> 00:09:40,240
The result is stunted shoots, disfigured in the hopeless, tenebrific atmosphere.
144
00:09:40,240 --> 00:09:44,840 The pernicious impact of the mountain is hard to estimate.
145
00:09:44,840 --> 00:09:49,200
Ask most researchers though, and they will tell you that they have not, in recent years,
146
00:09:49,200 --> 00:09:53,680 spent any time in the Sunlit Uplands.
147
00:09:53,680 --> 00:09:56,600 Whereas in the past, they enjoyed their jobs, this
148
00:09:56,600 --> 00:09:59,360 feeling has all but disappeared.
149
00:09:59,360 --> 00:10:04,560
The sense of belonging to an institution where there is a community of academics and clinicians
150
00:10:04,560 --> 00:10:10,240
that one can be proud to be part of and learn from - has simply vanished.
151
00:10:10,240 --> 00:10:16,360 We just plough the fields in the frigid darkness.
152
00:10:16,360 --> 00:10:18,440 Does it have to be like this?
153
00:10:18,440 --> 00:10:25,080
From time to time, there is a spark, an ephemeral attempt to remedy the current trajectory, a candle lit,
154
00:10:25,080 --> 00:10:30,040 throwing its meagre beams across the vast landscape.
155
00:10:30,040 --> 00:10:32,760 But it does not last.
156
00:10:32,760 --> 00:10:38,360 The problem is that we do not have the will to resist the mountain.
157
00:10:38,360 --> 00:10:44,200
Instead, we may even unwittingly contribute to it, extinguishing hope for the generations
158
00:10:44,200 --> 00:10:46,120 to come.
159
00:10:46,120 --> 00:10:50,840 Some younger researchers may never have seen the light. 160
00:10:50,840 --> 00:10:56,400
They might hear about times in the past when there was a sense of purpose, camaraderie
161
00:10:56,400 --> 00:10:59,080 and self-fulfillment, but
162
00:10:59,080 --> 00:11:05,240
the mountain's growth is relentless, leading to fewer and fewer people wanting to pursue
163
00:11:05,240 --> 00:11:08,080 a career in its shadow.
164
00:11:08,080 --> 00:11:14,040
There is a serious crisis in attracting and retaining people in biomedical and clinical
165
00:11:14,040 --> 00:11:15,720 science.
166
00:11:15,720 --> 00:11:23,080
If we recognize this but do nothing, we will also have contributed to the growing darkness.
167
00:11:23,080 --> 00:11:27,160 So that's it.
168
00:11:27,160 --> 00:11:29,200 It's a lovely piece.
169
00:11:29,200 --> 00:11:30,680 It really struck me.
170
00:11:30,680 --> 00:11:37,920
And it's kind of shocking that you're basically saying that paperwork
is one of the biggest
171
00:11:37,920 --> 00:11:39,800 problems facing science right now.
172
00:11:39,800 --> 00:11:44,640
It's not an obvious conclusion, but I think once you to think about it, it's
173
00:11:44,640 --> 00:11:47,520 the right conclusion to me at least.
174
00:11:47,520 --> 00:11:53,920
I think about other things that are huge, like P-Hacking, for instance, and the reproducibility
175
00:11:53,920 --> 00:11:54,920 crisis.
176
00:11:54,920 --> 00:11:58,640 And I talked to Dorothy Bishop.
177
00:11:58,640 --> 00:12:00,200 We talked all about those problems.
178
00:12:00,200 --> 00:12:01,640 Then those are definitely problems.
179
00:12:01,640 --> 00:12:07,160
Do you feel like this is an even bigger problem or a problem of the same order of magnitude?
180
00:12:07,160 --> 00:12:09,960 I think it is a bigger problem.
181
00:12:09,960 --> 00:12:19,640
And having been a researcher for over 30 years, I think what people of my vintage realize is
182
00:12:19,640 --> 00:12:23,160
that it doesn't have to be this way.
183
00:12:23,160 --> 00:12:25,200 It never was this way.
184
00:12:25,200 --> 00:12:26,480
And we were doing research,
185
00:12:26,480 --> 00:12:32,800
We didn't have calamitous consequences for either our institutions, the participants
186
00:12:32,800 --> 00:12:36,080 we've studied, or anyone else.
187
00:12:36,080 --> 00:12:44,560
And what has happened is, I don't think anybody in particular is at fault.
188
00:12:44,560 --> 00:12:53,600 It's a growing need to secure institutions from threats.
189
00:12:53,600 --> 00:13:01,080
And those are potentially legal threats, but it really boils down to making decisions
190
00:13:01,080 --> 00:13:03,560 under risk and uncertainty.
191
00:13:03,560 --> 00:13:09,520
As psychologists, some of your listeners would be very used to the idea that it's not really
192
00:13:09,520 --> 00:13:13,200 possible to bring those risks down to zero in the real world.
193
00:13:13,200 --> 00:13:17,320 But that's what I think people are attempting to do. 194
00:13:17,320 --> 00:13:24,160
Even though the evidence that the kind of instruments they're using would reduce the risk
195
00:13:24,160 --> 00:13:25,160 is minimal.
196
00:13:25,160 --> 00:13:27,320 It's absolutely minimal.
197
00:13:27,320 --> 00:13:29,400 So I think it's a bigger problem.
198
00:13:29,400 --> 00:13:36,080
It's not just clinical research, it's in all forms of research, and it's stopping us
199
00:13:36,080 --> 00:13:39,480 getting on with the things, that I said, that we're paid to do.
200
00:13:39,480 --> 00:13:44,720
Yeah, sometimes I feel like they've protected against every single imaginable risk, except
201
00:13:44,720 --> 00:13:50,040
for the risk that in doing so, they would completely stifle our ability to do meaningful
202
00:13:50,040 --> 00:13:51,040 work.
203
00:13:51,040 --> 00:13:54,680
That's the one risk, which nobody seems to be wanting to protect against it all.
204
00:13:54,680 --> 00:14:01,000 But there's no form of is this going tp destroyer ability to work?
205
00:14:01,000 --> 00:14:08,440
So I think it's our responsibility to articulate that problem, because we have been extremely
206
00:14:08,440 --> 00:14:12,840 compliant because most of us are very conscientious.
207
00:14:12,840 --> 00:14:13,840 That's the way we're built.
208
00:14:13,840 --> 00:14:15,920 We do the things we're asked to do.
209
00:14:15,920 --> 00:14:21,840
We don't say no, we don't push back, and we've never questioned this in a sort of concerted
210
00:14:21,840 --> 00:14:23,480 way.
211
00:14:23,480 --> 00:14:25,480 This is the time to do that.
212
00:14:25,480 --> 00:14:29,600
And I think you might say, and a lot of people said to me, "What can I do?
213
00:14:29,600 --> 00:14:31,680 What as an individual can I do?"
214
00:14:31,680 --> 00:14:36,040 But little nudges can have huge effects.
215
00:14:36,040 --> 00:14:44,640
So a little editorial like this has been read over 10,000 times by people, and some of my
216
00:14:44,640 --> 00:14:50,520
colleagues are saying they're using it to motivate themselves and
their groups to push
217
00:14:50,520 --> 00:14:52,920 back against things that aren't necessary.
218
00:14:52,920 --> 00:14:56,240 I'm now doing that on a regular basis, and nothing happens.
219
00:14:56,240 --> 00:15:00,520 It's not like my institution says, "Well, you're fired."
220
00:15:00,520 --> 00:15:05,240
I will just say, "I'm not doing that because there is no necessity to do that."
221
00:15:05,240 --> 00:15:07,800 And usually there isn't a necessity to do that.
222
00:15:07,800 --> 00:15:10,240 There isn't even a legal reason to do that.
223
00:15:10,240 --> 00:15:14,680
If you push back, you will see that it doesn't mean that something terrible is going to happen
224
00:15:14,680 --> 00:15:16,240 to you.
225
00:15:16,240 --> 00:15:24,760
And it's our responsibility to do this because the generations that follow are not going
226
00:15:24,760 --> 00:15:27,960 to be thanking us for not resisting this.
227
00:15:27,960 --> 00:15:29,200 Yeah.
228
00:15:29,200 --> 00:15:35,560
This reminds me of when I was a grad student, I had this side job to make money, transcribing
229
00:15:35,560 --> 00:15:39,760
interviews with biomedical scientists that was like a project, and that was being done in
230
00:15:39,760 --> 00:15:40,760 the library.
231
00:15:40,760 --> 00:15:42,520 It was like oral histories of people who had won some prize.
232
00:15:42,520 --> 00:15:43,960 I don't remember what prize it was.
233
00:15:43,960 --> 00:15:48,680
There was this guy, and I don't remember his name is from the
University of Utah, and
234
00:15:48,680 --> 00:15:51,880 he was some kind of wet lab scientist.
235
00:15:51,880 --> 00:15:58,440
And I remember he said, "Whenever I get a request for something, the first thing I do is
236
00:15:58,440 --> 00:16:01,880 I never respond to anything on the first request.
237
00:16:01,880 --> 00:16:04,320 If it's important, it will come back again."
238
00:16:04,320 --> 00:16:09,520
And he said, "You can make maybe 60% of requests will go away if you simply don't respond
239
00:16:09,520 --> 00:16:10,520
to them."
240
00:16:10,520 --> 00:16:14,400 Is that the kind of strategy that you could have been advocating?
241
00:16:14,400 --> 00:16:21,080
Sometimes I do that, but I think I'm also, I will verbally respond to it by saying, "I
242
00:16:21,080 --> 00:16:24,400 don't see the need for this, and I don't know why we have to do this.
243
00:16:24,400 --> 00:16:26,520
I'm not doing it."
244
00:16:26,520 --> 00:16:32,240
And I will sometimes do that if it's, I don't like reply to all in group emails.
245
00:16:32,240 --> 00:16:37,280
I don't usually do that, but in these cases, I think it's important because in a way,
246
00:16:37,280 --> 00:16:41,440
there's a setting a bar there for other people in your department to say, "Look, he's not
247
00:16:41,440 --> 00:16:42,440 doing this."
248
00:16:42,440 --> 00:16:43,440 Yeah.
249
00:16:43,440 --> 00:16:44,440 Why should I do that?
250
00:16:44,440 --> 00:16:48,200 And I think it's really important for us to resist that. 251
00:16:48,200 --> 00:16:54,440 It's, I'm not saying that administrators are malicious people.
252
00:16:54,440 --> 00:16:59,640
I'm just saying, this is the way they feel they need to protect the institution, or it's
253
00:16:59,640 --> 00:17:03,600 come down from an high, from central university.
254
00:17:03,600 --> 00:17:06,760 And they may not have thought that this is, this has a cost.
255
00:17:06,760 --> 00:17:11,920
All of these things have a cost, and the cost is academic time, which is actually costly
256
00:17:11,920 --> 00:17:13,600 for the university.
257
00:17:13,600 --> 00:17:14,600 Yeah.
258
00:17:14,600 --> 00:17:18,760 And how do your colleagues respond when you tell them no?
259
00:17:18,760 --> 00:17:20,600 I don't get anything back.
260
00:17:20,600 --> 00:17:26,000
Or I'll get an email from some of my colleagues saying, "Thank you for setting that."
261
00:17:26,000 --> 00:17:34,160
This is, I haven't had to reverse a decision I've made because usually there is no reason
262
00:17:34,160 --> 00:17:36,520
for that request coming through.
263
00:17:36,520 --> 00:17:40,400
And it's increasing, you know, the number of requests, I'm sure it is in your institution.
264
00:17:40,400 --> 00:17:46,640
The number of requests you get for simple things are just out of control.
265
00:17:46,640 --> 00:17:53,320
In the UK, over, I think over a decade, the number of academics doubled, but the number
266
00:17:53,320 --> 00:18:00,520
of administrators in UK universities, quadrupled, you know, those people have to do something,
267
00:18:00,520 --> 00:18:02,400 they're employed to do something.
268
00:18:02,400 --> 00:18:04,560 And they also employed to have initiatives.
269
00:18:04,560 --> 00:18:06,480 And those initiatives include, "Oh, why don't we do this?"
270
00:18:06,480 --> 00:18:09,040 I'm not blaming them.
271
00:18:09,040 --> 00:18:13,320
That's what they're employed to do, but we don't actually need that many of them.
272
00:18:13,320 --> 00:18:15,640 No, we got by fine in the past, right?
273
00:18:15,640 --> 00:18:16,640
We did.
274
00:18:16,640 --> 00:18:17,640 And yeah, that's the thing.
275
00:18:17,640 --> 00:18:22,360
I guess that's the thing you and I are old enough that we've seen this change in the
276
00:18:22,360 --> 00:18:24,080 course of our careers.
277
00:18:24,080 --> 00:18:30,040
Can you talk about, like, can you give any examples of things that you did research-wise
278
00:18:30,040 --> 00:18:35,000
when you were young academic just starting out that would never fly nowadays?
279
00:18:35,000 --> 00:18:40,200
Well, I think that's particularly the case with patient-related research, which is a special
280
00:18:40,200 --> 00:18:41,520 case.
281
00:18:41,520 --> 00:18:48,680
And it used to be possible to get an ethics application approved for observational or
282
00:18:48,680 --> 00:18:55,920
behavioral studies and perhaps imaging in patients within about two or three weeks.
283
00:18:55,920 --> 00:18:59,640 I got a major award last year.
284
00:18:59,640 --> 00:19:02,600 It started in December last year. 285
00:19:02,600 --> 00:19:05,960 We have only just got ethics approval.
286
00:19:05,960 --> 00:19:12,200
And then we still can't start this work because there is a division if you do clinical
287
00:19:12,200 --> 00:19:18,320
related research between hospitals, the NHS in the UK and the university, and
288
00:19:18,320 --> 00:19:22,360 everyone's really haggling for a bit of the pie.
289
00:19:22,360 --> 00:19:28,800
Despite the fact that this is really simple observation research, patients are very happy
290
00:19:28,800 --> 00:19:30,040 to get involved.
291
00:19:30,040 --> 00:19:34,040 There is no intervention, there's no risk here.
292
00:19:34,040 --> 00:19:39,880
Everything is about trying to get money out of this in terms of, you know, and that slows
293
00:19:39,880 --> 00:19:41,400 the process down.
294
00:19:41,400 --> 00:19:45,680 Is any hospital facility being used for this?
295
00:19:45,680 --> 00:19:49,000 Is there a room that isn't a university room that's being used?
296
00:19:49,000 --> 00:19:50,000 All right.
297
00:19:50,000 --> 00:19:51,000 All that kind of stuff.
298
00:19:51,000 --> 00:19:56,120
And you think, well, if you worked out how much it's costing to have administrators do this
299
00:19:56,120 --> 00:19:57,120 work?
300
00:19:57,120 --> 00:20:01,200 You would realize that this is, actually not worth the while.
301
00:20:01,200 --> 00:20:05,040 The money we're bringing in, in terms of research is not so big.
302
00:20:05,040 --> 00:20:08,200 We're not corporations that you have to worry.
303
00:20:08,200 --> 00:20:10,280 These are pennies we're talking about.
304
00:20:10,280 --> 00:20:12,280 But that's just giving you an example.
305
00:20:12,280 --> 00:20:15,680
What could be done in three weeks has still has not been done in ten months.
306
00:20:15,680 --> 00:20:16,680 Yeah.
307
00:20:16,680 --> 00:20:17,680 And it's going to be even longer.
308
00:20:17,680 --> 00:20:21,000 It's definitely the same here.
309
00:20:21,000 --> 00:20:23,760 Like you wouldn't get anything up on the ground very quickly.
310
00:20:23,760 --> 00:20:26,280 When I was thinking about this, yeah, go on.
311
00:20:26,280 --> 00:20:30,000
I was just saying, and of course, I was just talking here about observational studies.
312
00:20:30,000 --> 00:20:36,840
Imagine that I had a new intervention for a disease which means that your lifespan is
313
00:20:36,840 --> 00:20:38,200 really reduced.
314
00:20:38,200 --> 00:20:45,800
I can't get that intervention, drug or whatever it is, to you even try because of these processes.
315
00:20:45,800 --> 00:20:53,280
And I think if patients really understood that these administrative processes are blocking
316
00:20:53,280 --> 00:20:59,800
them being able to trial these new interventions, they would actually probably voice the same
317
00:20:59,800 --> 00:21:01,760 concerns that we are beginning to voice.
318
00:21:01,760 --> 00:21:06,000 So the other way of doing this is getting the public involved.
319
00:21:06,000 --> 00:21:07,000
Yeah.
320
00:21:07,000 --> 00:21:11,320
Were you surprised at how fast they manage to get the COVID vaccine out?
321
00:21:11,320 --> 00:21:14,320 Or do you think they could have got it out even quicker without,
322
00:21:14,320 --> 00:21:16,840 I mean, they sort of had a working vaccine in weeks.
323
00:21:16,840 --> 00:21:17,840 Right.
324
00:21:17,840 --> 00:21:22,920
So, you know, I work at Oxford and that's where it came from in the UK.
325
00:21:22,920 --> 00:21:32,360
And essentially, the regulatory bodies fast tracked those applications.
326
00:21:32,360 --> 00:21:34,240 They had an incentive to do that.
327
00:21:34,240 --> 00:21:38,120 It worked out perfectly well, but it showed us that it's possible.
328
00:21:38,120 --> 00:21:39,120 It did.
329
00:21:39,120 --> 00:21:46,000
Now, the risks of a vaccine are much higher than the kind of studies that most of us are doing.
330
00:21:46,000 --> 00:21:47,000
Right.
331
00:21:47,000 --> 00:21:50,120 We're not doing any interventional studies.
332
00:21:50,120 --> 00:21:54,240
Yet it's taken me 10 months and I still haven't got an approval to do this.
333
00:21:54,240 --> 00:21:59,840
Whereas a COVID vaccine which came with potential far higher risks as well as potential benefits
334
00:21:59,840 --> 00:22:03,920 could be fast tracked within a few weeks.
335
00:22:03,920 --> 00:22:05,880 So it shows us that it's possible.
336
00:22:05,880 --> 00:22:11,360
And it makes us think why on Earth couldn't we do this on a regular basis?
337
00:22:11,360 --> 00:22:12,600 Yeah.
338
00:22:12,600 --> 00:22:18,400
And I think back to when I was a grad student, and we used to just use the scanner, for free
339
00:22:18,400 --> 00:22:21,720 after hours because no one else was using it.
340
00:22:21,720 --> 00:22:23,240 It was just sitting there.
341
00:22:23,240 --> 00:22:27,480
And you could just go in and you know, we consented the research participants, you know, we had an IRB.
342
00:22:27,480 --> 00:22:32,960 We could scan a patient, scan the participant, participants.
343
00:22:32,960 --> 00:22:36,760
But like, you know, a grad student who hadn't really had any training was like allowed to
344
00:22:36,760 --> 00:22:41,160 just like enter the building after hours, scan a person.
345
00:22:41,160 --> 00:22:46,360
And when I think about like what that allowed me to do in terms of, by learning, like, I was
346
00:22:46,360 --> 00:22:47,880 just out of play around with stuff, right?
347
00:22:47,880 --> 00:22:51,320
I just tried out a million different things and like, I'd just make up, I'd just think of some
348
00:22:51,320 --> 00:22:53,400 paradigm over the weekend and
349
00:22:53,400 --> 00:22:57,600
I just, you know, coded it up and two days later I just ran it on somebody.
350
00:22:57,600 --> 00:23:00,440 And like most of them didn't work and some of them did.
351
00:23:00,440 --> 00:23:02,680 And those are the ones that grew into like lines of research later.
352
00:23:02,680 --> 00:23:05,320 I just think like, what about young people nowadays?
353
00:23:05,320 --> 00:23:10,960
They're not getting that opportunity to just like explore because everything is so hard to
354
00:23:10,960 --> 00:23:11,960 get started.
355
00:23:11,960 --> 00:23:18,160
You're absolutely right. After I did my PhD, I went to do a postdoc at MIT.
356
00:23:18,160 --> 00:23:23,520
And I could not believe the culture there because like everyone else, we chat over a coffee
357
00:23:23,520 --> 00:23:27,480 about a potential thought experiment.
358
00:23:27,480 --> 00:23:31,640
When I've been in Oxford doing my PhD, none of that actually translated into anything, those
359
00:23:31,640 --> 00:23:32,640 thought experiments.
360
00:23:32,640 --> 00:23:39,960
At MIT, full of postdocs from different parts of the world who are there for a short time,
361
00:23:39,960 --> 00:23:44,440
there are incentives to get on with things. People would just go, well, let's do it tonight.
362
00:23:44,440 --> 00:23:46,440 And I would say, what do you mean tonight?
363
00:23:46,440 --> 00:23:50,680 Well, let's code it up now and let's do it tonight.
364
00:23:50,680 --> 00:23:54,400 And we would be doing it that evening, just like you said. 365
00:23:54,400 --> 00:23:58,200 And most of those things didn't work out, but occasionally they did.
366
00:23:58,200 --> 00:24:01,880 But it also made you feel like you were doing science.
367
00:24:01,880 --> 00:24:02,880 You were at the cutting edge.
368
00:24:02,880 --> 00:24:03,880 You were trying something.
369
00:24:03,880 --> 00:24:04,880 It didn't work out.
370
00:24:04,880 --> 00:24:05,880 Okay, why doesn't it work out?
371
00:24:05,880 --> 00:24:07,600 Let's play with this.
372
00:24:07,600 --> 00:24:09,920 Those days are gone.
373
00:24:09,920 --> 00:24:19,920
Some of my other colleagues who really enjoyed the time to think about a problem, say that
374
00:24:19,920 --> 00:24:25,320
that has become extremely difficult for them to just sit there thinking for some time.
375
00:24:25,320 --> 00:24:29,440
And I guess if you're an administrator, walking passed an office seeing Stephen there and sitting
376
00:24:29,440 --> 00:24:32,920
there thinking, what on earth are you doing?
377
00:24:32,920 --> 00:24:34,240 You're not doing something useful.
378
00:24:34,240 --> 00:24:39,280
But of course, that is something useful. That's the whole purpose of our being in what we
379
00:24:39,280 --> 00:24:40,280 do, right?
380
00:24:40,280 --> 00:24:41,400 Yeah, right.
381
00:24:41,400 --> 00:24:44,440 So we need to get that back.
382
00:24:44,440 --> 00:24:46,960 And there is no reason we can't get that back.
383
00:24:46,960 --> 00:24:54,840
But we do have to express the problem, articulate it well, and put it on the table for people
384
00:24:54,840 --> 00:24:57,200 to understand that there is an issue.
385
00:24:57,200 --> 00:25:02,360 And pushing back is just the beginning of trying to change this.
386
00:25:02,360 --> 00:25:08,320
I also understand that trying to do this within an institution is all fine, but what we really
387
00:25:08,320 --> 00:25:14,840
need is a much bigger kind of framework to do this because
institutions need to feel like
388
00:25:14,840 --> 00:25:18,760 they're doing something that other places are doing too.
389
00:25:18,760 --> 00:25:24,920
There is no reason why that can't happen if we can get the push on this for this to happen.
390
00:25:24,920 --> 00:25:30,480
We just need to hold them by the hand and allow them to take a little bit more risk than
391
00:25:30,480 --> 00:25:34,760 they're doing because it didn't cause any problems in the past.
392
00:25:34,760 --> 00:25:37,880
And of course, they will sometimes say, well, look, there's an example of this.
393
00:25:37,880 --> 00:25:40,480 This went wrong.
394
00:25:40,480 --> 00:25:42,440 But that is what we live with.
395
00:25:42,440 --> 00:25:43,680 That's how it happens.
396
00:25:43,680 --> 00:25:45,680 There are always going to be problems.
397
00:25:45,680 --> 00:25:46,960 You can't get rid of them all.
398
00:25:46,960 --> 00:25:50,960 No, nothing is going to be perfect.
399
00:25:50,960 --> 00:25:55,960
Yeah, but like, apart from, you know, so if you describe like individual action we can take
400
00:25:55,960 --> 00:26:03,800
a distance of pushing back against some unreasonable requests here and there, and I'm just thinking
401
00:26:03,800 --> 00:26:08,960
of a couple of days ago, I received a request to do something truly meaningless and pointless, to
402
00:26:08,960 --> 00:26:13,640
attend a two hour meeting, which I would give a three minute, single slide presentation
403
00:26:13,640 --> 00:26:18,000
on something that I actually don't have any knowledge or familiarity with anyway.
404
00:26:18,000 --> 00:26:23,400
And I looked at my, I thought, oh, I hope I have, I hope I've already got something on
405
00:26:23,400 --> 00:26:24,400 at that time.
406
00:26:24,400 --> 00:26:26,960 And so I looked at my calendar and was like, oh, I don't.
407
00:26:26,960 --> 00:26:30,600 And I thought, but the person that sent the email doesn't know that.
408
00:26:30,600 --> 00:26:32,640 And so I wrote back and I was like, oh, I'd love to, but
409
00:26:32,640 --> 00:26:35,120 unfortunately, I've got a pre-existing commitment.
410
00:26:35,120 --> 00:26:38,240 And so, you know, we can make these little small pushbacks. 411
00:26:38,240 --> 00:26:43,840
But like, how do you see it like ramping up and really changing, like the whole enterprise?
412
00:26:43,840 --> 00:26:46,680 Like, do you think there's ways of scaling up?
413
00:26:46,680 --> 00:26:52,360 Well, I mean, you know, I would say you, you, you ducked out there.
414
00:26:52,360 --> 00:26:58,040
You could have said, I don't want to do this because you know, I'm,
I'm, this is not something that
415
00:26:58,040 --> 00:26:59,440 I have any expertise in.
416
00:26:59,440 --> 00:27:02,600 And I actually need those two hours to do something else.
417
00:27:02,600 --> 00:27:07,000
So I think we need to say these things rather than say, I don't, oh, sorry, I can't make
418
00:27:07,000 --> 00:27:09,520 it because I'm doing something else.
419
00:27:09,520 --> 00:27:11,600 So I'm going to point the finger back at you.
420
00:27:11,600 --> 00:27:12,600 OK, alright. I'll take that.
421
00:27:12,600 --> 00:27:13,600
It was,
422
00:27:13,600 --> 00:27:14,600
It was,
423
00:27:14,600 --> 00:27:20,200 It was a little bit, I was a little bit shady and yeah, not very...
424
00:27:20,200 --> 00:27:25,320
But seriously, I think what we really need to do is to get to departmental heads because
425
00:27:25,320 --> 00:27:32,760
they're the people who meet on the wider table of the university and start having a conversation
426
00:27:32,760 --> 00:27:33,760 with them.
427
00:27:33,760 --> 00:27:38,720
Now, of course, many departmental heads will not necessarily be sympathetic or even if they're
428
00:27:38,720 --> 00:27:41,880
sympathetic, they'll just say, shrug the shoulders and say, well, this is the way the
429
00:27:41,880 --> 00:27:43,520 world is.
430
00:27:43,520 --> 00:27:45,920 This is not the way the world needs to be.
431
00:27:45,920 --> 00:27:49,160 And we're the people who are doing the work, right?
432
00:27:49,160 --> 00:27:53,600
We're the people who are doing the work for the universities and we don't have to do it
433
00:27:53,600 --> 00:27:54,600 this way.
434
00:27:54,600 --> 00:27:57,640 So we, I'm not militant.
435
00:27:57,640 --> 00:28:02,080 I've never been anything but straightforward.
436
00:28:02,080 --> 00:28:11,200
But I think it is our responsibility as academics, intellectuals, people who are supposed to be
437
00:28:11,200 --> 00:28:16,280
thinking about the future and what we can do with it to actually take action.
438
00:28:16,280 --> 00:28:21,960
All action can make a big difference if it's cumulative or it's across the board.
439
00:28:21,960 --> 00:28:29,000
If 50% of your department agreed to this and said so to the departmental head, they need
440
00:28:29,000 --> 00:28:30,000 to think.
441
00:28:30,000 --> 00:28:31,000 Mm-hmm.
442
00:28:31,000 --> 00:28:32,000 Yeah.
443
00:28:32,000 --> 00:28:40,080
And do you think like, what do you think a young person entering the field can do to make
444
00:28:40,080 --> 00:28:43,200 their way through this new world that we find ourselves in? 445
00:28:43,200 --> 00:28:45,920 I mean, they don't have the air of the department head.
446
00:28:45,920 --> 00:28:47,680 Can they just protect themselves?
447
00:28:47,680 --> 00:28:51,480 Is that all they can hope to do as they get started?
448
00:28:51,480 --> 00:28:52,640 I think it's really difficult.
449
00:28:52,640 --> 00:28:58,320
It's difficult for us to navigate this sort of Byzantine complex of things you have to
450
00:28:58,320 --> 00:28:59,320 get through.
451
00:28:59,320 --> 00:29:02,720 It's not straightforward at all to me.
452
00:29:02,720 --> 00:29:06,400 So I have every sympathy for younger people starting this.
453
00:29:06,400 --> 00:29:12,120
If I had my big grant and I didn't have a preexisting grant, I would be tearing my
454
00:29:12,120 --> 00:29:15,880 hair out, my out, because it's 10 months and I haven't started.
455
00:29:15,880 --> 00:29:21,520
I did have a preexisting grant so I could just extend the old ethics and keep that going
456
00:29:21,520 --> 00:29:25,480 while we wasted time trying to get the new one. 457
00:29:25,480 --> 00:29:32,640
And so I think it's really difficult for younger people and we shouldn't forget that.
458
00:29:32,640 --> 00:29:37,680
Even when they get a grant, it's obviously just the beginning of the problem that they're
459
00:29:37,680 --> 00:29:39,520 confronting.
460
00:29:39,520 --> 00:29:46,080
So if I was in their position, I think I would also kind of document the problems they're
461
00:29:46,080 --> 00:29:51,880
having because without any documentation of how long it's taken you to get through this
462
00:29:51,880 --> 00:29:56,560 or why you've had a problem with this, it's meaningless.
463
00:29:56,560 --> 00:30:05,080
And departmental heads can't do anything unless they have real cases, scenarios because
464
00:30:05,080 --> 00:30:07,240 they can't really take this forward anywhere.
465
00:30:07,240 --> 00:30:08,640 It's just a little bit vague.
466
00:30:08,640 --> 00:30:15,000
So probably the best thing a young person can do is to document exactly what the hurdles
467
00:30:15,000 --> 00:30:21,120
have been and why it's been so frustrating and how long it's taken them.
468
00:30:21,120 --> 00:30:22,120 Yeah.
469
00:30:22,120 --> 00:30:25,800 And if they're hearing that from all sides, then that might lead to...
470
00:30:25,800 --> 00:30:26,800 Exactly.
471
00:30:26,800 --> 00:30:27,800
Imagine you're the head of department,
472
00:30:27,800 --> 00:30:32,360
You will after a while get frustrated with email after email telling you this is what's
473
00:30:32,360 --> 00:30:33,360 happening.
474
00:30:33,360 --> 00:30:35,680 You're going to have to do something.
475
00:30:35,680 --> 00:30:41,720
So what's happened in my own department is that we've had these conversations with heads
476
00:30:41,720 --> 00:30:50,240
of department but also with administrative heads so that they understand.
477
00:30:50,240 --> 00:30:56,080
They may not understand that there is a problem because this is just what they're supposed
478
00:30:56,080 --> 00:30:57,080 to do.
479
00:30:57,080 --> 00:30:59,320 This is what central universities told them to do.
480
00:30:59,320 --> 00:31:03,920
So I think it's really important to get them on your side and make them think, "Okay,
481
00:31:03,920 --> 00:31:07,200 I didn't realize that this is an issue."
482
00:31:07,200 --> 00:31:10,960 Because they would just say, "But this would take an hour to do.
483
00:31:10,960 --> 00:31:11,960 What's the problem?"
484
00:31:11,960 --> 00:31:16,640 And what they don't know is you're being asked to do 10 hours of this.
485
00:31:16,640 --> 00:31:17,640 That's so true.
486
00:31:17,640 --> 00:31:18,640 That's so true.
487
00:31:18,640 --> 00:31:19,640 Yeah.
488
00:31:19,640 --> 00:31:24,280
They all seem so trivial in isolation that when you complain, you feel like you're being
489
00:31:24,280 --> 00:31:25,280 a whiner for complaining.
490
00:31:25,280 --> 00:31:26,280
Well, but
491
00:31:26,280 --> 00:31:31,720
what they don't understand is that the thing you asked me, took me 10 minutes but a thousand
492
00:31:31,720 --> 00:31:33,800 people asked me for 10 minutes.
493
00:31:33,800 --> 00:31:39,160
And the other thing that they don't understand, I think, is that if you think of my 40-hour
494
00:31:39,160 --> 00:31:42,880
work week, which is really, of course, like a 50 or 60-hour work week or I don't know how
495
00:31:42,880 --> 00:31:48,800
you are, but most of us probably putting in more than 40, you could say, "Okay, well,
496
00:31:48,800 --> 00:31:53,240 it's only like this chore is only 1% of your work week.
497
00:31:53,240 --> 00:31:54,240
Why are you complaining?"
498
00:31:54,240 --> 00:32:01,840
But of my disposable time, 36 hours of my 40-hour work week is already fully taken up
499
00:32:01,840 --> 00:32:06,440
with teaching and ongoing responsibilities and things that are just scheduled and going
500
00:32:06,440 --> 00:32:07,440 to happen.
501
00:32:07,440 --> 00:32:14,320
It's like if you dig into the remaining time, it's actually only a
very small part of disposable
502
00:32:14,320 --> 00:32:15,320 time.
503
00:32:15,320 --> 00:32:19,320
It's like when you dig into it that you're taking much more than you think as a percentage,
504
00:32:19,320 --> 00:32:20,800 you said what I'm saying.
505
00:32:20,800 --> 00:32:28,440
Well, that is exactly the inspiration for a mountain of small things because that's exactly
506
00:32:28,440 --> 00:32:29,440 what I was thinking.
507
00:32:29,440 --> 00:32:34,280
I'm going to convey the idea that these are all trivial little things, but actually you
508
00:32:34,280 --> 00:32:35,280 put them together, it's
509
00:32:35,280 --> 00:32:38,520 a huge obstacle.
510
00:32:38,520 --> 00:32:44,000
That was the actual inspiration behind thinking about the title for this and that sort of concept
511
00:32:44,000 --> 00:32:45,440 for the editorial.
512
00:32:45,440 --> 00:32:57,280
Yeah, and you gave it just such like, poignant examples, training modules, forms and reports.
513
00:32:57,280 --> 00:33:02,400
What I found after some of this conversation with our administrators is they've said, "Well,
514
00:33:02,400 --> 00:33:09,760
actually X, Y and Z are not absolute requirements."
515
00:33:09,760 --> 00:33:16,520
But A is, everybody needs to do A, but X, Y and Z, when it comes to when push comes to
516
00:33:16,520 --> 00:33:21,120 shove, X, Y and Z don't really need to be done.
517
00:33:21,120 --> 00:33:22,120 That's been helpful.
518
00:33:22,120 --> 00:33:24,120 They've said, "Okay."
519
00:33:24,120 --> 00:33:28,120
So these are little small gains, but I think that's what we're going to have to do.
520
00:33:28,120 --> 00:33:32,160 We're going to have to dismantle the mountain one piece at a time.
521
00:33:32,160 --> 00:33:36,680
It's not going to be, we're going to sweep this thing away in one go, but the small gains
522
00:33:36,680 --> 00:33:37,680 are worth it.
523
00:33:37,680 --> 00:33:43,200
I think that's what I'd like to hope that your listeners will think about is that each of
524
00:33:43,200 --> 00:33:45,040 these small things are little wins.
525
00:33:45,040 --> 00:33:50,080
Yeah, do you feel optimistic or do you just kind of feel like, gosh,
I'm going to like,
526
00:33:50,080 --> 00:33:55,200
swimming is the tide, but I'm really going to get swept away at the end of the day.
527
00:33:55,200 --> 00:33:57,840 I'm not going to stop swimming, but I'm going to get swept away.
528
00:33:57,840 --> 00:34:04,640
Yeah, a very senior neuroscientist has described me as being like King
Canute in front
529
00:34:04,640 --> 00:34:13,360 of the sea, trying to whoosh the sea away before the deluge comes.
530
00:34:13,360 --> 00:34:19,480
I'm not necessarily super optimistic, but I have been really amazed by the reaction
531
00:34:19,480 --> 00:34:20,480 to these editorials.
532
00:34:20,480 --> 00:34:23,600 There are a couple of others.
533
00:34:23,600 --> 00:34:31,800
And how people have actually emailed me or texted me or gone on X to say, how good they
534
00:34:31,800 --> 00:34:39,140
feel that somebody actually is articulating this in a way that is coherent and perhaps can
535
00:34:39,140 --> 00:34:44,760 be used by as material in the argument.
536
00:34:44,760 --> 00:34:47,880 So I'm not necessarily naive about this.
537
00:34:47,880 --> 00:34:53,160
I'm not sure that we're going to be able to change everything here, but I mean, hey, if
538
00:34:53,160 --> 00:34:57,120 we don't actually try, then there is no hope.
539
00:34:57,120 --> 00:35:01,720
I have to compliment you on the way you just used the word X as if it was like, like it just
540
00:35:01,720 --> 00:35:02,720 rolled off your tongue.
541
00:35:02,720 --> 00:35:07,060
I think it's the first time I've ever heard somebody say X and not say, you know, the site
542
00:35:07,060 --> 00:35:11,160
formerly known as Twitter, so called X, you know, you just like, throw it right in there,
543
00:35:11,160 --> 00:35:13,960 right into your sounds without skipping a beat.
544
00:35:13,960 --> 00:35:16,720 That's very good.
545
00:35:16,720 --> 00:35:21,720
But you know, I think maybe when I could earlier I mentioned, like, you know, the sort of reproducibility
546
00:35:21,720 --> 00:35:24,280
crisis, right?
547
00:35:24,280 --> 00:35:32,440
That started with conversation and exposure and bringing into the sort of consciousness
548
00:35:32,440 --> 00:35:34,440 of everybody, right?
549
00:35:34,440 --> 00:35:38,120
20 years ago, nobody was thinking about those issues and then people started talking about
550
00:35:38,120 --> 00:35:44,160
them and then, you know, momentum built and I mean, still an issue, still a huge issue.
551
00:35:44,160 --> 00:35:46,640 But like, you feel that things are changing, right?
552
00:35:46,640 --> 00:35:50,400
Clearly there's like a shift has happened in the way we're doing science as a result of
553
00:35:50,400 --> 00:35:54,000 that and it didn't happen overnight or 10 or 20 years.
554
00:35:54,000 --> 00:36:00,560
So I think what you're doing, kick starting this conversation, I can hope, I mean, like,
555
00:36:00,560 --> 00:36:05,840
I'm a sort of naturally pessimistic person, but if there is hope, then
I think that it's
556
00:36:05,840 --> 00:36:11,000
going to be kind of, it's, it's just starting to talk about it and say, this is a problem
557
00:36:11,000 --> 00:36:15,240
that is bigger than the reproducibility crisis that is crushing science.
558
00:36:15,240 --> 00:36:21,000
Like this paper work obsession, like getting that to be kind of like a mainstream opinion,
559
00:36:21,000 --> 00:36:23,800
because you know, as soon as you say it, other people are like, hey, you're right.
560
00:36:23,800 --> 00:36:25,960 Like, I don't enjoy science anymore.
561
00:36:25,960 --> 00:36:30,800 I've not had exactly and I've not had any messages saying, I'm wrong.
562
00:36:30,800 --> 00:36:31,800 Yeah.
563
00:36:31,800 --> 00:36:32,800 So many really interesting things.
564
00:36:32,800 --> 00:36:38,560
Actually, I really like writing, writing ethics, ethics proposals and waiting 10 months
565
00:36:38,560 --> 00:36:39,840 for them.
566
00:36:39,840 --> 00:36:42,240 And you know, those online training modules are very helpful.
567
00:36:42,240 --> 00:36:47,080
Like, it's really important to know like what the symbol is for hazardous gases when
568
00:36:47,080 --> 00:36:49,320
I work in an office building.
569
00:36:49,320 --> 00:36:50,320 Yeah.
570
00:36:50,320 --> 00:36:51,320 Yeah.
571
00:36:51,320 --> 00:36:56,880
But I also think it's important to think about younger people because they also need to realize
572
00:36:56,880 --> 00:37:02,200
that this is a problem because otherwise, they're not going to understand that it wasn't
573
00:37:02,200 --> 00:37:03,200 always like this.
574
00:37:03,200 --> 00:37:04,200 Yeah.
575
00:37:04,200 --> 00:37:05,720 It really wasn't like this, right?
576
00:37:05,720 --> 00:37:12,160
And within our lifetimes, it has changed into this, this monster, which is sort of throttling
577
00:37:12,160 --> 00:37:14,080 our energy to do other stuff.
578
00:37:14,080 --> 00:37:18,400 And it really is having a detrimental effect on innovation.
579
00:37:18,400 --> 00:37:20,320 You don't have to the time to think.
580
00:37:20,320 --> 00:37:24,160 You don't have the time to play.
581
00:37:24,160 --> 00:37:26,960 And that's such an important part of doing research, right?
582
00:37:26,960 --> 00:37:28,360 Trying something out.
583
00:37:28,360 --> 00:37:29,360 Absolutely.
584
00:37:29,360 --> 00:37:31,560 It's really crucial.
585
00:37:31,560 --> 00:37:32,560 Yeah.
586
00:37:32,560 --> 00:37:33,560 Cool.
587
00:37:33,560 --> 00:37:43,480
Thank you so much for, you know, taking me out on my unexpected request to be on a podcast.
588
00:37:43,480 --> 00:37:47,720
I'm really glad, you know, to have met you and to, you know, kind of get your message,
589
00:37:47,720 --> 00:37:52,320
like help spread your message a little bit because I really enjoyed it.
590
00:37:52,320 --> 00:37:53,640 Thank you very much, Stephen. Thanks.
591
00:37:53,640 --> 00:37:54,640
Yeah.
592
00:37:54,640 --> 00:37:59,360
So good luck getting, finding some time today to do some real work, do some thinking, do
593
00:37:59,360 --> 00:38:00,360 some science.
594
00:38:00,360 --> 00:38:01,840 I'm going to try.
595
00:38:01,840 --> 00:38:02,840 Yeah. Definitely.
596
00:38:02,840 --> 00:38:03,840 Okay.
597
00:38:03,840 --> 00:38:04,840 All right.
598
00:38:04,840 --> 00:38:05,840 Thank you so much.
599
00:38:05,840 --> 00:38:06,840 Hey, yeah.
600
00:38:06,840 --> 00:38:07,840 Thanks, Stephen. Bye.
601
00:38:07,840 --> 00:38:08,840 Okay.
602
00:38:08,840 --> 00:38:09,840 Okay.
603
00:38:09,840 --> 00:38:10,840 Well, that's it for episode 31.
604
00:38:10,840 --> 00:38:14,440
Thanks a lot, Masud, for coming on the podcast and reading aloud your editorial and talking
605
00:38:14,440 --> 00:38:15,440 about it with me.
606
00:38:15,440 --> 00:38:16,440 Really enjoyed it.
607
00:38:16,440 --> 00:38:17,840 And I hope you guys did too.
608
00:38:17,840 --> 00:38:18,840 All right.
609
00:38:18,840 --> 00:38:19,840 Bye for now.
610
00:38:19,840 --> 00:38:19,840 See you next time.
611
00:38:20,840 --> 00:38:20,840
612
00:38:21,840 --> 00:38:21,840
613
00:38:22,840 --> 00:38:23,840
614
00:38:23,840 --> 00:38:28,000
[Music]
‘A mountain of small things’ with Masud Husain
Oct 21, 2024•38 min•Season 4Ep. 31
Episode description
In this episode, I talk with Masud Husain, Professor of Neurology and Cognitive Neuroscience at the University of Oxford, about his recent editorial ‘A mountain of small things’.
Husain M. A mountain of small things. Brain 2024; 147: 739. [doi]
Transcript
Transcript source: Provided by creator in RSS feed: download file