Whatever the context(s) of these events, the words are right.
The job of a scientist is to look for the truth, and the job of a teacher is to help people to empower themselves. I failed to do my job on both counts.
It is however not just the job of scientists.
The quote below was from a piece in the Lancet by Richard Horton.
Reading [Bertrand]Russell today is a resonant experience. Existential fears surround us. Yet today seems a long way from the dream of Enlightenment. Modern science is a brutally competitive affair. It is driven by incentives to acquire money (research funding), priority (journal publication), and glory (prizes and honours). Science’s metrics of success embed these motivations deep in transnational scientific cultures. At The Lancet, while we resist the idea that Impact Factors measure our achievements, we are not naive enough to believe that authors do not judge us by those same numbers. It is hard not to capitulate to a narrow range of indicators that has come to define success and failure. Science, once a powerful force to overturn orthodoxy, has created its own orthodoxies that diminish the possibility of creative thought and experiment. At this moment of planetary jeopardy, perhaps it is time to rethink and restate the purpose of science.
I am just musing on this. We like to think that ‘freedom’ was necessary for a modern wealthy state. We are not so certain, now. We used to think that certain freedoms of expression underpinned the scientific revolution. We are having doubts about this, too. Maybe it is possible to have atom bombs and live in a cesspool of immorality. Oops…
Direct URL for this post.
It is said that much of the foundations of 20th century physics was done in coffee houses (or in the case of Richard Feynman in strip bars), but things were once done differently in the UK
With neither institutional nor government masters to answer to, the British cyberneticians were free to concentrate on what interested them. In 1949, in an attempt to develop a broader intellectual base, many of them formed an informal dining society called the Ratio Club. Pickering documents that the money spent on alcohol at the first meeting dwarfed that spent on food by nearly six to one — another indication of the cultural differences between the UK and US cyberneticians.
The work of the British pioneers was forgotten until the late 1980s when it was rediscovered by a new generation of researchers… A company that I cofounded has now sold more than five million domestic floor-cleaning robots, whose workings were inspired by Walter’s tortoises. It is a good example of how unsupported research, carried out by unconventional characters in spite of their institutions, can have a huge impact.
A review from 2010 by Rodney Brooks of MIT of “The Cybernetic Brain: Sketches of Another Future” in Nature (For more on Donald Michie and “in spite of their institutions” see here).
Direct URL for this post.
I have had of all people a historian tell me that science is a collection of facts, and his voice had not even the ironic rasp of one filing-cabinet reproving another.
Jacob Bronowski | Science and Human Values
Direct URL for this post.
I have removed the name of the institution only because so many queue to sell their vapourware in this manner
Precision Medicine is a revolution in healthcare. Our world-leading biomedical researchers are at the forefront of this revolution, developing new early diagnostics and treatments for chronic diseases including cancer, cardiovascular disease, diabetes, arthritis and stroke. Partnering with XXXXX, the University of XXXX has driven … vision in Precision Medicine, including the development of a shitload of infrastructure to support imaging, molecular pathology and precision medicine clinical trials…… XXXXXX is now one of the foremost locations in a three mile radius to pursue advances in Precision Medicine.
And He declared to them, “It is written: ‘My house will be called a house of prayer. But you are making it ‘a den of robbers.'” Matthew 21:13
Direct URL for this post.
Statistics — to paraphrase Homer Simpson’s thoughts on alcohol — is the cause of, and solution to, all of science’s problems.
Direct URL for this post.
Digging deep into some of my old notes, I came across this obituary of John Ziman written by Jerry Ravetz. I know both through their written work and was lucky enough to meet and chat briefly with John Ziman not long before he died. Ziman’s book “Real Science” is for me the classic account of what has happened to science as it moved from a ‘way of life’ to a job.
Jerry Ravetz writes:
I first became aware of him through his 1960 radio talk Scientists – Gentlemen Or Players?, where he observed how a career in science was starting to change, from being a vocation to being a job.
There was a paradox running through his later career, to which he must have been sensitive. He was a “Renaissance man” in a way highly desirable for a scientist, but he did not exert the influence that he might have hoped to. This was due less to the passion he deployed in argument than the times in which he found himself. The age of such eminent scientist-savants as JBS Haldane, JD Bernal and Joseph Needham was passing, while a new generation of socially responsible scientists had yet to establish itself. Those who reminded scientists of their social responsibilities were viewed with suspicion; and those who had stopped doing research were treated as defectors.
Direct URL for this post.
“It appears to me, the doing what little one can to encrease [sic] the general stock of knowledge is as respectable an object of life as one can in any likelihood pursue.”
Darwin. Letter to his sisters from the Beagle. Quoted in the London Review of Books 23-May-2019, Rosemary Hill.
Direct URL for this post.
In a 1963 letter to molecular biologist Max Perutz, he wrote, “It is now widely realized that nearly all the ‘classical’ problems of molecular biology have either been solved or will be solved in the next decade…The future of molecular biology lies in the extension of research to other fields of biology, notably development and the nervous system.”
Sydney observed, and predicted, the flow of science: “Progress depends on the interplay of techniques, discoveries, and ideas, probably in that order of decreasing importance,” he said.
Man, the toolmaker. In this particularly case, a very special one.
Sydney Brenner (1927–2019) | Science [Obit of Sydney Brenner]
Direct URL for this post.
Some traits, such as adult height, are readily measured. The heritability of this trait is ∼60 to 80%. Attempts to characterize “height genes” have resulted in the identification of tens of thousands of genes, each of which contributes a small amount to this heritability. The plethora of factors is almost inevitable, given the vast number of cellular and physiological steps involved in the development of an adult human being. A model that accounts for ∼40% of height variability predicts individual heights to within 4 cm for 50% of people, but with errors of more than 10 cm for 5%. Thus, a sophisticated genomic analysis can predict height to some extent, but not well enough for use in ordering tailored clothing. Most direct-to-consumer genomic results are based on much less detailed analyses and many involve complex traits, so considerable skepticism is appropriate.
But such sensible comments, will not stem the hype — or the investors.
Direct URL for this post.
Scope for recognizing and accommodating exceptional individuals has been diminishing in British universities ever since. Hamilton published relatively few papers, in generally low status journals, and gained only a handful of grants much later in life. Bureaucratic measures of performance are increasingly important and judge the impact of an article only by the journal it is published in. This seriously undervalues radical originality, which although extremely rare is utterly vital to science. It is disturbing that a young Bill Hamilton today would probably find an academic career even more difficult to pursue.
Alan Grafen, in his obituary of Bill Hamilton (Biogr. Mems Fell. R. Soc. Lond. 50, 109–132 (2004)).
I post this excerpt following a discussion with somebody who had never heard of him. Hamilton’s enormous contributions to biology are not well known. You also have to wonder if the lack of a Nobel for biology diminishes medicine in the long run. Some things do indeed get worse.
Direct URL for this post.
Sydney Brenner has died. Not quite the last of the handful of scientists who made one of the two scientific revolutions of the 20th century. The first half belonged to physics, the second to the biology that he co-created.
A precocious boy—a student at the University of the Witwatersrand by the time he was 15—and bullied for it, reading was his connection to the wider world. Courses, he said, never taught him anything. The way to learn was to get a book that told you how to do things, and then to start doing them, whether it was making dyes or, later in life, programming computers. If he thought more deeply than the other great biologists of his age, which he did, it was surely because he read further, too.
Reading Brenner was a staccato of insights. I hadn’t come across the ‘courses’ quote before, but no surprises there.
I like statistics and spent most of my intercalated degree ‘using’ medical stats (essentially, writing programs on an IBM 360 mainframe to handle a large dataset, that I could then interrogate using the GLIM package from the NAG). Yes, the days of batch processing and punchcards. I found — and still find — statistics remarkably hard.
I am always very wary of people who say they understand statistics. Let me rephrase that. I am very suspicious of non-professional statisticians who claim that they find statistics intuitive. I remember that it was said that even the great Paul Erdos got the Monty Hall problem wrong.
The following is from a recent article in Nature:
What will retiring statistical significance look like? We hope that methods sections and data tabulation will be more detailed and nuanced. Authors will emphasize their estimates and the uncertainty in them — for example, by explicitly discussing the lower and upper limits of their intervals. They will not rely on significance tests. When P values are reported, they will be given with sensible precision (for example, P = 0.021 or P = 0.13) — without adornments such as stars or letters to denote statistical significance and not as binary inequalities (P < 0.05 or P > 0.05). Decisions to interpret or to publish results will not be based on statistical thresholds. People will spend less time with statistical software, and more time thinking.
There is lots of blame to go around here. Bad teaching and bad supervision, are easy targets (too easy). I think there are (at least) three more fundamental problems.
Science has been thought of as a form of ‘reliable knowledge’. This form of words always sounded almost too modest to me, especially when you think how powerful science has been shown to be. But in medicine we are increasingly aware that much modern science is not a basis for honest action at all. Blake’s words were to the effect that ‘every honest man is a prophet’. I once miswrote this in an article I wrote as ‘every honest man is for profit’. Many an error….
That picture that changed everything. Nice piece in Nature tells the story. (Image: NASA)
In climate science, you can check out of the lab anytime you like, but you can never leave.
Dave Reay, University of Edinburgh, quoted in Nature this week.
This is from an interview with Geoffrey Hinton who — to paraphrase Peter Medawar’s comments about Jim Watson — has something to be clever about. The article is worth reading in full, but here are a few snippets.
Now if you send in a paper that has a radically new idea, there’s no chance in hell it will get accepted, because it’s going to get some junior reviewer who doesn’t understand it. Or it’s going to get a senior reviewer who’s trying to review too many papers and doesn’t understand it first time round and assumes it must be nonsense. Anything that makes the brain hurt is not going to get accepted. And I think that’s really bad…
What we should be going for, particularly in the basic science conferences, is radically new ideas. Because we know a radically new idea in the long run is going to be much more influential than a tiny improvement. That’s I think the main downside of the fact that we’ve got this inversion now, where you’ve got a few senior guys and a gazillion young guys.
I would make a few comments:
All has been said before, I know, but no apology will be forthcoming.
Genome-wide study of hair colour in UK Biobank explains most of the SNP heritability.
Michael D. Morgan, Erola Pairo-Castineira, Konrad Rawlik, Oriol Canela-Xandri, Jonathan Rees, David Sims, Albert Tenesa & Ian J. Jackson
[Link to Nature Comm paper] https://doi.org/10.1038/s41467-018-07691-z
My guess is this is likely my last ‘research paper’ (although I now choose to redefine what counts as research). But not my last ‘thinking paper’. I cannot help but contrast the sheer volume of activity with that from our original papers on red hair. Things seemed so much simpler when we were young. But it is a nice coda to a career fugue.
Leading universities should pledge to actually read the work of applicants for research positions rather than use controversial metrics during the selection process, a Nobel prizewinner has argued.
No, not a spoof, but words from Harold Varmus. Sydney Brenner, a good while back, observed that people tended not to read papers anymore, they just xeroxed them.
Modesty seems to be under negative selection — among modern scientists, at least. So I warmed to this comment on a report of some recent work on the genetics of Africa and hunter-gatherers.
Deepti Gurdasani, a genetic epidemiologist at the Wellcome Sanger Institute in Hinxton, UK. But it’s plausible, she adds. “There is literally nothing in Africa that is not possible since we have no idea what humans were doing on the continent 5,000 years ago.”
This is from an article in Nature.
Under pressure to turn out productive lab members quickly, many PhD programmes in the biomedical sciences have shortened their courses, squeezing out opportunities for putting research into its wider context. Consequently, most PhD curricula are unlikely to nurture the big thinkers and creative problem-solvers that society needs.
That means students are taught every detail of a microbe’s life cycle but little about the life scientific. They need to be taught to recognize how errors can occur. Trainees should evaluate case studies derived from flawed real research, or use interdisciplinary detective games to find logical fallacies in the literature. Above all, students must be shown the scientific process as it is — with its limitations and potential pitfalls as well as its fun side, such as serendipitous discoveries and hilarious blunders.
And from a letter in response
My father designed stellar-inertial guidance systems for reconnaissance aircraft and, after he retired, would often present his work to physics and engineering students. When they asked him what they should study to prepare for such a career, he would reply: “Read the classics,” by which he meant Aristotle, Ralph Waldo Emerson, Jean-Jacques Rousseau and Blaise Pascal.
The best scientific and technical progress does not come out of a box. It is more likely to emerge from trying to fit wild, woolly and tangential ideas into useful societal and economic contexts.
As the historian Norman Davies once said:
“Since no one is judged competent to offer an opinion beyond their own particular mineshaft, beasts of prey have been left to prowl across the prairie unchecked.”
Or as the Economist once put it”
“…professors fixated on crawling alone the frontiers of knowledge with a magnifying glass.”
This is the tragedy of our age: 90% right and 100% wrong. And that is even before we get to medicine.
When working in Africa in the 1980s with my good friend Victor Pretorius, I heard a legend about an important tribe in Central Africa, the Masai. The legend claimed that a genius member of the tribe in the nineteenth century or earlier had the idea that cow’s urine was the safest fluid for washing cooking utensils. Compared with the previous practice of using far from clean river water, it avoided the dangers of dysentery and probably saved many lives. This simple and effective public heath practice was cast out by medical missionaries who had quite different ideas, more religious than medical, about what was clean and what was dirty. Neither the original genius, nor the missionaries, knew anything about the epidemiology of water-borne disease. Whether or not there is any substance to this legend, it has stayed in my mind as a metaphor appropriate for many of our problems today. Inventions such as Newcomen’s steam engine, Faraday’s electrical machines, and the idea that fresh urine is a sterile fluid, all came long before their scientific understanding.
James Lovelock, A Rough Ride to the Future. This is like so much of real discovery in clinical medicine, although the academy gets to write the history of how it is supposed to work.
This is from David Hubel, although the citation is not to hand.
Most importantly, today’s organization of science tends to deprive a young scientist of one of the most important learning experiences, that of thinking up a project of one’s own and carrying it through; deciding for oneself, independently, whether to persist or to give up and switch over to something else.
I read this book so long ago I cannot remember when. But Perutz had a way with words ( as well as molecules).
What is Life? helped to make influential biologists out of several physicists: Crick, Seymour Benzer and Maurice Wilkins, among others. But there’s no indication from contemporary reviews that many biologists grasped the real significance of Schrödinger’s code-script as a kind of active program for the organism. Some in the emerging science of molecular biology were critical. Linus Pauling and Max Perutz were both damning about the book in 1987, on the centenary of Schrödinger’s birth. Pauling considered negative entropy a “negative contribution” to biology, and castigated Schrödinger for a “vague and superficial” treatment of life’s thermodynamics. Perutz grumbled that “what was true in his book was not original, and most of what was original was known not to be true even when the book was written”.
From an obituary of Paul Boyer.
“Paul Boyer was approaching the finish line of his career when he risked everything with a jaw-dropping proposal. He addressed one of the most important, as-then-unanswered questions in biochemistry”
“We were attending a UCLA seminar in 1972 when I noticed that he wasn’t paying attention to the speaker. Afterwards, Paul approached us in a very excited state. This was surprising because he was known for his calm demeanour. He confessed that he had spent the hour thinking about old unexplained data. He asked: “What would you say if I told you that it doesn’t take energy to make ATP at the catalytic site of ATP synthase,” (as was universally held at the time) “but rather that it takes energy to get ATP off the catalytic site?” This was a eureka moment.
As is often the case with transformational ideas, early reactions were negative. When the Journal of Biological Chemistry rejected our manuscript containing data supporting this concept, Boyer told me without animosity that he could see why they would do that — “It was a very striking claim.”
Well, I have never had an idea to compare with this. But sitting through talks that do not light my fire, I have always found conducive to thinking creatively about something else. Its similar to the way that some writers practice their craft better in a coffee shop than in a silent office. Intellectual white noise.
Remember: the best ideas are not in the literature. If they were…..
‘True science thrives best in glass houses where everyone can look in. When the windows are blacked out, as in war, the weeds take over; when secrecy muffles criticism, charlatans and cranks flourish’.
Max Perutz (1914-), Austrian born biochemist. Shared 1962 Nobel Prize for X-ray crystallography of haemoglobin.
These are a few words from the author of “Lost in Math: How Beauty Leads Physics Astray”, but they speak to me at least of an intellectual honesty that is (as the author argues) increasingly rare in the academy.
I am not tenured and I do not have a tenure-track position, so not like someone threatened me. I presently have a temporary contract which will run out next year. What I should be doing right now is applying for faculty positions. Now imagine you work at some institution which has a group in my research area. Everyone is happily producing papers in record numbers, but I go around and say this is a waste of money. Would you give me a job? You probably wouldn’t. I probably wouldn’t give me a job either.
What typically happens when I write about my job situation is that everyone offers me advice. This is very kind, but I assure you I am not writing this because I am asking for help. I will be fine, do not worry about me. Yes, I don’t know what I’ll do next year, but something will come to my mind.
What needs help isn’t me, but academia: The current organization amplifies rather than limits the pressure to work on popular and productive topics. If you want to be part of the solution, the best starting point is to read my book.
A quote from an earlier post I particularly like”
While the book focuses on physics, my aim is much more general. The current situation in the foundations of physics is a vivid example for how science fails to self-correct. The reasons for this failure, as I lay out in the book, are unaddressed social and cognitive biases. But this isn’t a problem specific to the foundations of physics. It’s a problem that befalls all disciplines, just that in my area the prevalence of not-so-scientific thinking is particularly obvious due to the lack of data.
I would make two observations. First, I think science is self-correcting — in the long run, at least. Just not when measured in lifetimes. Second, this takes me back to John Horgan’s book, and in particular how some domains of science are more easily corruptible that others (to be less combative, I might say, ‘less robust’). If you want to understand the modern medical research complex, you have to understand this.
And no, I wouldn’t have thought the effect was measurable. Wrong again.
From the results presented here it is clear that there has been a slow but steady decline in the frequency of certain variants in the Icelandic gene pool that are associated with educational attainment. It is also clear that education attained does not explain all of the effect. Hence, it seems that the effect is caused by a certain capacity to acquire education that is not always realized.
Fine thoughts, with words and a life to match
The departure of scientific reality from what common sense suggests is going on (the sun going round the Earth, for example) no longer threatens political institutions, but it threatens the human psyche just as much as it did in Galileo’s day. Dr Hawking’s South Pole of time was 13.7 billion years in the past—three times as old as the Earth. His mathematics showed that the universe, though finite in time, might be infinite in space.
No philosophy that puts humanity anywhere near the centre of things can cope with facts like these. All that remains is to huddle together in the face of the overwhelmingness of reality. Yet the sight of one huddled man in a wheelchair constantly probing, boldly and even cheekily demonstrating the infinite reach of the human mind, gave people some hope to grasp, as he always wished it would.
The Economist’s obit of Stephen Hawking
Article in Nature. I largely agree, although my views are as much based on the hype-upon-hype that characterises so much of medical research, especially cancer. I do not have a reference, but whatever one’s views about the late David Horrobin, his Lancet article about cancer trials — written when he was dying from lymphoma — is worth a read. What a mess!
Key quotes from this article:
In 2017, my colleagues and I completed a study of all 48 cancer drugs approved by the European Medicines Agency between 2009 and 2013 (C. Davis et al. Br. Med. J. 359, j4530; 2017). Of the 68 clinical indications for these drugs (reasons to use a particular drug on a patient), only 24 (35%) demonstrated evidence of a survival benefit at the time of approval. Even fewer provided evidence of an improved quality of life for symptoms such as pain, tiredness and loss of appetite (7 trials; 10%). Most indications (36 of 68) still lacked such evidence three or more years after approval. Other groups in other regions have observed similar trends. For example, a 2015 study demonstrated that only a small proportion of cancer drugs approved by the FDA improved survival or quality of life (C. Kim and V. Prasad JAMA Intern. Med. 175, 1992–1994; 2015).
But the key point he makes is:
I believe that the low bar also undermines innovation and wastes money.
When assessments — whether in medicine or education — are flawed the loss in value is not in short term financial costs, but in what might have happened 10 years down the road.