At the same time, there has been a growing “pull” from the UK and other richer nations for doctors and nurses from Africa, as their own health systems have struggled to train and retain sufficient local healthcare workers while demand from ageing populations continues to rise.
I am aware of the issue but keep being pulled back to the claims about how expensive it is to train doctors (in the UK or other similar countries). Yes, I know the oft wheeled out figures, but I am suspicious of them.
Awhile back I was sat in a cafe close to the university campus. I couldn’t help but listen in on the conversation of a few students who were discussing various aspect of university life, and their own involvement in student politics. I couldn’t warm to them: they were boorish and reminded me of a certain Prime Minister. But I did find myself in agreement on one point: many UK universities are too big and if you are really serious about undergraduate education, you need smaller institutions than is the norm in the Russell group. You can have large institutions and teach well — the Open University is the classic example historically — but Russell group universities are not designed for the same purpose.
A few months back there was an interview in the Guardian with Michael Arthur, the Vice Chancellor of University College, London (UCL). In it he said some extraordinary things. Not extraordinary in the sense that you have might not have heard them before, or that they were difficult to grasp. Just extraordinary in their banality of purpose.
UCL like many universities in the UK has and will continue to rapidly expand undergraduate student numbers. The interviewer asked him whether or not UCL was not already too big. Arthur replied:
“We want to be a global player,” says Arthur. “Round the world, you’re seeing universities of 90,000, 100,000 students. If you have critical mass, you can create outstanding cross-disciplinary research on things like climate change. You can do research that makes a difference.” He mentions a treatment recently developed at UCL that makes HIV, the virus that causes Aids, untransmittable. If UCL didn’t increase student numbers, thus maximising fee revenue, such research would have to be cut back. “To me,” Arthur says, “that is unthinkable.”
The tropes are familiar to those who have given up serious thinking and have short attention spans: ‘global player’, ‘critical mass’, ‘cross disciplinary’, ‘make a difference’, and so on. Then there is the ‘maximising fee revenue’ so that research is not cut back — “that is unthinkable”
Within the sector it is widely recognised that universities lose money on research. In the US in the Ivy League, endowments buffer research and in some institutions, teaching. In the UK, endowments outwith Oxbridge are modest, and student fees fund much research. As research volume and intensity increases, the need for cross subsidy becomes ever greater. This is of course not just within subjects, but across the university and faculties.
That universities lose money on research is a real problem. For instance, in medicine much research is funded by charities who do not pay the full costs of that research. Governments pretend they fill this gap, but I doubt that is now the case. Gaps in research funding are therefore being made up out of the funds that are allocated to educate doctors, or students in other subjects. And anybody who has been around UK universities for a while knows that a lot of the research — especially in medicine — would have at one time being classed as the D of R&D. This sort of work is not what universities are about: it is just that the numbers are so large that they flatter the ‘research figures’ for the REF (research excellence framework).
Pace the students in the cafe, few can mount any argument that once you have grown beyond several thousand students that the student experience and student learning worsen. Phrases such as ‘research-led teaching’ and ‘exposure to cutting edge research’ are common, but the reality is that there is little evidence to support them in the modern university. They are intended as fig leaves to mask some deeper stirrings. Arthur states that it ‘would be unthinkable’ to cut back on research. He may believe that, but I doubt if his self-righteousness is shared by the majority of students who spend much of their lives paying off student debts.
A few years ago, whilst on a flight to Amsterdam, I chatted with a physicist from a Dutch university. We talked about teaching and research. He was keen on the idea of situating institutions that resembled US liberal arts colleges (as in small colleges) within bigger and more devolved institutions. I doubt that would be practical in the UK — the temptation for the centre to steal the funds is something VCs (Vice Chancellors not Venture Capitalists, that is) would not be able to resist. The late Roger Needham, a distinguished Professor of Computing at Cambridge, and former head of Microsoft Research in Cambridge, pointed out that most IP generated by universities was trivial and that the most important IP we produced were educated and smart students. He was perhaps talking about PhDs and within certain domains of knowledge, but I will push beyond that. Educating students matters.
And contrary to what Arthur thinks many of the world’s best universities have far fewer students than UCL even before its recent metastatic spread.
Medical students have higher rates of depression, suicidal ideation, and burnout than the general population and greater concerns about the stigma of mental illness. The nature of medical education seems to contribute to this disparity, since students entering medical school score better on indicators of mental health than similarly aged college graduates. Roughly half of students experience burnout, and 10% report suicidal ideation during medical school
This is from the US, and I do not know the comparable figures for the UK. Nor as I really certain what is going on in a way that sheds light on causation or what has changed. By way of comparison, for early postgraduate training in the UK, I am staggered by how many doctors come through it unscathed. I don’t blame those who want to bail out.
Direct URL for this post.
An economist may have strong views on the benefits of vaccination, for example, but is still no expert on the subject. And I often cringe when I hear a doctor trying to prove a point by using statistics.
There were some critical comments about this phrase used by Wolfgang Münchau in a FT article. The article is about how ‘experts’ lose their power as they lose their independence. This is rightly a big story, one that is not going away, and one the universities with their love of mammon and ‘impact’ seem to wish was otherwise. But there is a more specific point too.
Various commentators argued that because medicine took advantage of statistical ideas that doctors talked sense about statistics. The literature is fairly decisive on this point: most doctors tend to be lousy at statistics, whereas the medical literature may or (frequently) may not be sound on various statistical issues.
Whenever I hear people talk up the need for better ‘communication skills’ or ‘communication training’ for our medical students, I question what level of advanced statistical training they are referring to. Blank stares, result. Statistics is hard, communicating statistics even harder. Our students tend to be great at communicating or signalling empathy, but those with an empathy for numbers often end up elsewhere in the university.
Direct URL for this post.
The main story is about an ‘anti-vaxxer’ who had informed the university that he/she was opposed to receiving any vaccinations, but the university had not noticed or acted upon this advice till after the student had started univeristy. Cardiff university were ordered to pay £9K to the anti-vaxxer healthcare student.
But this caught my eye even more.
In a separate case summary, also published on 1 July, the OIA said that it had told Wrexham Glyndwr University to compensate eight students who had complained about the quality of a healthcare-related course.
The watchdog said that the students had complained that a key part of the course had not been taught as promised, meaning that they were not given the necessary skills to practise safely. Some teaching hours were cancelled for some modules, and the group also complained about the behaviour of a staff member, who they said was “unapproachable and aggressive”.
The OIA, which ruled that the complaint was partly justified, said that Glyndwr should refund tuition fees of £2,140 to each student, and pay an additional £1,500 compensation to each of them for the inconvenience caused.
Direct URL for this post.
Our present pattern of medical education is only one of several that are operating more or less successfully at the present time: good medicine can be taught and practised under widely varying conditions.
Henry Miller. ‘Fifty Years after Flexner’, 1966.
In my last post, I used a familiar Newton quote: ‘the job of science is to cleave nature at the joints’. We can never understand the entirety of the universe, all we can do is to fragment it, in order to make it amenable to experimentation or rational scrutiny. Before you can build anything you have to have taken other familiar things apart. Understanding always does violence to the natural world.
In this series of posts I have already listed some of the many things that confound attempts to improve medical education. But I don’t think we now need just a series of bug fixes. On the contrary, we need radical change — as in a new operating system — but radical change we have had before, and there are plenty of examples that we can use to model what we want. And as I hinted at in my last post, medical exceptionalism (and in truth pride) blind medical educationalists to how other domains of professional life operate. This soul searching about professional schools is not confined to medicine. There are debates taking place about law schools  and engineering schools , and corresponding debates about the role of the professions in society more generally (have the professions a future — professional work has, but who is going to do it?) .
The conventional medical degree has two components: the preclinical years (which I used to to call the prescientific years, simply because rote learning is so favoured in them); and the clinical years. This divide has been blurred a little, but does not seriously alter my argument — the blurring has in any case been a mistake IMHO. The preclinical years have some similarities with other university courses, for good and bad. The clinical years are simply a mess. They aspire to a model of apprenticeship learning that is impossible to deliver on.
All is not lost, however. We know we can do some things well. Let me consider the ‘clinical’ first, before moving back to the ‘preclinical’.
Registrar training day in any speciality can work well. We know how to do it. There is a combination of private study, formal courses, and day-to-day supervised and graded practice. Classic apprenticeship. This doesn’t mean it is always done well — it isn’t — but in practice we know how to put it together. Let me use dermatology as an example.
In the UK and a few other countries, you enter dermatology after having done FY (foundation years 1 & 2) and a few years of internal medicine, having passed the MRCP exams along the way (the College tax). I refer to this as pre-dermatology training. At this stage, you compete nationally for training slots in dermatology.
This pre-dermatology training is unnecessary. We know this to be the case because most of the world does not follow this pattern, and seems to manage OK in terms of quality of their dermatologists. (This ‘wasted years’ period was painfully pointed out to me when I started dermatology training in Vienna: ‘you have wasted four years of your life’, I was told. I wasn’t pleased, but they were right and I was wrong). Why you ask, does the UK persist? Three explanations come to mind. First, the need for cheap labour to staff hospitals. Second, the failure to understand that staff on a training path need to supplement those who provide ‘core service’: much as senior registrars were supernumerary in some parts of the UK at one time. Finally, an inability to realise that we might learn from others.
Providing good apprenticeship training in dermatology is (in theory) very straightforward. Book learning is required, formal courses online can supplement this book learning, and since trainees are grouped in centres, interpersonal learning and discussion is easy to organise. Most importantly, trainees work with consultants, over extended periods of time, who know what they are trying to achieve: the purpose of the apprenticeship is to produce somebody who can replace them in a few years time. You do not need to be deep into educational theory to work well within this sort of environment, indeed you should keep any ‘educationalists’ at arms length.
Where this model does not work well, is in the ‘predermatology’ training. The obvious point is that much of this pre-dermatology work is not necessary and where it is, it should be carried out by those who are embarking on a particular career or by non-training staff (who may or may not be doctors). In the UK, if you have a FY doctor attached to a dermatology ward, they will rotate every few months through a range of specialties, and it is likely that they will have no affinity for most of them. Such jobs are educationally worthless as dermatology is an outpatient specialty. Ironically the only value of such jobs, is for those who have already committed to a career in dermatology. I will return to the all too familiar objections of what I propose in another blog post, but for training in many areas of medicine, including GP, radiology, pathology, psychiatry, what I have said of dermatology, holds.
We could frame my argument in another way. If you cannot hold onto the tenets of apprenticeship learning — extended periods of graded practice under the close supervision of a small group of masters and novices, it is not a training post.
I am now going to jump to the other end of medical education: what are medical schools for?
Current undergraduate medical education is a hybrid of ‘education’ and ‘training’. Universities can deliver high class education (I said can, not do), but they cannot deliver high class clinical training. They do not have the staff to do it, and they do not own the ‘means of production’. Apprenticeship learning does not work given the number of students, and in any case, teaching of medical students is a low priority for NHS hospitals who have been in a ‘subsistence’ mode for decades. Things will only get worse.
Some (but not all) other professional schools or professions organise things differently. A degree may be necessary, but the bond between degree and subsequent training is loose. Unlike medicine, it is not the job of the university to produce somebody who is ‘safe’ and ‘certified’ on the day of graduation.
What I propose is that virtually all the non-foundational learning is shifted into the early years of apprenticeship learning where the individuals are paid employees of the NHS (or other employer). I talked about what foundational learning is in an earlier post, and here I am arguing that it is the foundational learning which universities should deliver. Just as professional service firms, law firms or engineering schools may prefer graduates with particular degrees, they know that they need to train their apprentices in a work environment, an environment in which they are paid (as with all apprenticeships the training salary reflects the market value to the individual of the professional training they receive). What becomes of medical schools?
The corpus of knowledge of the determinants of health and how to promote health, as well as how to diagnose and care for those who are sick is vast. Looked at in financial terms, or numbers of workers, it is a large part of the modern economy, and is of interest way beyond the narrow craft of clinical medicine. The fundamental knowledge underpinning ‘health’ includes sciences and arts. Although modern medicine likes to ride on the coat-tails of science, it is in terms of practice, a professional domain that draws eclectically from a broad scholarship and habits of mind. Medical science has indeed grown, but as a proportion of the domains of knowledge that make up ‘health’ it has shrunk.
Simply put, we might expect many students to study ‘health’, and for the subset of those who want to become doctors we need to think about the domains that are most suitable for ‘practising doctors’. Not all who study ‘health’, will want to be ‘practising doctors’, but of those who do, there may be constraints on what modules they should take. The goal is to produce individuals who can be admitted into a medical apprenticeship when they leave university.
I will write more about ‘health’ in the next post, and contrast it with what we currently teach (and how we teach it). The later part of training (genuine apprenticeship), as in the dermatology example, I would leave alone. But what I am suggesting is that we totally change the demands put on medical schools, and place apprenticeship learning back where it belongs.
 Stolker C. Rethinking the Law School. Cambridge University Press; 2014
 Goldberg DE, Somerville M, Whitney C. A Whole New Engineer: The Coming Revolution in Engineering Education. Threejoy Associates; 2014
 Susskind RE. The end of lawyers? : rethinking the nature of legal services. Oxford; New York: Oxford University Press; 2010
 Susskind R, Susskind D. The Future of the Professions. Oxford University Press, USA; 2015
 Rees J. The UK needs office dermatologists. BMJ. 2012;345:35.
In the previous post, I talked about some of the details of how undergraduate clinical teaching is organised. It is not an apprenticeship, but rather an alienating series of short attachments characterised by a lack of continuity of teacher-pupil contact. This is not something easily fixed because the structure is geared around the needs of the NHS staff who deliver the bulk of student teaching, rather than what we know makes sense pedagogically. I likened it to the need to put up with getting through security when you travel by plane: you want to get somewhere, but just have to grin and bear the humiliation. This is not a university education. I am not saying that individual teachers are to blame — far from it — as many enjoy teaching students. It is a system problem.
It is not possible to make sense of either undergraduate medical education or postgraduate training without looking at the forces that act on the other. It is also far too easy to assume that ‘the system’ in the UK is the only way to organise things, or indeed, to think it is anywhere near optimal. A damning critique of medicine (and much else in society) in the UK is our inability to learn from what others do.
The formative influences on (undergraduate) medical education are those conditions that were operating over half a century ago. At that time, a medical degree qualified you to enter clinical practice with — for many students — no further formal study. And much clinical practice was in a group size of n=1.
In the 1950s the house year (usually 6 months surgery and 6 months medicine) was introduced. Theoretically this was under the supervision of the university, but in practice this supervision was poor, and the reality was that this was never going to work in the ‘modern NHS’. How can the University of Edinburgh supervise its graduates who work at the other end of the country? In any case, as has been remarked on many occasions, although the rationale for the house year was ‘education’, the NHS has never taken this seriously. Instead, housepersons became general dogsbodies, working under conditions that could have come from a Dickens novel. In my own health board, the link between master and pupil has been entirely broken: apprenticeship is not only absent from the undergraduate course, but has been exiled from a lot of postgraduate training (sic). House doctors are referred to as ‘ward resources’, not tied to any group of supervising doctors. Like toilet cisterns, or worse…
Nonetheless, the changes in the 1950 and other reforms in the 1960s established the conventional wisdom that the aim of undergraduate medical education was not to produce a ‘final product’ fit to travel the world with their duffel-shaped leather satchel in hand. Rather, there would be a period of postgraduate training leading to specialist certification.
This change should have been momentous. The goal was to refashion the undergraduate component; and allow the postgraduate period to produce the finished product (either in a specialty, or in what was once called general practice). It is worth emphasising what this should have meant.
From the point of view of the public, the key time for certification for medial practice was not graduation, but being placed on the specialist register. The ability to practice independently was something granted to those with higher medical qualification (MRCP, MRCPysch etc) and who were appointed to a consultant post. All other posts were training posts, and practice within such roles was not independent but under supervision. Within an apprenticeship system — which higher professional training largely should be — supervision comes with lots of constraints, constraints that are implicit in the relation between master and pupil, and which have stayed largely unchanged across many guilds and crafts for near on a thousand years.
What went wrong was no surprise. The hospitals needed a cadre of generic dogbodies to staff them given the 24 hour working conditions necessary in health care. Rather than new graduates choosing their final career destination (to keep with my airport metaphor) they were consigned to a holding pattern for 2-7 years of their life. In this service mode, the main function was ‘service’ not supervised training. As one of my former tutees in Edinburgh correctly told me at graduation: (of course!)he was returning to Singapore, because if he stayed in the NHS he would just be exploited until he could start higher professional training. The UK remains an outlier worldwide in this pattern of enforced servitude.
The driving force in virtually all decision making with the UK health systems is getting through to the year-end. The systems live hand-to-mouth. They share a subsistence culture, in which it almost appears that their primary role is not to deliver health care, but to reflect an ideology that might prove attractive to voters. As with much UK capitalism, the long term always loses out to the short term. What happened after the realisation that a graduating medical students was neither beast nor fowl, was predictable.
The pressure to produce generic trainees with little meaningful supervision in their day-to-day job, meant that more and more of undergraduate education was sacrificed to the goal of producing ‘safe and competent’ FY (foundation years 1 & 2) doctors, doctors who again work as dogsbodies and cannot learn within a genuine apprenticeship model. The mantra became that you needed five years at medical school, to adopt a transitory role, that you would willingly escape from as soon as possible. Furthermore the undergraduate course was a sitting duck for any failings of the NHS: students should know more about eating disorders, resilience, primary care, terminal care, obesity, drug use… the list is infinite, and the students sitting ducks, and the medical schools politically ineffective.
What we now see is an undergraduate degree effectively trying to emulate a hospital (as learning outside an inpatient setting is rare). The problem is simply stated: it is not possible to do this within a university that does not — and I apologise if I sound like an unreconstructed Marxist — control the means of production. Nor is it sensible to try and meld the whole of a university education in order to produce doctors suitable for a particular time-limited period of medical practice, that all will gladly leave within a few years of vassalage.
Medicine is an old profession, (I will pass on GBS’ comments about the oldest profession). In medicine the traditional status of both ‘profession’ and ‘this profession’ in particular has been used to imagine that medicine can stand aloof from other changes in society. There are three points I want to make on this issue: two are germane to my argument, whilst the other, I will return to in another post.
The first is that in the immediate post-Flexner period to the changes in medical education in the 1950s and 1960s, few people in the UK went to university. Doctors did go to university even if the course was deemed heavily vocational, with a guaranteed job at the end of it. Learning lots of senseless anatomy may not have compared well with a liberal arts eduction but there was time for maturing, and exposure to the culture of higher learning. Grand phrases indeed, but many of us have been spoiled by their ubiquity. Our current medical students are bright and mostly capable of hard work, but many lack the breadth and ability to think abstractly of the better students in some other faculties. (It would for instance, be interesting to look at secular changes in degree awards of medical students who have intercalated.) No doubt, medical students are still sought after by non-medical employers, but I suspect this is a highly self-selected group and, in any case, reflects intrinsic abilities and characteristics as much as what the university has provided them with.
The second point, is that all the professions are undergoing change. The specialist roles that were formalised and developed in the 19th century, are under attack from the forces that Max Weber identified a century ago. The ‘terminally differentiated’ individual is treated less kindly in the modern corporate state. Anybody who has practiced medicine in the last half century is aware of the increasing industrialisation of medical practice, in which the battle between professional judgment and the impersonal corporate bureaucracy is being won by the latter 
My third point is more positive. Although there have been lots of different models of ‘professional training’ the most prevalent today is a degree in a relevant domain (which can be interpreted widely) following by selection for on the job training. Not all those who do a particular degree go onto the same career, and nor have the employers expected the university to make their graduates ‘fit for practice’ on day 1 of their employment. Medicine has shunned this approach, still pretending that universities can deliver apprenticeship training, whilst the GMC and hospitals have assumed that you can deliver a safe level of care by offloading core training that has to be learned in the workplace, to others. No professional services firm that relies on return custom and is subject to the market would behave in this cavalier way. Patients should not be so trusting.
In the next post, I will expand on how — what was said of Newton — we should cleave nature at the joints in order to reorganise medical education (and training).
 Re; the enforced servitude. I am not saying this work is not necessary, nor that those within a discipline do not need to know what goes on on the shop floor. But to put it bluntly, the budding dermatologist should not be wasting time admitting patients with IHD or COPD, or inserting central lines or doing lumbar punctures. Nor do I think you can ethically defend a ‘learning curve’ on patients given that the learner has committed not to pursue a career using that procedure. The solution is obvious, and has been discussed for over half a century: most health care workers need not be medically qualified.
 Which of course raises the issue of whether certification at an individual rather than an organisational level makes sense. In the UK the government pressure will be to emphasise the former at the expense of the latter: as they say, the beatings will continue until moral improves.
 Rewards in modern corporations like the NHS or many universities are directed at generic management skills, not domain expertise. University vice-chancellors get paid more than Nobel prize winners at the LMB. In the NHS there is a real misalignment of rewards for those clinicians who their peers recognise as outstanding, versus those who are medical managers (sic). If we think of some of the traditional crafts — say painting or sculpture – I doubt we can match the technical mastery expertise of Florence. Leonardo would no doubt now by handling Excel spreadsheets as a manager (see this piece on Brian Randell’s homepage on this very topic).
In the previous post I laid out some of the basic structures of the ‘clinical years’ of undergraduate medical degrees. In this post I want to delve a little deeper and highlight how things have gone wrong. I do not imagine it was ever wonderful, but it is certainly possible to argue that things have got a lot worse. I think things are indeed bad.
When I was a medical student in Newcastle in 1976-1982 the structure of the first two clinical years (years 3 and 4) were similar, whereas the final year (year 5) was distinct. The final year was made up of several long attachments — say ten weeks medicine and 10 weeks surgery — and there were no lectures or any demands on your time except that you effectively worked as an unpaid houseman, attached to a firm of two or three consultants. The apprenticeship system could work well during these attachments. The reasons for this partly reflected the fact that all parties had something to gain. Many if not most students chose where they did their attachments (‘if you like fellwalking, choose Carlisle etc), and had an eye on these units as a place to do your house jobs the following year. The consultants also had skin in the game. Instead of relying on interviews, or just exam results, they and all their staff (junior docs, nurses etc) got a chance to see close up what an individual student was like, and they could use this as a basis for appointing their houseperson the following year. If a houseman was away, you acted up, and got paid a small amount for this. At any time if you didn’t turn up, all hell would break out. You were essential to the functioning of the unit. No doubt there was some variation between units and centres, but this is how it was for me. So, for at least half of final year, you were on trial, immersed in learning by doing / learning on the job / workplace learning etc. All the right buzzwords were in place.
As I have said, years 3 and 4 were different from final year, but similar to each other. The mornings would be spent on the ward and the afternoons — apart from Wednesdays — were for lectures. I didn’t like lectures (or at least those sort of lectures) so I skipped them apart from making sure that I collected any handouts which were provided on the first day (see some comments from Henry Miller on lectures below ).
The mornings were ‘on the wards’. Four year 3 students might be attached to two 30 bedded wards (one female, one male), and for most of the longer attachments you would be given a patient to go and see, starting at 9:30, breaking for coffee at 10:30 and returning for an hour or more in which one or more of you had to present you findings before visiting the bedside and being taught how to examine the patient. The number of students was small, and there was nowhere to hide, if you didn’t know anything.
For the longer attachments (10 weeks for each of paediatrics, medicine and surgery) this clinical exposure could work well. But the shorter attachments especially in year 4 were a problem, chiefly because you were not there long enough to get to know anybody.
The design problem was of course that the lectures were completely out of synchrony with the clinical attachments. You might be doing surgery in the morning, but listening to lectures on cardiology in the afternoon. Given my lack of love for lectures, I used the afternoons to read about patients I had seen in the morning, and to cover the subject of the afternoon lectures, by reading books.
I don’t want to pretend that all was well. It wasn’t. You might turn up to find that nobody was available to teach you, in which case we would retreat to the nurses canteen to eat the most bacon-rich bacon sandwiches I have ever had the pleasure of meeting (the women in the canteen thought all these young people needed building up with motherly love and food 🙂 ).
The knowledge of what you were supposed to learn was, to say the least, ‘informal’; at worst, anarchic. Some staff were amazingly helpful, but others — how shall I say — not so.
In reality, everybody knew that years 3 and 4 were pale imitations of year 5. The students wanted to be in year 5, because year 5 students — or at last most year 5 students — were useful. The problem was that the numbers (students and patients) and the staffing were not available. It was something to get through, but with occasional moments of hope and pleasure. Like going through security at airports: the holiday might be good, but you pay a price.
The easiest way to summarise what happens now is to provide a snapshot of teaching in my own subject at Edinburgh.
Year 4 (called year 5 now, but the penultimate year of undergraduate medicine) students spend two weeks in dermatology. Each group is made up of 12-15 students. At the beginning of a block of rotations lasting say 18 weeks in total, the students will have 2.5 hours of lectures on dermatology. During the two week dermatology rotation, most teaching will take place in the mornings. On the first morning the students have an orientation session, have to work in groups to answer some questions based on videos they have had to watch along with bespoke reading matter, and then there is an interactive ‘seminar’ going through some of the preparatory work in the videos and text material.
For the rest of the attachment students will attend a daily ‘teaching clinic’, in which they are taught on ‘index’ patients who attend the dermatology outpatients. These patients are selected from those attending the clinic and, if they agree, they pass through to the ‘teaching clinic’. The ‘teacher’ will be a consultant or registrar, and this person is there to teach — not to provide clinical care during this session.
Students will also sit in one ‘normal’ outpatient clinic as a ‘fly on the wall’, and attend one surgical session. At the end of the attachment, there is a quiz in which students attempt to answer questions in small groups of two or three. They also get an opportunity to provide oral feedback as well as anonymous written feedback. Our students rate dermatology highly in comparison with most other disciplines, and our NHS staff are motivated and like teaching.
When I read through the above it all sounds sort of reasonable, except that…
Students will pass though lots of these individual attachments. Some are four weeks long but many are only 1 or 2 weeks in duration. It is demanding to organise such timetables, and stressful for both students and staff
My critique is not concerned with the individuals, but the system. It is simply hard to believe that this whole edifice is coherent or designed in the students’ interest. It is, as Flexner described UK medical school teaching a century ago, wonderfully amateur. Pedagogically it makes little sense. Nor in all truthfulness is it enjoyable for many staff or many students. Each two weeks a new batch will arrive and groundhog days begins. Again. And again. And if you believe the figures bandied about for the cost of medical education, the value proposition seems poor. We could do better: we should do better.
 Lectures. Henry Miller, who was successively Dean of Medicine and Vice Chancellor at Newcastle described how…
“Afternoon lectures were often avoided in favour of the cinema. The medical school was conveniently placed for at least three large cinemas….in one particularly dull week of lectures we saw the Marx brothers in ‘A Day at the Races’ three times.”
In the previous post in this series (Late night thoughts #5: Foundations) I wrote about the content or material of medical education, hinting at some of the foundational problems (pardon the meta). We have problems distinguishing between knowledge that is essential for some particular domain of medical practice, and knowledge that is genuinely foundational. The latter is largely speciality independent, less immediate than essential knowledge, and is rightly situated within the university. The expertise necessary to teach foundational knowledge lies within universities.
What I have not made explicit so far in this essay is also important. The best place to learn much essential knowledge is within the hospital, and during a genuine apprenticeship. There are various ways we can hone a meaningful definition or description of apprenticeship but key is that you are an employee, that you get paid, and you are useful to your employer. Our current structures do not meet any of these criteria.
Kenneth Calman in the introduction to his book ‘Medical Education’ points out that medical education varies enormously between countries, and that there is little evidence showing the superiority of any particular form or system of organisation. It is one of the facts that encourages scepticism about any particular form, and furthermore — especially in the UK — leads to questioning about the exorbitant costs of medical education. It also provides some support for the aphorism that most medical students turn into excellent doctors despite the best attempts of their medical schools.
Across Europe there have been two main models of clinical training (I am referring to undergraduate medical student training, not graduate / junior doctor training). One model relies on large lectures with occasional clinical demonstrations, whereas the UK system — more particularly the traditional English system — relies on ‘ clerkships’ on the wards.
At Newcastle when I was a junior doctor we used to receive a handful of German medical students who studied with us for a year. They were astonished to find that the ‘real clinical material’ was available for them to learn from, with few barriers. They could go and see patients at any time, the patients were willing, and — key point— the clinical material was germane to what they wanted to learn. The shock of discovering this veritable sweetshop put some of our students to shame.
The English (and now UK) system reflects the original guiding influence of the teaching hospitals that were, as the name suggests, hospitals where teaching took place. These hospitals for good and bad were proud of their arms length relationship with the universities and medical schools. The signature pedagogy was the same as for junior doctors. These doctors were paid (poorly), were essential (the place collapsed if they were ill), and of course they were employees. Such doctors learned by doing, supplemented by private study using textbooks, or informal teaching provide locally within the hospital or via the ‘Colleges’ or other medical organisations. Whatever the fees, most learning was within a not-for-profit culture.
It was natural to imagine or pretend that what worked at the postgraduate level would work at the undergraduate level, too. After all, until the 1950s, medical education for most doctors ended at graduation where, as the phrase goes, a surgeon with his bag full of instruments ventured forth to the four corners of the world.
This system may have worked well at one stage, but I think it fair to say it has been failing for nearer a century than half a century. At present, it is not a system of education that should be accepted. There are two reasons for this.
First, medicine has (rightly) splintered into multiple domains of practice. Most of the advances we have seen over the last century in clinical medicine reflect specialisation, specialisation as a response to the growth of explicit knowledge, and the realisation that high level performance in any craft relies not solely on initial certification, but daily practice (as in the ‘practice of medicine’). Second, what might have worked well when students and teachers were members of one small community, fails within the modern environment. As one physician at Harvard / Mass General Hospital commented a few years back in the New England Journal of Medicine: things started to go awry when the staff and students no longer ate lunch together.
Unpicking the ‘how’ of what has happened (rather than the ‘why’ which is, I think obvious), I will leave to the next post. But here is a warning. I first came across the word meliorism in Peter Medawar’s writing. How could it not be so, I naively thought? But of course, historians or political scientists would lecture me otherwise. It is possible for human affairs to get worse, even when all the humans are ‘good’ or at least have good intentions. The dismal science sees reality even more clearly: we need to only rely on institutions that we have designed to work well — even with bad actors.
We sought out an examiner who would understand that anatomy was being taught as an educational subject and not simply for the practice of surgery. I thought I had found such a man in an old colleague. I listened while he asked the student to name the successive branches emerging from the abdominal aorta in a cadaver. When we got to the inferior mesenteric he asked what viscera were supplied by that vessel. The student gave a complete and correct answer but did not know the exact amount of the rectum supplied. The examiner asked me what I thought and I said that I thought he was very good, that the only question he had missed was the last one, which in my opinion, was trivial. No, said the anatomist, by no means trivial. You have to know that before you can excis the rectum safely.
My mind still boggles at the thought of a newly graduated doctor undertaking the total excision of the rectum on the faint remembrance of the anatomy he learned as a student.
George Pickering, “Quest for Excellence in Medical Education: A Personal Survey”
When I was a medical student I read this book by Sir George Pickering. It was published in 1978, and I suspect I read it soon after the Newcastle university library acquired it. Why I came across it I do not know, but at the time ‘new volumes’ were placed for a week or two on a shelf adjacent to the entrance, before being assigned their proper home (or ‘final resting place’). It was a way to find things you didn’t know you might enjoy. I liked this book greatly, and have returned it on many occasions. Parts of it are wonderfully dated (and charming), but it remains a wonderful young man’s book written by an old man. Now I am an old man, who read it first as a young man.
Roger Schank summarise the problems of education this way:
There are only two things wrong with the education system:
- What we teach, and
- How we teach it
George Pickering’s quote relates to ‘what we teach’ — or at least what we expect students to know — but in clinical medicine ‘what we teach’ and ‘how we teach’ are intimately bound together. This may be true for much education, but the nature of clinical exposure and tuition in clinical medicine imposes a boundary on what options we can explore. The other limit is the nature of what we expect of graduates. People may think this is a given, but it is not. If you look worldwide, what roles a newly qualified doctor is asked to fill vary enormously (something I discovered when I worked in Vienna).
Here is another quote, this time from the philosopher, Ian Hacking, who has written widely on epistemology, the nature of causality and the basis of statistics (and much else).
Syphilis is signed by the market place where it is caught; the planet Mercury has signed the market place; the metal mercury, which bears the same name, is therefore the cure for syphilis.
Ian Hacking | The Emergence of Probability
Well, of course, this makes absolutely no sense to the modern mind. We simply do not accept the validity of the concept of entities being ‘signed‘ as a legitimate form of evidence. But no doubt medical students of the time would have been taught this stuff. Please note, those priests of Evidence Based Medicine (EBM), that doctors have always practiced Evidence Based Medicine, it is just that opinions on what constitutes evidence change. Hacking adds:
He [Paracelsus] had established medical practice for three centuries. And his colleagues carried on killing patients.
I am using these quotes to make two points. The first, is that there is content that is correct, relevant to some clinical practice and which medical students do not need to know. This may seem so obvious that it is not necessary to say it. But it is necessary to say it. Pickering’s example has lots of modern counterparts. We could say this knowledge is foundational for some medical practice, but foundational is a loaded term, although to be fair I do not know a better one. The problem with ‘foundational’ is that it is widely used by academic rent seekers and future employers. Students must know this, students ‘must’ know X,Y and Z. I once started to keep a list of such demands, but Excel spreadsheets have limits. You know the sort of thing: ethics, resilience, obesity, child abuse, climate change, oral health, team building, management, leadership, research, EBM, professionalism, heuristics and biases etc. Indeed, there is open season on the poor undergraduate, much of which we can lay blame for at the doors of the specialist societies and the General Medical Council (GMC).
My second point, stemming from the second quote, is to remind that much of what we teach or at least ask students to know is wrong. There is a feigned ignorance on this issue, as though people in the past were stupid, whereas we are smart. Yes, anatomy has not changed much, and I am not chucking out all the biochemistry, but pace Hacking, our understanding of the relation between ‘how doctors work’ and ‘what underpins that knowledge’ is opaque. We can — and do — tell lots of ‘just-so’ stories that we think explain clinical behaviour, that have little rational or experimental foundation. Clinicians often hold strong opinions on how they arrive at particularly decisions: there is a lot of data to suggest that whilst you can objectively demonstrate clinical expertise, clinicians often have little insight into how they actually arrive at the (correct) diagnosis (beyond dustbin concepts such as ‘pattern recognition’ or ‘clinical reasoning’).
If you are a dermatologist, and you wish to excise a basal cell carcinoma (BCC, a common skin cancer) from the temple, you need to be aware of certain important anatomical structures (specifically the superficial temporal artery, and the temporal branch of the facial nerve). This knowledge is essential for clinical practice. It is simple to demonstrate this: ask any surgeon who operates in this area. Of course, if you are a lower GI surgeon, this knowledge may not be at your finger tips. Looked at the other way, this knowledge is in large part specialty specific (or at least necessary for a subset of all medical specialties). What happens if you damage these structures is important to know, but the level of explanation is not very deep (pardon the pun). If you cut any nerve, you may get a motor or sensory defect, and in this example, you may therefore get a failure in frontalis muscle action.
This knowledge is not foundational because it is local to certain areas of practice, and it does not form the basis or foundation of any higher level concepts (more on this below). The Pickering example, tells us about what a GI surgeon might need to know, but not the dermatologist. Their world views remains unrelated, although the I prefer the view of the latter. There is however another point. We should be very careful about asking medical students to know such things. So what do we expect of them?
I find the example of anatomical knowledge as being essential compelling. But only in terms of particular domains of activity. Now, you may say you want students to know about ‘joints’ in general, and there may well be merit in this (Pickering, I suspect, thought so), but knowing the names of all the bones in the hand or foot is not essential for most doctors. If we move beyond ‘essential’ what is left?
At one time anatomy was both essential and foundational. And I am using the term foundational here to mean those concepts that underpin not just specialty specific medicine, but medicine in the round. A few examples may help.
Whatever branch of medicine you practice, it is hard to do so without some knowledge of pharmacology. How deep you venture , is subject to debate, but we do not think knowing the doses and the drug names in the BNF is the same as knowing some pharmacology.
Another example. I would find it very hard to converse with a dermatologist colleague without a (somewhat) shared view of immunology or carcinogenesis. Every sentence we use to discuss a patient, will refer and make use of concepts that we use to argue and cast light on clinical decisions. If you want to explain to a patient with a squamous cell carcinoma (SCC) who has had an organ transplant why they are at such increased risk of tumours, it is simply not possible to have a meaningful conversation without immunology or carcinogenesis (and in turn, genetics, virology, and histopathology). And for brevity, I am putting to one side, other key domains such as behaviour and behaviour modification, ethics, economics and statistics etc.
To return to my simple anatomical example of the excision of the BCC. The local anatomy is essential knowledge, but it is not foundational. What is foundational is knowing what might happen if you cut any nerve.
Let me try and put the above in the context of how we might think about medical education and medical training.
Foundational knowledge is specialty (and hence career) independent. Its function is to provide the conceptual framework that underpin much clinical practice. This not to say that the exact mix of such knowledge applies to all clinical domains, but we might expect most of it to be familiar to most doctors. But none of it will, years later, have the same day-to-day immediacy of ‘essential knowledge’ — think of my example of the temporal branch of the facial nerve for the dermatologist excising facial tumours on a weekly basis.
In this formulation, the core purpose of undergraduate medical education is to educate students in such knowledge. The purpose is not therefore to produce doctors at graduation who are ‘just not very good doctors’ but graduates who are able to pursue specialty training and make sense of the clinical world around them. The job of a medical school is to produce graduates who can start clinical training in an area of their choice. They are now in a position to — literally — understand the language of the practising doctors that surround them. They are not mini-doctors, but graduates, embarking on a professional career.
By contrast most specialty knowledge is not foundational, but essential for those within that specialty — not medical students. If you learn dermatology, you might come across things that help you learn respiratory medicine or cardiology but to be blunt, not very often. Specialties are not foundational domains of knowledge. You do not need to know dermatology to understand cardiology or vice versa.
The best place to learn the ‘foundations’ are universities. Anatomy, again may be an exception, but if you want to learn immunology, genetics, statistics or psychology you have, I think, no alternative. Hospitals simply cannot provide this.
On the other hand, using Seymour Papert’s metaphor, if you to want learn French you should go to Frenchland, if you want to learn maths, you should go to Mathland and if you want to learn doctoring, you need to go to doctorland. Medical schools are not the place to learn how to find you way around doctorland — how could they be?
NB: I will use the epithet TIJABP, but as subsequent posts will confirm, I am serious.
One of the problems in learning clinical medicine is the relation between an overall schema of what you have to learn and the detail of the various components that make up the schema. I can remember very early in my first clinical year, seeing a child with Crohn’s disease, and subsequently trying to read a little about this disorder. My difficulty was that much of what I read, contrasted Crohn’s with various other conditions — ulcerative colitis, Coeliac and so on. The problem was that I didn’t know much about these conditions either. Where was I too start? A wood and the trees, issue.
I have, pace Borges written about maps and learning before. This is my current riff on that theme. I am going to use learning how to find your way around Edinburgh as my example. There is a simple map here.
The centre of Edinburgh is laid out west to east, with three key roads north of the railway station. You can imagine a simple line map — like a London underground map — with three parallel main roads: Prince’s street, George Street and Queen street. You can then add in a greater level of detail, and some arterial routes in and out of the city centre.
If you were visiting Edinburgh for the first time, you could use this simple schema to try and locate places of interest. If you were lost and asked for help, it night prove useful. You could of course remember this simple plan — which is the most northerly of these three streets and so on — or perhaps use a simple cognitive prosthesis such as a paper map.
Students learn lots of these maps when they study medicine, because they are asked to find their way around lots of cities. They also forget many of them. The more complete the map, the harder it is to recall. If they have to navigate the same terrain most days, their recall is better. No surprises there. If you challenge a student you can literally see them reproducing the ‘map tool’ as they try and answer your question. Just like if you ask them the causes of erythema nodosum, you can literally see them counting their list on their fingers.
There are obvious differences between novices and experts. Experts don’t know need to recall the maps for multiple cities, instead they reside in the city of their specialty. Experts also tend not be good at recalling long lists of the causes of erythema nodosum, rather they just seem to recall a few that are relevant in any particular context. The map mataphor provides clues to this process.
If you challenge experts they can redraw the simple line figure that I started this piece with. They can reproduce it, although as the area of coverage is increased I suspect their map may begin to break the rules of 2D geometry: they move through the city professionally, but they are not professional cartographers.
The reason for this failure is that experts do not see the ‘line map’ in the mind’s eye, but actually see the images of the real geography in their mind as they move through it. They can deduce the simple line graph, but this is not what they use diagnostically to find their way around. By contrast, they see the images of the roads and building and can navigate based on those images. They have their own simulation, that they can usually navigate without effort. Of course, when they first visited Edinburgh, they too probably crammed a simple line graph, but as they spent time in the city, this simple cognitive tool, was replaced by experience.
This sort of way of thinking was AFAIK first highlighted by the US philosophers Dreyfus and Dreyfus. They pointed out novices use ‘rule based’ formal structures, whereas experts did not. This is obvious in craft based perceptual subjects such as dermatology (or radiology or histopathology). Experts don’t use check list to diagnose basal cell carcinomas or melanoma, they just compare what they see with a personal library of exemplars. The cognitive basis for this ability, taking advantage of the idea of ‘familial likeness’, has been studied for a long time, although I do not think the problem is solved in any sort of formal way. It is usually very fast — too fast for the explicit scoring methods promoted by most clinicians and educators.
Although this way of thinking is easiest to appreciate in perceptual subjects such as dermatology, most clinicians do not view things this way — even when the experimental evidence is compelling. Some believe the explicit rules they use to teach students, are how they do it themselves. Others believe that experts are fluent in some high level reasoning that students to not possess . They like to think that their exams can test this higher level ‘deep’ reasoning. I think they may be mistaken.
There are some ideas that follow from my story.
Clayton Christensen gets mixed press: he cannot be accused of not pushing his ideas on ‘disruption’ to — well — disruption. So, his long history of predicting how a large number of universities will be bankrupt in a few years due to ‘innovation’ and ‘digital disruption’ I take with a pinch of salt (except I would add: an awful lot should be bankrupt). But I am glad I have read what he writes, and what he says in the following excepts from an interview makes sense to me:
Fortunately, Christensen says that there is one thing that online education will not be able to replace. In his research, he found that most of the successful alumni who gave generous donations to their alma maters did so because a specific professor or coach inspired them.
Among all of these donors, “Their connection wasn’t their discipline, it wasn’t even the college,” says Christensen. “It was an individual member of the faculty who had changed their lives.”
“Maybe the most important thing that we add value to our students is the ability to change their lives,” he explained. “It’s not clear that that can be disrupted.”
We know several factors that are dramatically important in promoting learning in university students: the correct sort of feedback, and students who understand what feedback is about (and hence can use it); and close contact. Implicit in the latter is that there is continued contact with full time staff. When stated like this it is easy to understand why the student experience and faculty guided learning is so poor in most UK medical schools. The traditional way of giving timely feedback has collapsed as the ward / bedside model of teaching has almost disappeared; and teaching is horribly fragmented because we have organised teaching around the working lives of full time clinicians, rather than what students need (or what they pay for). When waiting times are out of control, when ‘bodies’ are queued up on trolleys, and when for many people getting a timely appointment to see a NHS doctor is impossible, it is self evident that a tweak here and there will achieve very little. Without major change things will get much worse.
When MIT under Chuck Vest put all of their coursewhere on line it merely served to illustrate that the benefits of MIT were not just in the materials, but in ‘being there’. And ‘being there’ is made up on other students, staff, and the interactions between these two groups.
Medical schools were much smaller when I was a medical student (1976-1982). Nevertheless, there was remarkably little personal contact, even then. Lectures were to 130+ students, and occasional seminars were with groups of 10-12. Changing perspective, students did recognise the Dean of Medicine, and could name many of the lecturers who taught them. Integration of the curriculum had not totally disrupted the need for a course of lectures from a single person, and the whole environment for learning was within a physical space that was — appropriately enough — called a medical school: something obvious to the students was that research and teaching took place in the same location. For the first two years, with one possible exception, I was fairly confident that nobody knew my name. If a student passed a lecturer in the street, I doubt if the lecturer would recognise the student, let alone be able to identify them by name.
Two members of staff got to know me in the first term of my opening clinical year (year 3): Nigel Speight, a ‘first assistant’ (senior registrar / lecturer) in paediatrics; and Sam Shuster, the Professor of Dermatology in Newcastle, who I started a research project with. For paediatrics, I was one of four junior students attached to two 30-bedded-wards, for ten weeks. It was very clear that Nigel Speight was in charge of us, and the four of us were invited around to his house to meet his kids and his wife. It was interesting in all sorts of ways — “home visits” as we discovered in general practice, often are — but I will not go into detail here.
Sam invited me around for an early evening dinner and I met his wife (Bobby), and we talked science, and never stopped — except to slag off Margaret Thatcher, and Milton Friedman. Meeting Sam was — using Christensen’s phrase — my ‘change of life’ moment. As I have written elsewhere, being around Sam, was electric: my pulse rate stepped up a few gears, and in one sense my cortical bradycardia was cured.
There are those who say that meaningful personal contact is impossible in the modern ‘bums on seats’ research university. I do not agree, although it is not going to happen unless we create the necessary structures, and this does not involve bloody spreadsheets and targets. First, even in mega-universities like the Open University, with distance learners, it was shown to be possible. Second, in some collegial systems, close personal contact (and rapid verbal feedback!) is used to leverage a lot of private study from students. In the two years I did research under Sam’s supervision (as an undergraduate — not later when I worked for him as a full time researcher), I doubt that I spent more than six hours one-to-one with him.
How you leverage staff time to promote engagement and learning is the the single most important factor in giving students what they need (and often what they want, once they know what that is ). We will continue to fail students until we realise what we have lost.
We have a very clear idea of how apprenticeship has worked over the last nine hundred years or so within Europe. The core ideas are of course much older, and the geography wider. But we have written records of the creation of the various social structures that led to the rapid changes in society that led in turn via the Renaissance to the Enlightenment and modern capitalism. We can trace so many of the norms that have guided my professional life: Royal Colleges, corporations, guilds, “masters and apprentices”, universities, certification and the granting of monopoly, and ‘professionalism’, to name but a few.
Apprenticeship is a powerful pedagogical model, but one that can only take place when a number of conditions are met. In medicine the ‘apprentice’ model is widely discussed, assumed, and contrasted with the ‘bums on seats’ lecture, the latter, the now signature pedagogy of the modern ‘massified’ university. It is also used to justify the high costs of training of education in medicine and some craft university courses.
At the level of higher professional training in medicine (or in the training of research academics) apprenticeship still can work well. There is an asymmetry between master and pupil (the master does know best, but cannot always justify why he knows best); long term interaction between both parties is required; and, at its best the pupils will model their behaviours on the master. Apprenticeship is not passive — it is not ‘shadowing’ (although a period of shadowing may be required); it will require the pupil to undertake tasks that can be observed and critiqued — you cannot learn complicated tasks based on passive observation. Chimps are highly intelligent, and yet learning to crack nuts using stones takes years and years, not because the young chimps do not watch their mothers, but because the mothers never watch (and hence correct) the young chimps. This requirement is not just required for motor tasks but for any complicated set of ‘thinking’ procedures that require accuracy and fluency. In medicine, surgeons are ahead of physicians on this, and have been for a long time.
In medieval times, becoming a master meant more than being a ‘journeyman’ — the level of professional expertise was greater, and it was recognised that teaching required another level of competence, and breadth. The master is not one step ahead on the way to perfection, but several. We prefer those teaching ‘A’ level physics, to have more than an ‘A’ level in physics themselves. And whatever domain expertise a master possesses, we know that experience of the problems or difficulties learners face, is important.
Still, in comparison with say school teaching the demands on the master (with regard to being a ‘professional’ educator) are modest. They know the job — they do not need to check out the syllabus — as they are effectively training people to do the same job they do day-to-day. They probably also have little need of theory and, in a sensible system, their reputation may be accurate.
In higher professional training in medicine, apprenticeship is still possible — it is just that it is harder than it once was (as to why, that is for another day). Similarly, at one time higher education was in large part viewed as a type of apprenticeship. Students were not staff, but they were not treated as schoolchildren, rather they were —at best— viewed as co-producers of knowledge within a university. If you were studying physics, the goal was to get you to approach the world like a physicist might. This may persist in a few institutions for a minority of students, but it is not the norm anymore.
In undergraduate medicine apprenticeship died a long time ago, although its previous health may well have been exaggerated. There is little long term personal interaction, with students passed around from one attachment to another, with many of the students feeling unwanted (‘burden of teaching’, ‘teaching load‘ etc). Staff and students can walk past each other in the street, none the wiser. Apprentices are — by definition — useful. It is this utility that underpinned the business model that formalised training and acceptance or rejection into the guild. But sadly — through no fault of their own – medical students are rarely useful. If they were useful they would be paid: they are not. Historically, students might have got paid to cover house officer absences (I did), but that world no longer exists. Nor are we able to return to it.
Whereas the master has an implicit model of the goals of training, that is no longer the case in undergraduate education, in which literally 500 or individuals are engaged in educating students for roles that they individually have little knowledge of. Instead of personal interaction, over a long time period, based on a common world view, medical schools create complicated management systems to process students, with the predictable lack of buy-in from those who are doing the educating.
There is a deeper point here. Much though a lot of UK postgraduate medical training is poor, it is possible to improve it within a framework that is known to work. Many doctors know how to do it (although the same cannot be said of the NHS). Undergraduate medical education is in a different place (like much of university education). At graduation, you step form one world into another, but just as with caterpillars and butterflies, the structures and environment we need to create are very different.
Not the word I usually associate with student descriptions of their emotional state on being taught (except after the exam). Sadly. But the word featured in a teaching management meeting today. Made me smile.
Contrast this with the quote from a book on reforming engineering education, “A Whole New Engineer”
Go into the bathrooms at the Massachusetts Institute ofTechnology (MIT) and you will see an acronym scrawled on the walls of the stalls: IHTFP. It means “I Hate This F** king Place.” (IHTFP is also found in the service academies and other elite engineering programs.) Whether this remains the true sentiment of MIT students today or merely a tradition handed down from generation to generation isn’t clear….
Direct URL for this post.
Today’s (Scottish) Daily Telegraph ran with a story about the shortage of paediatricians in Scotland. The Herald had a similar story, too. It is not just paediatrics that has major shortages. The same can be said about dermatology, radiology and a host of other areas of medicine. And that is not to mention GP land, which normally seems to attract most ‘government’ attention.
I find none of this surprising. The NHS has long been in subsistence mode, eating the seed corn (or to use that other phrase, ‘eating its young’), spending its moral and cultural capital at an alarming rate. Management is notable by its absence, whereas the administrators think they are ‘managers’, in part, because they can’t administrate and stay sane. By lack of management I meant those functions of management we see in most corporations or freestanding institutions. Changes in demography have not happened suddenly; the relation between age and health care provision, has been well known for a century or more; the impact of family structure and geography on care provision of elderly relatives evident since the early 1960s; changes in work force have been growing for at least 40 years; and UK medicine has a long history of ignoring why people wish to leave either the UK (or want to leave the NHS). The attempt to run health care as a Taylor-like post-industrial service industry using staff who value their autonomy and professionalism, may not end well for doctors — or patients.
All the above, management should have been grappling with over the last quarter century: instead they have been AWOL. Meanwhile, politicians engage in speculative future-selling, where electoral vapourware is often a vehicle for the maintenance of political power. Given the state of UK politics (as in the BxxxxT word), it seems reasonable not to give politicians the benefit of the doubt any more. As individuals, no doubt, most of them mean well, and love their kids etc, but the system they have helped co-create, cannot command respect (that is now electorally obvious).
There are however some aspects of this that bear on what keeps me awake at night: how we educate — and to a lesser extent—how we train doctors.
Well, words from the past, to remeasure the future. I am on (a sort of) sabbatical for six months. There is a project, about which I will say more later. My writings here will change, too.
My “Beginner’s Guide” to the messy world of medical education over at Wonkhe.
N.Y.U. said that it had raised more than $450 million of the $600 million that it anticipates will be necessary to finance the tuition plan. About $100 million of that has been contributed by Kenneth G. Langone, the founder of Home Depot, and his wife, Elaine, for whom the medical school is named.
To date, only a handful of institutions have tried to make medical education tuition-free, according to Julie Fresne, senior director of student financial services of the Association of American Medical Colleges, a nonprofit organization that represents medical schools.
Those who rent seek on biomedical knowledge wish to seek to define the norms of what is foundational. What is foundational for the practice of medicine should be contested more. Anatomy for surgeons is an easy case to make. But for most non-surgeons, the case for much anatomy is far from simple.
In any historical account of the ascent of modern medicine, Versalius looms large. But this Nature article (Sex, religion and a towering treatise on anatomy) intrigues me for a not so obvious reason: the counterpoint between how such knowledge was represented and understood.
Even Vesalius realized that his images could be confusing, and devised an ingenious method to explain them. A letter or number was printed onto the image of each body part, with a separate key. Unfortunately, the characters were often too small to pick out against the swirling background….
Faced by such challenges, many medics might have given up on the images. Indeed, when we reconstructed what early modern readers and scholars found fascinating about the Fabrica, it was evidently the text. The clear majority of sixteenth- and seventeenth-century readers who annotated the book focused on that and left no traces of having engaged with the illustrations. Sixteenth-century reviews of the Fabrica confirm this impression, because they tended to discuss only the text.
This is no surprise. The Fabrica’s scholarly readership was trained in the traditions of Renaissance humanism, which put a strong emphasis on textual analysis. Even if they found it difficult to interpret visual information, medical practitioners were expert at making sense of long Latin texts.
Here are the figures for skincancer909 my online textbook of skin cancer for medical students. The site was rewritten and updated in the final quarter of last year (with videos). Usage is 80% from search, with the rest from direct links. In June about 4,600 sessions. Local usage (Edinburgh) is around 5%. I am pleased, but financially poorer.
A revered teacher, Seldin was known for his pithy expressions, including: “A good medical education leaves much to be desired”; “One of the dangers of a medical education is that it leads to graduation from medical school”; and “The greatest crime is to do the right thing for the wrong reason.”
An obit of Donald W Seldin – The Lancet. I do not know the source of my favourite cognate aphorism:
most students turn into good doctors despite the earnest attempts of medical educationalists
I posted this awhile back, but it still makes me smile. I wrote:
Well my knowledge of Neanderthals is rather limited to the work showing that some of them would likely had red hair. But now a reviewer (Clive Gamble) in Nature of a book on Neanderthals states that
Wynn and Coolidge conclude that today, Neanderthals would be commercial fishermen or mechanics, based on their enormous strength and ability to learn the motor procedures needed. Their capacity for empathy might even have made them competent physicians, the authors say, although a lack of mathematical ability means that they would never have been able to graduate from medical school. Neanderthals would also make excellent army grunts, with their high levels of pain tolerance, and would be good tacticians in small combat units. They would never rewrite the tactical manual — although tearing it up, however thick, would not be a problem.
“It’s quite obvious that we should stop training radiologists,” said Geoffrey Hinton, an AI luminary, in 2016. In November Andrew Ng, another superstar researcher, when discussing AI’s ability to diagnose pneumonia from chest X-rays, wondered whether “radiologists should be worried about their jobs”. Given how widely applicable machine learning seems to be, such pronouncements are bound to alarm white-collar workers, from engineers to lawyers.
The Economist’s view is (rightly) more nuanced than Hinton’s statement on this topic might suggest, but this is real. For my own branch of clinical medicine, too. The interesting thing for those concerned with medical education is whether we will see the equivalent of the Osborne effect (and I don’t mean that Osborne effect).
Discussing the shortage of GPs, a locus GP writes:
As so often, there are several factors. Many GPs have retired early – the causes are often quoted as falling GP work earnings, disenchantment with CCGs, the CQC, and revalidation. I think more significant is the sense that we have suffered a loss of control of our work, with QOF making us ask pointless questions about emergency contraception to 45-year-olds, prescribing software pop-ups that order us around for petty savings, warnings about FGM on computers of doctors in areas with no ethnic minorities. These are very harmful to our sense of doing a worthwhile job.
And we have failed to recruit new GPs. Quite an achievement when one considers that the training is three years rather than seven, salaries are good, and there is no out-of-hours work if you don’t want it. How have the deaneries managed that? I have talked to many young doctors and most of the ones who have done F2 in general practice have felt exploited and hated it. They feel that have been chucked in at the deep end. There seems to have been lots of investment in the system for training registrars who often work at the practices of the doctors in the training hierarchy, but very little in F2 practices – who are after all the shop window that we need to perform well if we are to attract new GPs.
Maybe it is just me, but I find many of the graphics in the BMJ hard to follow. The image below is from a clinical update on “Depression and anxiety in patients with cancer” (BMJ 28 April 2018, p116-120). It occupies two whole pages. I am not certain what problem the graphic is trying to solve. For me, it just induces a sense of incomprehension. Or nausea.
In dermatology, there was a famous US academic known for producing slides with numerous arrows, many involving feedback. It was professional cargo-cult science (as the BMJ is cargo-cult education). Sam Shuster always cautioned: more than 3 or 4 arrows per slide, usually means bullshit.
That which is simple is wrong; that which is complicated is useless (Paul Valery).
My third issue is more nuanced. The biggest reason for cataract blindness is the dearth of ophthalmologists. Orthopaedic surgeons in Leicester faced with a backlog of carpal tunnel surgery decided to train intensively one theatre nurse. As a result, our carpal tunnel surgery service is probably the best and most cost effective in the country. Having a person who is not a “fully qualified doctor and surgeon” operating on cataracts could be the best solution.
Quote from, John Sandford-Smith, retired ophthalmologist, Leicester. BMJ 2018;360:k640
This sort of argument is old, and largely correct. But you can only scale such a process with the help of some form of certification, because without it, there is no durable career structure. And without this, there is no investment.
The Medical Council of India has asked state councils to investigate the problem of “ghost” teachers in medical colleges following the discovery of more than 400 fake teachers in four colleges in three states.
He said that the Maharishi Markandeshwar College in Ambala, established in 2008, has an annual intake of 150 students.“It needs a minimum of 108 faculty members as per the Medical Council of India norms. On paper it has 145 teachers listed. But that’s on paper. The college would retain a majority of the names adding new ones off and on. During inspections the doctors would appear and then disappear once the inspections were over.”
We learn that:
These “ghost teachers” are registered as faculty members drawing a hefty salary, but have never taken a single class. Most of them run private clinics, and only attend the college when there is an impending medical inspection.
Well, one wit once remarked that the collective noun for a group of professors is “An absence of professors”.
As the surgeon reached for a scrub brush, the medical student lingered back, his thumbs incessantly and rhythmically tapping on the screen of his phone. The surgeon peered at him with frustration, annoyed that again his student appeared more interested in his smartphone than the pathology. In an effort to engage him back to the case, the surgeon asked: “Can you tell me what tendons lie in each of the extensor compartments in the hand?” The student’s head snapped up, and he quickly rattled off the answer with ease. Smiling momentarily, he then asked, “Could I get your thoughts on this new video describing nerve transfers rather than tendon transfers for radial nerve injuries that was just uploaded to our educational portal? See, I have it pulled up right here, it was just presented last week at the plenary session…”