Article
Comment
Royalty
5 min read

What a monarch’s meeting teaches about politics and permanence

A monarch meeting a prime minister is a symbol of a deeper truth in a fleeting world.

Graham is the Director of the Centre for Cultural Witness and a former Bishop of Kensington.

A frail old lady, the late Queen, rises from a sofa to shake hands with an approaching woman.
The longest serving monarch meets the shortest serving prime minister.
The Royal Family.

Just about the last constitutional act of our late Queen was to give an audience to Liz Truss, the (temporary as it turned out) Prime Minister and to ask her to form a government. The pictures of a frail but smiling monarch, weakened, but still doing her job, couldn’t help but evoke a mix of admiration and affection, especially when we look back and consider that this was just two days before she died.  

But those pictures raised some questions. A Prime Minister, and a political party that forms a government, is normally chosen by the people. Queen Elizabeth was not. Neither is King Charles. She was, and he now is, our monarch by virtue of birth, something that can seem scandalous to republicans, and even to many who liked the Queen, or admire the King as decent people, but have their doubts about the monarchy. To our democratic instincts, it feels, at least to some, distinctly odd, a relic of a hierarchical past, a hangover from a less enlightened age.  

But perhaps something more significant was hidden in that act. The idea of a constitutional monarch – a figure whose position is out of our hands, as it were – formally asking a politician to form a government - acts as a reminder to us that the will of the people is not the last word, or even the first word. It tells us that, important as democracy is (‘the worst form of government except for all those other forms that have been tried from time to time’, as Winston Churchill famously put it), there is an order, an authority that stands above and beyond the will of the people. When it has worked well, the monarchy, a source of rule above that of people and parliament, has always been a symbol and pointer to a divine authority that can work through, but essentially stands above all human government. 

Because, of course, ’the will of the people’, and governments that claim to enact the will of the people, sometimes get things badly wrong. History, even that of democracies, is littered with tales of nations that have elected bad governments, or regimes that went on to enact a rule of terror in the name of ‘the people’, or where a majority has oppressed minorities. Republics of various kinds have ended up as oppressive and authoritarian. Even Hitler was elected in the first place. 

That a Prime Minister only governs at the pleasure of the Monarch is a reminder of a deeper truth - that all governments are subject to a higher accountability.

Of course, there are good monarchs and bad ones. For most of our lives, those of us who live in the UK are fortunate to have had a very good monarch in Queen Elizabeth, and we hope and pray Charles will prove to be one too. Bad monarchs, whose personal failings and moral selfishness betray the office they hold, blur the picture. They tell a different story, that authority is in itself abusive, oppressive and not to be trusted. But at its best, the continuous institution of the monarchy has served as an anchor for us, pointing away from itself to an unchanging divine presence in the course of history. The fact that a Prime Minister only governs at the pleasure of the Monarch is a reminder of a deeper truth - that all governments are subject to a higher accountability, to a moral law they did not invent, a law that tempers justice with mercy, that our lives are subject to a deeper and more lasting reality than the shifting sands of politics or times and that there is an even higher loyalty than that which we may have felt to our late Queen, or to our democratic political system. 

At the coronation, King Charles will be presented with an orb – a symbol of the world with a cross perched on top of it. It is a sign that ultimate power in this world belongs not to the King, or even the people, but to God. It is a reminder to the King, and to us, that he (and we) are accountable to an authority that stands beyond our own desires, or even the general will of the people. It is an authority represented by a cross – the symbol of love and self-sacrifice for the good of our neighbour, or even our enemy. It is one of those valuable reminders that stops any ruler from starting to think he can become a despot.  

As our constitutional system has evolved, it is the custom that Monarchs don’t get involved in the nitty-gritty of politics and it’s vital that they don’t. That is left, quite properly, to the crucial hard work of democratically elected government and politicians, who have to get on with the important but messy business of governing, working out what to do about the cost of living crisis, how to respond to conflict in Ukraine, or how to respond to those fleeing to our shores from war-torn or poverty-stricken parts of the world.  

The monarchy is a symbol of ultimate permanence, not the source of that permanence 

Over past decades, Queen Elizabeth kept to this custom. She avoided expressing opinions on particular political issues and disputes because that wasn’t her role. Her role was to be a reminder that there is an order of things beyond the temporal, a moral structure to the world that is just given, not created by us, a structure that tells us that compassion, truthfulness, integrity, justice and honesty matter in all the calculations and compromises of political decision making. 

The Queen’s death removed something steady and sure from our lives, as most of us have never known another monarch. Her death shook our sense of permanence, as the Archbishop of Canterbury put it at her funeral. Many of the vox pops we heard during the period of mourning pointed to that longing for permanence, the sense she gave of something enduring and reliable. Yet she was a symbol of ultimate permanence, not the source of that permanence.  

As King Charles is crowned, he becomes a pointer to the unshakeable and steady presence that surrounds us, upholds us and all things - the God that Christians see revealed in Jesus Christ. Queen Elizabeth understood that and showed it in her own faith – the one aspect of her personal life that she was quite open about. And there are signs that King Charles understands that too. Faith in that God is meant to be the foundation of a monarch’s rule. It can also provide a sure foundation for our individual and less public lives too, a sense of permanence in the changes and chances of this fleeting and unstable world.  

Review
Books
Care
Comment
Psychology
7 min read

We don’t have an over-diagnosis problem, we have a society problem

Suzanne O’Sullivan's question is timely
A visualised glass head shows a swirl of pink across the face.
Maxim Berg on Unsplash.

Rates of diagnoses for autism and ADHD are at an all-time high, whilst NHS funding remains in a perpetual state of squeeze. In this context, consultant neurologist Suzanne O’Sullivan, in her recent book The Age of Diagnosis, asks a timely question: can getting a diagnosis sometimes do more harm than good? Her concern is that many of these apparent “diagnoses” are not so much wrong as superfluous; in her view, they risk harming a person’s sense of wellbeing by encouraging self-imposed limitations or prompting them to pursue treatments that may not be justified. 

There are elements of O-Sullivan’s argument that I am not qualified to assess. For example, I cannot look at the research into preventative treatments for localised and non-metastatic cancers and tell you what proportion of those treatments is unnecessary. However, even from my lay-person’s perspective, it does seem that if the removal of a tumour brings peace of mind to a patient, however benign that tumour might be, then O’Sullivan may be oversimplifying the situation when she proposes that such surgery is an unnecessary medical intervention.  

But O’Sullivan devotes a large proportion of the book to the topics of autism and ADHD – and on this I am less of a lay person. She is one of many people who are proposing that these are being over diagnosed due to parental pressure and social contagion. Her particular concern is that a diagnosis might become a self-fulfilling prophecy, limiting one’s opportunities in life: “Some will take the diagnosis to mean that they can’t do certain things, so they won’t even try.” Notably, O’Sullivan persists with this argument even though the one autistic person whom she interviewed for the book actually told her the opposite: getting a diagnosis had helped her interviewee, Poppy, to re-frame a number of the difficulties that she was facing in life and realise they were not her fault.  

Poppy’s narrative is one with which we are very familiar at the Centre for Autism and Theology, where our team of neurodiverse researchers have conducted many, many interviews with people of all neurotypes across multiple research projects. Time and time again we hear the same thing: getting a diagnosis is what helps many neurodivergent people make sense of their lives and to ask for the help that they need. As theologian Grant Macaskill said in a recent podcast:  

“A label, potentially, is something that can help you to thrive rather than simply label the fact that you're not thriving in some way.” 

Perhaps it is helpful to remember how these diagnoses come about, because neurodivergence cannot be identified by any objective means such as by a blood test or CT scan. At present the only way to get a diagnosis is to have one’s lifestyle, behaviours and preferences analysed by clinicians during an intrusive and often patronising process of self-disclosure. 

Despite the invidious nature of this diagnostic process, more and more people are willing to subject themselves to it. Philosopher Robert Chapman looks to late-stage capitalism for the explanation. Having a diagnosis means that one can take on what is known as the “sick role” in our societal structures. When one is in the “sick role” in any kind of culture, society, or organisation, one is given social permission to take less personal responsibility for one’s own well-being. For example, if I have the flu at home, then caring family members might bring me hot drinks, chicken soup or whatever else I might need, so that I don’t have to get out of bed. This makes sense when I am sick, but if I expected my family to do things like that for me all the time, then I would be called lazy and demanding! When a person is in the “sick role” to whatever degree (it doesn’t always entail being consigned to one’s bed) then the expectations on that person change accordingly.  

Chapman points out that the dynamics of late-stage capitalism have pushed more and more people into the “sick role” because our lifestyles are bad for our health in ways that are mostly out of our own control. In his 2023 book, Empire of Normality, he observes,  

“In the scientific literature more generally, for instance, modern artificial lighting has been associated with depression and other health conditions; excessive exposure to screen time has been associated with chronic overstimulation, mental health conditions, and cognitive disablement; and noise annoyance has been associated with a twofold increase in depression and anxiety, especially relating to noise pollution from aircraft, traffic, and industrial work.” 

Most of this we cannot escape, and on top of it all we live life at a frenetic pace where workers are expected to function like machines, often subordinating the needs and demands of the body. Thus, more and more people begin to experience disablement, where they simply cannot keep working, and they start to reach for medical diagnoses to explain why they cannot keep pace in an environment that is constantly thwarting their efforts to stay fit and well. From this arises the phenomenon of “shadow diagnoses” – this is where “milder” versions of existing conditions, including autism and ADHD, start to be diagnosed more commonly, because more and more people are feeling that they are unsuited to the cognitive, sensory and emotional demands of daily working life.  

When I read in O’Sullivan’s book that a lot more people are asking for diagnoses, what I hear is that a lot more people are asking for help.

O’Sullivan rightly observes that some real problems arise from this phenomenon of “shadow diagnoses”. It does create a scenario, for example, where autistic people who experience significant disability (e.g., those who have no perception of danger and therefore require 24-hour supervision to keep them safe) are in the same “queue” for support as those from whom being autistic doesn’t preclude living independently. 

But this is not a diagnosis problem so much as a society problem – health and social care resources are never limitless, and a process of prioritisation must always take place. If I cut my hand on a piece of broken glass and need to go to A&E for stiches, I might find myself in the same “queue” as a 7-year-old child who has done exactly the same thing. Like anyone, I would expect the staff to treat the child first, knowing that the same injury is likely to be causing a younger person much more distress. Autistic individuals are just as capable of recognising that others within the autism community may have needs that should take priority over their own.   

What O’Sullivan overlooks is that there are some equally big positives to “shadow diagnoses” – especially as our society runs on such strongly capitalist lines. When a large proportion of the population starts to experience the same disablement, it becomes economically worthwhile for employers or other authorities to address the problem. To put it another way: If we get a rise in “shadow diagnoses” then we also get a rise in “shadow treatments” – accommodations made in the workplace/society that mean everybody can thrive. As Macaskill puts it:  

“Accommodations then are not about accommodating something intrinsically negative; they're about accommodating something intrinsically different so that it doesn't have to be negative.” 

This can be seen already in many primary schools: where once it was the exception (and highly stigmatised) for a child to wear noise cancelling headphones, they are now routinely made available to all students, regardless of neurotype. This means not only that stigma is reduced for the one or two students who may be highly dependent on headphones, but it also means that many more children can benefit from a break from the deleterious effects of constant noise. 

When I read in O’Sullivan’s book that a lot more people are asking for diagnoses, what I hear is that a lot more people are asking for help. I suspect the rise in people identifying as neurodivergent reflects a latent cry of “Stop the world, I want to get off!” This is not to say that those coming forward are not autistic or do not have ADHD (or other neurodivergence) but simply that if our societies were gentler and more cohesive, fewer people with these conditions would need to reach for the “sick role” in order to get by.  

Perhaps counter-intuitively, if we want the number of people asking for the “sick role” to decrease, we actually need to be diagnosing more people! In this way, we push our capitalist society towards adopting “shadow-treatments” – adopting certain accommodations in our schools and workplaces as part of the norm. When this happens, there are benefits not only for neurodivergent people, but for everybody.

Support Seen & Unseen

Since Spring 2023, our readers have enjoyed over 1,500 articles. All for free. 
This is made possible through the generosity of our amazing community of supporters.

If you enjoy Seen & Unseen, would you consider making a gift towards our work?
 
Do so by joining Behind The Seen. Alongside other benefits, you’ll receive an extra fortnightly email from me sharing my reading and reflections on the ideas that are shaping our times.

Graham Tomlin
Editor-in-Chief