Explainer
Creed
Seven Deadly Sins
Sin
9 min read

Reverse psychology: understand goodness then sin

Psychologist Roger Bretherton concludes our series on the seven deadly sins with a subversive proposition: we don’t understand sin because we don’t understand goodness.
An abstract shadow of a human reaching an hand skyward is overlaid by a trace of orange line that becomes a circle
Jr Korpa on Unsplash.

Over the last month or so, here at Seen & Unseen, we have been writing together about the seven deadly sins: greed, lust, gluttony, pride, sloth, anger, envy. If you have seen the David Fincher film, you probably have the grisly murders that illustrate them etched on your retina. But if there is one theme that comes up in all our articles on the subject, it is the fact that sin as a concept no longer carries any weight in our culture. A word that once bore all the heft of heaven and hell, is now the branding for a mildly indulgent discount day at the local health spa.  

One way of responding to the downgrading of sin as a meaningful and useful term, is to argue that we need a return to sin. Sin needs a come-back tour, a conceptual rehabilitation. We need to re-populate the word with meaning to make it current and plausible again. Without a consistent shared language of moral failure, of falling short, of ethical deficiency, it is difficult to imagine how responsible human community can be viable. Alasdair MacIntyre, the virtue ethicist, suggested that the problem with our culture is that multiple ethical games are being played. We are not just disagreeing about what the rules should be, but moreover what game it is we are meant to be playing. Our culture is a babel of voices, proposing conflicting versions of what a good life looks like. Consequently, in moral dialogue, we often fail to understand one another. As MacIntyre puts it: ‘my move to queen-bishop-three, is countered by your lob over the net.’

The etymology of the word sin is that it is an old English word originally derived from archery, meaning to miss the mark.

I am no etymologist. Very occasionally I dabble in a bit of New Testament Greek. But to be honest, I don’t know what I’m doing, and whenever I pronounce Greek root words they sound like items from the IKEA stocklist. And, given my tendency to talk to myself when I write, it’s almost inevitable that sooner or later the ever-attentive Alexa will accidentally order me a bedside lamp in response to what I thought was the Greek for bowels. That said, my understanding of the etymology of the word sin is that it is an old English word originally derived from archery, meaning to miss the mark. Miss by an inch, a foot, a mile – it’s all called sin (assuming archers in Old England preferred imperial units of measurement). Shoot the entire quiver in the opposite direction – that’s sin too. Linguists may tell us that this is an apocryphal origin myth, but it doesn’t matter. Whether in archery or ethics, the point remains the same. Sin is a relative term. It is relative to whatever it is we wanted to do, or aspired to become, but missed. 

Could it be then, that the root of our current cultural anomie is not so much that sin as a concept has been emptied of meaning (though it has), but that we no longer have any consensual agreement on what a good person should look like? We have no shortage of imagination when it comes to inventing new contents for the empty container of sin, but our thinking about goodness is woefully uninspired. Without a target to aim at, sin becomes vacuous. To illustrate this point, I’d like to tell you about two of the most eminent psychologists of the last hundred years. 

Hobart Mowrer and the psychology of sin 

Let’s start with a history lesson. O. Hobart Mowrer (1907-1982) is perhaps one of the most eminent, innovative and bemusing contributors to the short history of academic psychology. At the pinnacle of his career in 1953, he was elected president of the American Psychological Association (APA), the largest and most esteemed society of psychologists in the world. But as a life-long sufferer of recurrent depression, the announcement sent him into a deep psychological crisis that left him incapacitated for nearly four months. In 1959, he addressed the APA convention in Cincinnati with one of the most unusual and controversial papers of the decade, Constructive Aspects of the Concept of Sin in Psychotherapy, in which he argued that the euphemisms for sin preferred by psychologists (wrongdoing, immorality, irresponsibility etc.) were not sufficiently powerful to convey the distress of a guilty conscience. He had an ambivalent off-again/on-again relationship with institutional religion, but he was nevertheless impressed with the words of Jesus in the Sermon on the Mount that we should do our good deeds in secret. This principle – you are your secrets – became central to his development of Integrity Therapy, a group approach to psychological distress through confession. In the 1970s, the tide of cultural opinion turned against Mowrer, his Integrity Groups were accused of brainwashing their participants, and their popularity waned. Some say he subsequently retracted his views on sin. His episodes of depression continued to dog him, and in 1982, at the age of 75, he died of suicide, having long advocated this as a reasonable course of action in certain circumstances.    

This thumbnail sketch hardly does justice to the sensitive suffering genius of O. Hobart Mowrer. There is no space to recount the academic innovations that make him still one of the most cited psychologists in history. He coined the term ‘pathogenic secret’, the idea that sin – by which he meant the things that secretly bother our consciences – makes us sick. I think he was probably right about that. Take for example a freely available open-access list of what might be considered sin. It includes sexual immorality, impurity, sensuality, enmity, strife, jealousy, fits of anger, rivalries, dissensions, divisions, envy, intoxication, orgies, and things like that. It is not a particularly systematic or comprehensive list. It just happens to be the list of examples that Paul the apostle came up with in a first-century letter to residents of what is now central Turkey. They hardly make pleasant reading, but they make a great episode of Succession. 

Paul called these ‘works of the flesh’. Works, because they are things we do, situations we manufacture.  Flesh, not in opposition to physical pleasure, but because these kinds of responses (strife, envy, rage etc.) seem to be patterned bodily reactions, part of our conditioning, written into our muscle. The fact that Paul refers to them in plural (works of the flesh) is more significant than it initially appears. For him the unethical life is an incoherent ragbag of reactions, a series of plays and tactics designed to gain immediate personal gratification. As long as we come out on top, or at least think we have, they have done their job. But if we live by them, if we hand ourselves over to the universe of their self-centred cynical logic, our identity fragments, our sense of coherence shatters. We can no longer imagine who we would be if our greed, pride, lust or whatever, was taken away from us. Our appetites become our identity.  

This is why I tend to think that the instincts that led Mowrer to develop a therapy emphasising integrity was right.  Provided of course, that we embrace the full definition of integrity, rather than simply taking it as a synonym for honesty. The integrity to which Mowrer’s groups aspired was not just the truthfulness that comes from the disclosure of secrets, but the inner harmony that comes from the restoration of wholeness. What looked superficially like an unhealthy preoccupation with sin, was in fact Mowrer’s pursuit of the unified state of self that accompanies goodness. A sentiment that leads us to psychologist number two.  

Martin Seligman and the psychology of goodness 

Fast forward four and half decades from Mowrer’s election as APA president. It is 1998 and another newly-minted APA president, Martin Seligman – arguably the most famous clinical psychologist in the world – is making his inaugural address. Seligman is unique among world-famous psychologists in many ways, not least of which being his claim that he was ‘called’ to be a clinical psychologist. He later told a conference at Lambeth Palace that as a young research scientist, during his deliberations on whether to follow the path into clinical practice, he woke one night from a dream of visiting the Guggenheim Museum in New York. As he admired the architecture of the iconic building, God himself – a giant bearded old man – lifted the roof and boomed: I want you to be a clinical psychologist. Of course, as a secular Jewish academic, Seligman doesn’t believe in God, but this doesn’t deter him from openly admitting that his vocation in clinical psychology was not wholly chosen but issued from the unspoken depths of his being. 

This deep-seated sense of vocation may go some way to explaining why his 1998 inaugural address has gone down as a turning point in the history of psychology. His central assertion was that up until then psychology had been obsessed with the negative (what’s wrong with us) and that it was time to rebalance the discipline with a refocus on human flourishing, which at the time he formulated as the pleasant life, the engaged life, and the meaningful life. This is viewed by some as the birth of Positive Psychology, which according to Seligman was a corrective to the ‘rotten to the core’ view of human beings that had dominated the discipline since Freud. It is difficult to imagine a proposal more diametrically opposed to that of Mowrer. While Mowrer argued for a renewed awareness of ‘sin’, Seligman asserted to the contrary that it was the vast ignorance of goodness that bedevilled contemporary psychology. We already had a rich, ever-expanding science of what was wrong with people, what we needed was an equally detailed, every bit as expansive science of what was right with us. The explosive growth of positive psychology over the last few decades has been a response to that call, to develop a full-blown science of goodness. 

In the years that followed, Seligman revised his idea of the good life multiple times, perhaps his most ambitious proposal being the concept of Prospective Psychology; the idea that we as a species are best defined not by our past but by our future. He held this future-orientation to be so characteristic of human nature that we could name the species after it- homo prospectus.  We are defined not by what we have been, but by what we are yet to be. It is an insight he shares with many of the thoughtful people who have pondered ethics over the years. Take for example a freely available open-access list of what might be considered virtue. It includes love, joy, peace, patience, kindness, goodness, faithfulness, gentleness, self-control. It is not a particularly systematic or comprehensive list. It just happens to be the list of examples that Paul the apostle came up with in a first-century letter to residents of what is now central Turkey. I often feel better about life, just reading it. 

Paul called these beautiful qualities, ‘fruit of the Spirit’. It is significant that in the original New Testament Greek, the word fruit is singular: καρπὸς (No! Alexa! I don’t need garden furniture!) He doesn’t write fruits, but fruit of the Spirit. He points toward some kind of unity, harmony, consistency in these qualities. Mowrer no doubt would have called it integrity. Goodness is much more than the avoidance of naughtiness; it is the restoration of wholeness to our shattered and divided selves. Sin is much more than the cheeky indulgences we succumb to at the end of a bruising day at work; it is the misdirection and derailment of all we could become. And this is where the tale of two psychologists terminates. Contrary to what the marketing executives may have conspired to tell you, sin is not your friend; it is the enemy of your genuine divine magnificence. That’s what makes deadly sin so deadly. 

Article
Belief
Creed
Identity
Truth and Trust
5 min read

Calls to revive the Enlightenment ignore its own illusions

Returning to the Age of Reason won’t save us from post-Truth

Alister McGrath retired as Andreas Idreos Professor of Science and Religion at Oxford University in 2022.

In the style of a Raeburn portrait, a set of young people lounge around on their phones looking diffident
Enlightened disagreement (with apologies to Henry Raeburn).
Nick Jones/Midjourney.ai.

Is truth dead? Are we living in a post-truth era where forcefully asserted opinions overshadow evidence-based public truths that once commanded widespread respect and agreement? Many people are deeply concerned about the rise of irrational beliefs, particularly those connected to identity politics, which have gained considerable influence in recent years. It seems we now inhabit a culture where emotional truths take precedence, while factual truths are relegated to a secondary status. Challenging someone’s beliefs is often portrayed as abusive, or even as a hate crime. Is it any surprise that irrationality and fantasy thrive when open debate and discussion are so easily shut down? So, what has gone wrong—and what can we do to address it? 

We live in an era marked by cultural confusion and uncertainty, where a multitude of worldviews, opinions, and prejudices vie for our attention and loyalty. Many people feel overwhelmed and unsettled by this turmoil, often seeking comfort in earlier modes of thinking—such as the clear-cut universal certainties of the eighteenth-century “Age of Reason.” In a recent op-ed in The Times, James Marriott advocates for a return to this kind of rational thought. I share his frustration with the chaos in our culture and the widespread hesitation to challenge powerful irrationalities and absurdities out of fear of being canceled or marginalized. However, I am not convinced that his proposed solution is the right one. We cannot simply revert to the eighteenth century. Allow me to explain my concerns. 

What were once considered simple, universal certainties are now viewed by scholars as contested, ethnocentric opinions. These ideas gained prominence not because of their intellectual merit, but due to the economic, political, and cultural power of dominant cultures. “Rationality” does not refer to a single, universal, and correct way of thinking that exists independently of our cultural and historical context. Instead, global culture has always been a bricolage of multiple rationalities. 

The great voyages of navigation of the early seventeenth century made it clear that African and Asian understandings of morality and rationality differed greatly from those in England. These accounts should have challenged the emerging English philosophical belief in a universal human rationality. However, rather than recognizing a diverse spectrum of human rationalities—each shaped by its own unique cultural evolution—Western observers dismissed these perspectives as “primitive” or “savage” modes of reasoning that needed to be replaced by modern Western thought. This led to forms of intellectual colonialism, founded on the questionable assumption that imposing English rational philosophies was a civilizing mission intended to improve the world. 

Although Western intellectual colonialism was often driven by benign intentions, its consequences were destructive. The increasing influence of Charles Darwin’s theory of biological and cultural evolution in the late nineteenth century led Darwin’s colleague, Alfred Russel Wallace, to conclude that intellectually and morally superior Westerners would “displace the lower and more degraded races,” such as “the Tasmanian, Australian and New Zealander”—a process he believed would ultimately benefit humanity as a whole. 

We can now acknowledge the darker aspects of the British “Age of Reason”: it presumed to possess a definitive set of universal rational principles, which it then imposed on so-called “primitive” societies, such as its colonies in the south Pacific. This reflected an ethnocentric illusion that treated distinctly Western beliefs as if they were universal truths. 

A second challenge to the idea of returning to the rational simplicities of the “Age of Reason” is that its thinkers struggled to agree on what it meant to be “rational.” This insight is often attributed to the philosopher Alasdair MacIntyre, who argued that the Enlightenment’s legacy was the establishment of an ideal of rational justification that ultimately proved unattainable. As a result, philosophy relies on commitments whose truth cannot be definitively proven and must instead be defended on the basis of assumptions that carry weight for some, but not for all. 

We have clearly moved beyond the so-called rational certainties of the “Age of Reason,” entering a landscape characterized by multiple rationalities, each reasonable in its own unique way. This shift has led to a significant reevaluation of the rationality of belief in God. Recently, Australian atheist philosopher Graham Oppy has argued that atheism, agnosticism, and theism should all be regarded as “rationally permissible” based on the evidence and the rational arguments supporting each position. Although Oppy personally favours atheism, he does not expect all “sufficiently thoughtful, intelligent, and well-informed people” to share his view. He acknowledges that the evidence available is insufficient to compel a definitive conclusion on these issues. All three can claim to be reasonable beliefs. 

The British philosopher Bertrand Russell contended that we must learn to accept a certain level of uncertainty regarding the beliefs that really matter to us, such as the meaning of life. Russell’s perspective on philosophy provides a valuable counterbalance to the excesses of Enlightenment rationalism: “To teach how to live without certainty, and yet without being paralyzed by hesitation, is perhaps the chief thing that philosophy, in our age, can still do for those who study it.” 

Certainly, we must test everything and hold fast to what is good, as St Paul advised. It seems to me that it is essential to restore the role of evidence-based critical reasoning in Western culture. However, simply returning to the Enlightenment is not a practical solution. A more effective approach might be to gently challenge the notion, widespread in some parts of our society, that disagreement equates to hatred. We clearly need to develop ways of modelling a respectful and constructive disagreement, in which ideas can be debated and examined without diminishing the value and integrity of those who hold them. This is no easy task—yet we need to find a way of doing this if we are to avoid fragmentation into cultural tribes, and lose any sense of a “public good.” 

Support Seen & Unseen

Since Spring 2023, our readers have enjoyed over 1,500 articles. All for free. 
This is made possible through the generosity of our amazing community of supporters.

If you enjoy Seen & Unseen, would you consider making a gift towards our work?

Do so by joining Behind The Seen. Alongside other benefits, you’ll receive an extra fortnightly email from me sharing my reading and reflections on the ideas that are shaping our times.

Graham Tomlin
Editor-in-Chief