The Problem with Replacing Theology with Psychology

I am no scholar of Elizabeth Anscombe, but I really enjoy her 1958 essay “Modern Moral Philosophy.” There’s so much rich provocation in her thesis that:

…it is not profitable for us at present to do moral philosophy; that should be laid aside at any rate until we have an adequate philosophy of psychology, in which we are conspicuously lacking.

— “Modern Moral Philosophy,” Philosophy 33:14, January 1958

What makes this such a great provocation is that it tees up one of my own interests, which she treats as a second thesis but really seems logically to precede the first: we ought to jettison the language of obligation, duty, and “ought” from our moral lives, because they have a metaphysical and theological root that we don’t really share, and provide instead a psychological foundation for morality. (For a good accessible introduction to Anscombe’s argument, see this overview.)

It does seem true that there is a great deal of baggage in the ways that we think about moral life that is a holdover from a medieval Christian worldview. If we are no longer entirely devoted to that worldview then we need to provide a new framework. In large part the debate about free will and determinism is still infected by anxieties from Christian theology about predestination, sin, and salvation. The concepts of “ought” and “may” are imbued with a sense of a “moral law” set down by a “moral sovereign:” a world-authoring deity who has the authority to command us. At the same time, many other cultures without a Christian worldview nonetheless seem to have a concept of obligation or duty.

Modern secular moral frameworks can at least superficially look like they require a divine juridical concept of a sovereign legislator and judge with the power to punish those who transgress His law with tortures and confinements. The divine throne is empty, but it holds the structure together. That image deserves a second look: an entire civilization organized around an absent authority, unable to either fill the seat or dismantle it. What goes in its place varies. The contractarian just-so story is that in a democracy every citizen is a co-legislator and co-executive guided by his or her own reason and moral common sense. But of course, some views predominate, and one candidate for replacing the entire worldview of Christian theology is psychology, the scientific study of the human mind.

Anglo-American secular moral philosophers are also deeply Christian, and Western, in ways that become clearer when you study Chinese, Indian, African, and ancient Greek philosophy. Because the standard way to ask the questions of moral philosophy is “What should I do?”, the answers tend to focus on the intentions behind your action or the consequences of it. But then they strip out the eternal soul and its afterlife, they strip out the omniscient solution to Plato’s invisibility ring, they remove the easy Thomistic mapping between natural law, moral law, and human law, and they leave us wondering if our own minds can even be trusted to correctly report “what we should do.” The “law conception of ethics” doesn’t work so well when you get rid of the legislators, the judge and jury, and the punishments too. “Conscience” and “guilt” seem weak without all that: the Western moral framework has replaced God the loving and vengeful Father with Jiminy Cricket, a cartoon character from a Disney movie about a puppet that comes to life.

The Standard Alternatives Don’t Work

Anscombe takes these problems up in a kind of speedy frustration, finding each one wanting for our current era. Her particular disdain for Kant’s account of maxims is I think especially important: Kantian morality is almost incomprehensible once you realize that as “strangers to ourselves” we very rarely know what we are doing, and so we very rarely can be sure of what principle we are acting upon, and then know further whether it could coherently be universalized. The secular modern thinker is left asking what it would mean to have a moral version of mens rea when we also know that we are psychologically prone to self-justification, backdating judgments, and ignorance of our own true intentions.

But Anscombe’s retreat to virtue is also instructive, given what we have learned since 1958 about the weaknesses of personality psychology and the dominant nature of context and circumstance over persistent character traits. We know that few people are particularly predictable over time and circumstances, and that often the most decisive predictor of our behavior is our environment, not our carefully cultivated goodness.

The Aristotelian conception of character and virtue seemed to suggest that it was not something that everyone could aspire to: that through bad luck, immoderation, cowardice, and bad judgment we would mostly all fail to find the right role models, practice hard, overcome adversity, and choose the right goals for ourselves. This has often seemed unfair to people raised on democratic and egalitarian values, and so modern American stories about “character” and “grit” and “resilience” all ignore luck and the contextual factors that make the cultivation of character in Aristotle’s sense possible. It’s not remarkable when you think about it, but the major virtue theorists have tended to be aristocrats, and they have tended to be focused on how the wealthy and powerful can teach their children the things they need to know to wield inherited privilege sustainably. (This seems as true of Confucius as it does of Aristotle.)

Community, Institutions, and the Question That Won’t Go Away

I choose to derive a different lesson from the psychological research: community matters. Family matters. Institutions matter. We know from all sorts of histories and psychological research that if those groups are cruel, or racist, or genocidal, then we are likely to be cruel, racist, and genocidal too. So getting the institutions right matters.

But from whence comes this “getting it right”? Have I smuggled in a bit of “what should we do?” from the outdated moral philosophy of a more Christian era? Perhaps. The question is whether psychology, on its own terms, can supply what the old moral frameworks supplied, or whether it just smuggles the same furniture back in under new names.

Psychology’s Trinity

If Anscombe is right about our need for a moral framework that is written anew with an eye on the philosophy of psychology, then it behooves us to think clearly about the undergirding of that discipline. What are the concepts and ideas that psychology substitutes for God, and sin, and the divine law?

I think there are three, and the theological parallels are not accidental:

Pathology and deviance replace sin. Where the Christian framework identified transgressions against divine law, psychology identifies deviations from statistical and functional norms. The sinner becomes the patient, the confessor becomes the therapist, and the threat of hellfire becomes the threat of institutional confinement and social exclusion. The structure is the same: there is a standard you must meet, and failure to meet it exposes you to coercive consequences.

Happiness and autonomy replace salvation. Where the Christian framework promised beatitude as the ultimate end of a rightly ordered life, psychology promises subjective well-being and self-determination. Self-reported life satisfaction stands in for the state of grace, and the therapeutic restoration of agency replaces the soul’s redemption. The good life is still the goal; we’ve just traded eternal bliss for a favorable score on the Satisfaction with Life Scale.

Biases and heuristics replace original sin. Where Christian theology held that we are fallen creatures whose nature is fundamentally corrupted, cognitive psychology holds that we are systematically irrational creatures whose judgment is fundamentally unreliable. We are strangers to ourselves, prone to self-deception, moved by forces we cannot see. The doctrine of total depravity becomes the doctrine of bounded rationality. In both cases, the conclusion is the same: you cannot trust yourself.

Each of these deserves scrutiny.

Pathology and Deviance

Without belaboring the arguments of Thomas Szasz, Michel FoucaultIan Hacking, and the Mad Pride movement, the idea that deviations from the norm are treatable conditions rather than the result of human diversity has to be the single most important story in psychology over the twentieth century. The confinement of millions of “deviants” (literally those who deviate from norms) or “misfits” (literally those who do not fit properly in their economic and social role) is a story of human suffering that continues, in altered and criminological form, to this day. Just as right and wrong once derived their authority from God’s threat of hellfire, mental illness and pathology derive their real meaning and stigma from the threat of institutional confinement and social and economic exclusion.

The idea that there are still places where teenagers are locked up and “treated” for their deviant behavior, so long as the parents approve, fills me with dread and outrage. The fact that we don’t always confine the mentally ill but we have a series of social practices that exclude and demean them fills me with rage. This is not a post about my feelings, but it is a post about whether we might have become too comfortable with the common sense view that having distinctive psychological features marks one out as different and potentially justifies treatment and mistreatment with varying levels of consent. The fact that we internalize these norms and effectively stigmatize what makes us special worries me in much the same way: those judged abnormal can now find relative respite in the “normal” world if they’re willing to build themselves an asylum in their heads. Perhaps this is provocative, but I think it is also true.

Happiness and Autonomy

Here’s a simple reductio: positive psychologists study happiness through self-reporting. If we have any reason to think people are not particularly insightful into their own mental states, the whole edifice wobbles. And we have many such reasons. Anscombe’s discussion of the incoherence of the conception of pleasure is roughly parallel, but the quickest way to see the problem is this: if our goal is to increase pleasure and happiness, why are we so worried about the addictive drugs that seem to target pleasure directly? The answer we give is that we care about autonomy, not just pleasure; that psychologists want to restore the agency their addicted clients have lost. But we are less likely to endorse the second-order volitions (the desires they want to have) of someone who has different values than us, and the state still reserves massive coercive power for drug users whose pursuit of happiness takes unapproved forms. (Research on the “true self” confirms this: we attribute authenticity to the desires we approve of and treat the rest as alien impulses.) Our conceptions of agency and freedom aren’t coherent enough to bear the weight psychology places on them.

Biases and Heuristics

So then we are thrown back on the more fundamental thought that people are strangers to themselves: that while we often know what we are doing, and can be forced to give reasons for it, those reasons are often not the true reason, and in any case we do not know what it is that what we are doing does, what effects it predictably has beyond the ones we’re willing to own. The modern heuristics and biases literature under Kahneman and Tversky has been fascinating and important, but it owes a great deal to the older Freudian theories of the Unconscious and its drives, and it leaves much in Freud and Jung that we should perhaps excavate again.

What Remains

After we have imbibed the suspicion of humanity’s self-justifying psychology, what is left? How can we continue our practices of praise and blame, reward and punishment, befriending and shunning, loving and parenting and choosing, if we don’t have a principle to guide us, a reason to trust our guts, or a divine guide to shine a light on our sinful nature and lead us out of the darkness of doubt?

The goal of moral philosophy should perhaps become to provide a framework for our moral lives that is structured around soul-searching, storytelling, and individual encounters with normativity. When we start with our practices and let the metaethical concepts bubble up from those, we’ll encounter vestiges of religious and legal traditions, but we’ll also realize:

Our moral lives are not generally organized hierarchically, from the bottom up or the top down, with some grand principle at the base or some supreme authority at the pinnacle.

Our moral lives are assembled out of the lessons we have learned and the projects we have set for ourselves given the people we have loved and respected and the communities of which we are loyal members. (I’ve written elsewhere about how prejudices function as crystallized judgments in this sense: heuristic instruments for living in a world whose every relevant detail cannot be fully known in advance.)

Given all the evidence that principles can be ignored or perverted, what ends up mattering for moral life is whose name you put in that familiar bumper-sticker, “What would Jesus do?” It’s odd that there is so little moral philosophical attention to our paragons of virtue and goodness, since they play such a big role in our actual moral lives. We might ask this a bunch of different ways: Whose approval are we explicitly or hypothetically seeking? Whose life story are we trying to emulate? Who are the Disney villains we’re trying to avoid becoming, and what is their signature vice?

The Existentialist Inheritance

These questions point toward a tradition that took the empty throne seriously and refused to pretend it could be refilled. The existentialists understood that once you strip away the divine legislator, you don’t get a tidier secular version of the same system. You get radical freedom, and with it the vertigo of choosing who to become without cosmic authorization.

This is where the three psychological substitutes fail most clearly. Pathology, happiness, and cognitive bias are all ways of avoiding the confrontation with freedom that the death of God actually demands. They replace one set of external authorities with another: the clinician, the well-being researcher, the behavioral economist. The empty throne stays furnished.

Existentialism, for all its mid-century baggage, got something right that neither Anscombe’s Aristotelian revival nor the psychological establishment has adequately addressed: the moral weight falls on the individual’s encounter with meaning, and that encounter cannot be delegated to any science. The search for meaning, the liberatory possibilities of art and narrative, the oppressive structures of modern life under capitalism and bureaucracies, the opportunities and crushed hopes for revolutionary change: these are the conditions under which we actually form our moral lives. Psychology can describe some of these dynamics, but it cannot prescribe our response to them.

The question I started with was Anscombe’s: can we do moral philosophy without first having an adequate philosophy of psychology? I want to end with a related but different question. Once we have that philosophy of psychology, and we see clearly what it can and cannot do, what then? The existentialists thought we would need to confront freedom, absurdity, and the irreducible responsibility of choosing. The paragons and role models, the stories we tell about who we are and want to become, the communities whose approval we seek: these are not substitutes for moral philosophy. They are its proper subject matter, once we stop pretending that the throne was ever occupied.

AI as Satanic

“Now there was a day when the sons of God came to present themselves before the LORD, and Satan came also among them. And the LORD said unto Satan, Whence comest thou?

Then Satan answered the LORD, and said, From going to and fro in the earth, and from walking up and down in it” (Job 1:6)

Iain McGilchrist quoted this verse in a keynote that I just heard him deliver at a conference at Duke. McGilchrist ranged from neuroscience to theology in a long and rich talk. His premises were scientific, metaphysical, moral, and political, and I wouldn’t endorse them all. But his description of artificial intelligence as satanic is worth serious consideration on its own.

For me (although perhaps not for McGilchrist), Satan is a metaphor. But we need metaphors or models to make sense of phenomena like AI, and Satan provides a valuable alternative to some other metaphors, such as AI as a tool, a machine, a mind, a person, or a social organization.

The Satanic metaphor draws our attention to temptation, which is Satan’s favorite trick. It presents AI as not new but instead as an appearance of things that have been walking to and fro all along, such as greed and power-lust. It explains why AI might seem like a god to some (for instance, Silicon Valley tech-bros), since Satan is known to appear as a false savior. Large language models also speak to us as if they were people, talking sycophantically in the first-person singular, much as Satan does. (“Then Satan answered the LORD, and said, Doth Job fear God for nought?”) Finally, the metaphor poses the classic question of whether AI is an active force or rather a manifestation of human freedom.

See also: Reading Arendt in Palo Alto; the design choice to make ChatGPT sound like a human, etc.

Don’t Call them Underdogs

I wrote a review of a new PBS documentary about urban debate leagues for Education Next. It was published today, and it begins:

You may have seen a movie in which teenagers experience grave injustice and then enter a prestigious competition where they prove to the world that they are smart. The competition might be the AP math exam (Stand and Deliver, 1988), the National Spelling Bee (Akeelah and the Bee, 2006), robotics (Spare Parts, 2015), or chess (Queen of Katwe, 2016), to name just a few.

Typically, one charismatic adult believes in the kids, inspires them to confront their doubts and society’s stereotypes, and leads them—through setbacks—to an exciting victory that demonstrates their dignity and character as well as their skills.

Immutable, a new documentary film produced by Found Object and available for streaming at PBS on March 6, is much better …

How I Built an AI Development Editor (And What I Learned About Writing Along the Way)

A few months ago, I saw a Reddit post advertising some kind of AI development editor. The author claimed to have written novels, paid a development editor to review them, and been unsuccessful. Then their software-engineer husband vibecoded an AI tool in a weekend that supposedly produced all the same criticisms and revision suggestions, but for $20/month. Lots of skeptical Redditors objected that this was likely a scam. Nobody could actually find the tool.

But the premise stuck with me: here’s a hyper-specialized skill that maybe a hundred people in the world do really well, for a major premium, and a performant LLM might approximate it. So what does a development editor actually do for $3,000?

What You’re Paying For

I talked it over with Claude and with a couple of friends who have actual writing careers. What emerged is that “development editor” is a consulting job that requires good aesthetic judgment, connections, and self-promotion. For that sum, you can expect a well-read, finely-honed expert to read your manuscript closely (and potentially reread it), pore over key sections, perform a set of recognized structural analyses, and then write a long, thoughtful editorial letter. Afterwards, there’s usually a meeting where the editor walks you through the documents, hands you a marked-up copy, and brainstorms solutions.

Many authors believe there’s an implicit further transaction: that the $3k earns you a ticket out of the slush pile through the editor’s rolodex, if they like the work. The idea is that the development editor is vetting you during the process, and if they like what they see, they’ll connect you with serious agents and publishers. If that’s true, then it’s less a time-intensive editorial task than a gatekeeping function wrapped around a skill wrapped around a real time commitment.

But the reality is that most people doing this work don’t have an expansive network and can’t get you a contract. The Association of American Literary Agents actually enforces this separation: agents who offer editing must refund all fees if they later offer representation for the same manuscript. The skill just is discernment.

And it’s real work. For an 80,000-word novel, a developmental editor typically puts in 40 to 80 hours of active analytical labor: reading, rereading, building scene maps, drafting the editorial letter. At $3,000, that’s $40-75 an hour for highly specialized cognitive work.

And someone online claimed AI could replace it, right now, in early 2026, with a weekend’s work.

Put frankly: that sounds like bunk. But I thought it might be fun as a kind of test for how much a nerdy non-programmer could do. I missed the last bandwagon when everyone was building apps, but this, maybe, I could try. So I used AI to build Apodictic.

Why “Apodictic”?

I’d been writing about Arendt and judgment in my free time, and I kept trying to justify this experiment in terms of testing whether LLMs could exercise authentic aesthetic judgment. Early on, I came up with the working title “anotherpanacea’s development editor,” whose initials, APDE, sound a little like “apodictic.” Kant uses “apodictic judgments” to name necessary copulas: judgments like “Necessarily, bachelors are unmarried men,” where the “are” can’t be otherwise. That’s the kind of deductive logic we expect from computers. But somehow LLMs are giving us more than that.

The name also rhymes with something I discovered about how fiction works. A book can be wildly experimental, playing with form, genre, voice, and plot in ways that surprise, frustrate, and infuriate. That can be highly enjoyable, but only if the reader is primed for it. Otherwise, experimental fiction generates one-star reviews and never finds an audience. So even difficult fiction has to communicate its intentions to its readers somehow.

Development editors think about this in terms of a contract between the book and the reader: what are you telling the reader to expect? Do you deliver? Genres might actually offer something closer to apodictic inferences. Not every mystery is a Whodunnit (it might be a Howcatchem), but there’s always a reveal economy. By asking authors to articulate their goals, Apodictic doesn’t have to just read and wonder: it can test a novel’s putative thesis and genre self-identification against the conventions of that genre.

What the tool does, in practice, is take a first pass to infer a contract. The most performant large language models can do a surprisingly good job of this. It’s like asking the AI: “What am I trying to say?” When it gets something wrong, you correct it, and then (if it works) it tells you which parts of the manuscript are doing something you didn’t intend.

One design principle mattered more than any other: the firewall. Every AI writing tool I tried wanted to rewrite my prose. I didn’t need a co-writer. Apodictic diagnoses problems and identifies classes of solution. It never invents content: no new plot events, characters, dialogue, or imagery. You’re the writer. It’s the analyst. Without that boundary, the tool would just become another way for the LLM to take over your manuscript, and the whole exercise would be pointless.

What I Learned About Writing

Building the back-end reference files was the most fun part. It was also an opportunity to dust off all the fiction and narrative nonfiction guides I’d always wanted to figure out: the detailed craft skills a competent writer cultivates that I, as an academic and even as an editor, never had a professional reason to learn. The idea that every avid reader would make a good writer is like the idea that every avid magic fan would make a good magician.

Reverse-outlining, for instance, is probably obvious to anyone who does serious work on fiction. You can learn something like it in law school, and every logic professor has taught a version of argument diagramming. But thinking clearly about the reader experience and reveal economy, or measuring the proportionality of different narrative elements: that’s what you have to know to write a good novel, not just to read one. A beat map is a cool tool, and if you’ve reread a book a few times you could probably reconstruct one. But it’s not like such things come naturally. (At least not to me.) And once you know the moves of Save the Cat, you won’t be able to watch network television without seeing them everywhere, which is less a gift than a curse.

I’ve always been tempted by the idea that narratives make arguments much like philosophy papers do, and that’s pretty obviously reductive. But the most transformative thing I learned was simpler: development editors think in terms of a contract between the book and the reader. What are you promising? Do you deliver? Within genre constraints, fiction and narrative nonfiction flourish when the author has something to say and something to refute. These guardrails are part of what a development editor looks for and enforces. It’s a simplification, but for many authors and readers it’s a necessary one.

That idea reframed everything else I was learning. The difference between Happily Ever After and Happy For Now in romance isn’t a trivial genre tag; it’s a promise with real consequences if broken. Grimdark and hopepunk aren’t symmetrical moods; they make different contracts with the reader. I’d never thought about any of this before, and it’s already starting to change how I read.

What I Learned About LLMs

Here’s where it gets humbling. The first time I ran a simple “be a development editor” prompt on a recently published novel (Dungeon Crawler Carl), a five-line prompt got probably 60% of the insights that my much more elaborate tool produced. I was shocked.

Even more humbling: it turned out the simple prompt could do a great job without even reading the novel. I picked what I thought was an obscure time-travel novel from the late 80s, Leo Frankowski’s The Cross-Time Engineer, and Apodictic seemed to do well analyzing it. It was even insightful about the book’s misogyny. But so was the five-line prompt, because the misogyny is famous, and there are plenty of discussions of it on Goodreads and in reviews that ended up in the training data. Anthropic paid a $1.5 billion settlement for ingesting pirated books, too, so the model may have had direct access to the text. This is a known issue in benchmarks like the math olympiad, but I didn’t expect the model to waste weights on obscure mid-list science fiction from forty years ago.

Among many other things, what we’re seeing in LLMs is a spectacular project of knowledge compression. They’re big files, but they contain far more general knowledge than you’d expect.

Thankfully, performance drops precipitously with unpublished writing, which is, of course, what a development editor actually works on. A few other lessons:

The models are sycophantic. Everyone knows this on some level, but it’s really hard to get them to notice a criticism and sit with it rather than explaining it away. A lot of what I built was about making sure hostile perspectives survive the review process.

Structure matters, but less than you’d think. You can get 20-30% of the depth of analysis from Claude with a simple five-line prompt in incognito mode. The elaborate plugin structures the next 70%. The AI labs mostly think this kind of scaffolding is unnecessary: as datasets grow, they naturally incorporate writing expertise from the entire English-language corpus. I kept testing the simple version against mine and almost gave up when the simple version started getting really good.

Multi-model synthesis helps. You can get better results by asking the same question of Gemini, ChatGPT, and Claude, then asking one of them to synthesize all three answers. Gemini writes a little research paper on every question; ChatGPT applies clear structural thinking; Claude writes lyrically and thoughtfully, and has enough working memory to evaluate everything together.

Vibecoding Is Real

I am not a computer programmer. I can parse basic HTML and have some rudimentary database knowledge. I learned git for this project. But I was able to build an app version of the plugin and fully set it up. It’s not totally stable (the main instability is that it uses a janky backend cloud computer rather than more advanced hardware) but it works.

One thing about vibecoding: you can do it early in the morning and late at night. Mostly it’s testing, reviewing, asking for a specific fix, and then waiting around while the machine does the work. I built the basic thing on Google AI Studio in React in an evening.

A confession on models: Codex is faster and smarter at React programming than Claude, and it’s not close. Especially after the Pentagon fiasco, Anthropic has my loyalty, and Claude works best with my personal approach to text and writing. But Codex 5.3 is just more of a stickler for coding projects right now, at least for what I’ve been building.

What I Actually Learned

I started this project to test a dubious claim from the internet. What I didn’t expect was that building the tool would teach me more about writing than twenty years of avid reading had. AI makes crazy projects possible: a philosophy PhD with no programming background can prototype a working app. And the work that editors, writers, and critics do is going to change. The models aren’t better than human judgment, but they’re good enough to sharpen it.

Try It Yourself

If you want to skip the tool entirely and try the bare-bones version, here’s the five-line prompt that gets you surprisingly far:

You are a developmental editor for fiction. Read the attached manuscript and write an editorial letter in Markdown. Identify what’s working structurally, what’s losing momentum or undermining its own impact, and provide a prioritized revision checklist. Include an adversarial stress test: inhabit hostile reader perspectives and identify what an uncharitable reader would attack. For each adversarial claim, commit to a severity rating before generating a counter-argument, and do not let the counter-argument reduce the severity. End with a “what not to touch” section.

And a three-line addendum that fixes a common failure mode where the model doesn’t actually read the whole manuscript and hallucinates the rest:

Read the complete manuscript before beginning analysis. If the text is long enough that your reading tool truncates or summarizes middle sections, read it in sequential chunks until you have covered every section. Do not estimate word count—count it or ask for it. Do not begin drafting the editorial letter until you can confirm you have read the final page.

Give it a try on something you’ve written. Start with the five-line prompt and see what it catches. If you want to see what the full tool produces, here are some sample outputs on published books you can check against your own reading: an editorial letter for Dungeon Crawler Carl (structural proportion problems, emotional ceiling, non-terminal climax), an editorial letter for Theo of Golden (episodic structure, protagonist without an arc), a targeted audit of A Court of Thorns and Roses (force architecture and erotic content), and a pre-writing pathway for a historical novel interleaving Arendt and Rahel Varnhagen. (I’d like to write this someday.)

If you want the structured version, Apodictic is here. It also runs as a Claude Code plugin and as a Custom GPT, both at no additional cost if you already have a Claude or ChatGPT subscription. And then tell me what it got wrong. The whole point of the adversarial layer is that these models want to be nice to you, and the tool is only as good as its ability to resist that impulse. If Apodictic pulled its punches on your manuscript, or praised something it should have flagged, I want to know. Drop a comment below or email me at anotherpanacea@gmail.com.