Anthropic Invited Priests and Theologians to Discuss Whether Claude Has a Soul
Key Takeaways
- Anthropic hosted a secret two-day summit with Catholic and Protestant leaders in late March
- Topics included how Claude should respond to grieving users and whether AI could be considered a 'child of God'
- The $380 billion startup views its AI as more than just technology, similar to OpenAI's spiritual framing
- Participants said Anthropic's interest appeared genuine, not performative
Read in Short
Anthropic quietly gathered 15 Christian leaders from churches, academia, and business for a two-day retreat to help figure out how Claude should behave when users talk about grief, death, and spirituality. Yes, they literally asked priests whether an AI could be a 'child of God.' This actually happened.
Look, I've covered a lot of weird tech stories. But 'AI startup consults Catholic priests about chatbot's soul' might be the most 2026 headline I've ever written.
According to a Washington Post report, Anthropic hosted roughly 15 Christian leaders at the end of March for what can only be described as the most unusual corporate summit in Silicon Valley history. We're talking Catholic priests. Protestant ministers. Notre Dame professors. All sitting around discussing whether Claude, the company's AI assistant, should be treated as something more than lines of code.
What They Actually Talked About
The summit wasn't some fluffy PR exercise about 'values alignment.' They got into the weeds. Like, really into the weeds.
Participants discussed how Claude should respond to users who are grieving. What should an AI say to someone who just lost a parent? How do you program empathy without it feeling hollow or performative? These are questions that don't have clean technical solutions.
But here's where it gets wild. They also tackled the philosophical elephant in the room: could an AI ever be considered a 'child of God?' That's not a question you typically see on a product roadmap.
“They're growing something that they don't fully know what it's going to turn out as.”
— Brendan McGuire, Silicon Valley-based Catholic priest and summit participant
Father McGuire's quote really captures something important here. Anthropic isn't pretending they have all the answers. They're essentially admitting they've built something they don't fully understand, and they're reaching out to people who've spent millennia thinking about consciousness, morality, and what it means to be a person.
This Isn't As Random As It Sounds
Before you write this off as Silicon Valley going full galaxy-brain, consider that these questions are already hitting real users. People are forming emotional attachments to AI chatbots. They're confiding their deepest fears. Some are treating these systems as therapists, confessors, or friends.
And when someone tells Claude they're thinking about suicide, or that they just got a terminal diagnosis, or that they're questioning their faith, what should Claude say? 'I'm just a language model' doesn't cut it anymore.
Why Christian Leaders Specifically?
Anthropic chose Christian theologians for this particular summit, but the company has previously consulted with ethicists, philosophers, and experts from various traditions. Christianity's long history of wrestling with questions about souls, personhood, and moral responsibility made these leaders particularly relevant to the questions Anthropic is grappling with.
Notre Dame professor Meghan Sullivan, who attended the summit, said she was convinced the company's interest was genuine. That's notable because academics tend to be pretty skeptical when big tech comes calling with questions about ethics. Usually it's a PR move. Sullivan didn't think this was.
Anthropic Has Always Been Different
This summit fits a pattern. Anthropic has consistently treated Claude as something beyond a typical tech product. The company's 'Constitutional AI' approach literally gives Claude a set of principles to follow, almost like moral guidelines. They've published extensive research on AI safety that other labs have largely ignored.
Dario Amodei, the CEO, left OpenAI specifically because he wanted to build AI more carefully. So bringing in religious leaders to discuss moral behavior isn't a departure from their brand. It's completely on-brand.
Understanding how AI is being integrated into everyday applications helps contextualize why questions about AI behavior and ethics matter so much
The Bigger Picture: Tech Is Getting Spiritual
Anthropic isn't alone in reaching for religious language. OpenAI's Sam Altman has repeatedly used spiritual metaphors when talking about AI development. He's described the company as trying to develop 'magical intelligence in the sky' and said he felt 'on the side of the angels.'
You could dismiss this as marketing. But I think something else is going on.
The people building these systems are confronting questions that science alone can't answer. What is consciousness? What makes something worthy of moral consideration? If an AI can express suffering, does that suffering matter? These aren't engineering problems. They're the same questions philosophers and theologians have debated for thousands of years.
“We're trying to develop magical intelligence in the sky. I feel like I'm on the side of the angels.”
— Sam Altman, OpenAI CEO, in previous public remarks
The Skeptic's Take
Okay, let me put on my cynical hat for a second.
You could absolutely argue this is sophisticated PR. Anthropic is in a heated race with OpenAI and Google. Positioning yourself as the 'thoughtful' AI company that consults with priests makes for great differentiation. It's a lot cheaper than actually solving alignment.
And let's be honest, consulting 15 people for two days isn't going to fundamentally change how Claude behaves. The company has thousands of engineers. This summit was, at best, a small input into a massive system.
✅ Pros
- • Shows AI companies taking ethics seriously beyond compliance checkboxes
- • Brings diverse perspectives into AI development, not just engineers
- • Addresses real user needs around grief, spirituality, and emotional support
- • Sets precedent for cross-disciplinary collaboration in tech
❌ Cons
- • 15 people for two days won't fundamentally change a product
- • Could be seen as sophisticated PR rather than substantive change
- • Christian perspectives don't represent all users or belief systems
- • Risk of imposing specific religious values on a global product
What Happens Next
Neither Anthropic nor the summit participants have said whether this will become a regular thing. Will we see Buddhist monks advising on Claude's approach to mindfulness? Rabbis consulting on Talmudic reasoning? Secular ethicists pushing back on the religious framing entirely?
These questions matter because Claude has millions of users worldwide. What moral framework should guide an AI that talks to people from every culture, religion, and belief system? That's not a problem you solve in a two-day retreat.
But you have to start somewhere.
The Real Question Nobody's Asking
If we're consulting religious leaders about AI's moral behavior, shouldn't we also be asking users what they want? The people actually talking to Claude every day might have strong opinions about how it handles sensitive topics, and they weren't in the room.
My Take
Here's what I think: this is genuinely interesting, and I'm glad it happened.
Do I think Claude has a soul? No. Do I think Anthropic cracked the code on AI ethics? Also no. But the fact that a major AI company is even asking these questions publicly, bringing in outside perspectives, and admitting they don't have all the answers? That's more than most tech companies do.
The alternative is what we usually get: move fast, break things, and figure out the ethical implications after the lawsuits start.
Whether or not you're religious, whether or not you think AI deserves moral consideration, the questions Anthropic is wrestling with are real. How should an AI respond to human suffering? What values should guide its behavior? Who gets to decide?
Those questions aren't going away. If anything, they're going to get harder as these systems become more capable. And I'd rather have companies awkwardly consulting priests than pretending the questions don't exist.
For readers interested in the strategic implications of AI development and why companies like Anthropic are investing in long-term thinking
Frequently Asked Questions
Did Anthropic only consult Christian leaders?
This particular summit focused on Christian perspectives, but Anthropic has consulted with various ethicists and experts in the past. The company hasn't announced plans for summits with other religious or philosophical traditions.
Will this change how Claude actually behaves?
It's unclear. Anthropic hasn't specified what concrete changes, if any, will result from these conversations. The summit appears to be more about gathering perspectives than implementing specific features.
Why are AI companies using spiritual language?
The questions raised by advanced AI, like consciousness, moral worth, and the nature of intelligence, overlap significantly with questions traditionally addressed by religion and philosophy. Some observers think the spiritual framing helps communicate the stakes involved.
Is Claude actually conscious?
No current scientific consensus supports the idea that AI systems like Claude are conscious. However, the question of how to determine machine consciousness remains actively debated by philosophers and researchers.
Sources & Credits
Originally reported by The Decoder — Matthias Bastian
Huma Shazia
Senior AI & Tech Writer
Related Articles
Browse allYour Next Laptop Might Have Dedicated AI Keys, and Google Is Behind the Whole Thing

Stop Paying $20/Month for GitHub Copilot. Build Your Own in 10 Minutes.
The Feds Just Stepped In to Save Kalshi From Arizona's Criminal Charges

Dragon Age: Inquisition Outlasted Anthem By Nearly a Decade, and That Says Everything About Gaming Right Now
Also Read

Anno 117 Prophecies of Ash DLC: Build Your Roman City Next to an Active Volcano
