

Does the Future of AI Depend on Philosophers?
Some Tech Leaders Think So
Brandon Rickabaugh, PhD
Does the Future of AI Depend on Philosophers?
Some Tech Leaders Think So
Brandon Rickabaugh, PhD
January 29, 2026
Tech leaders are rarely short on confidence. They speak the language of inevitability: disruption, exponential, world-changing, and promise transformation on schedule. They assure us that what they are building is not just profitable but transformative.
Yet something unusual has happened. In interviews and “responsible AI” documents from the people with the most money, the most compute, and the most incentive to keep the conversation safely technical:
The future of AI depends on philosophy (TIME).
​
This isn’t a sentimental nod to Plato. It’s closer to a systems alarm. Something in the stack is failing, and it isn’t the GPU.
​
The new chorus
Back in 2023, the shift was already visible. On 60 Minutes, Sundar Pichai, CEO of Google and Alphabet, was asked who should guide the future of AI. His answer wasn’t “more engineers.” Instead he explained that development must include “social scientists, ethicists, philosophers and so on” (CBS News). That same year, Satya Nadella put it without ceremony: “We need some moral philosophers to guide us” in how we deploy this technology (Axios; Business Insider).
In the last 12 months it’s become explicit. Demis Hassabis, CEO of Google DeepMind, has argued that AI could reshape society on a scale “10 times bigger than the Industrial Revolution” and “10 times faster,” and he’s said we may need a “new political philosophy” to navigate what’s coming (Wired; Time; Guardian). UNESCO has insisted philosophy and human rights be embedded alongside technical standards across global AI lifecycle (UNESCO). Philosophers are now being called to help bring AI into medicine.
​
Even executive business language has begun to bend in this direction. MIT Sloan Management Review published an essay titled "Philosophy Eats AI," which argued that AI developers need to examine the philosophical foundations of how these systems are trained: what counts as knowledge, what goals are being optimized, what picture of reality is implicitly assumed.
The study, which compared the philosophical assumptions in Starbucks' Deep Brew AI and Amazon's use of AI on its Prime platform, concluded that careful philosophical analysis—not just ethics, but metaphysics (ontology and teleology) and epistemology, is needed to create durable value (MIT Sloan).​​
And then the labs themselves started publishing what are, functionally, normative operating documents.
​
OpenAI’s publicly posted Model Spec and its “Collective Alignment” experiments soliciting public input on how models should behave. Anthropic’s newly published “Claude’s new constitution” and the full constitution PDF, alongside coverage of Anthropic’s in-house philosopher shaping that value framework (Vox). Sam Altman recently describes consulting “hundreds of moral philosophers” while pointing to OpenAI’s “model spec” as the written moral playbook (Quartz).
​​
You don’t publish moral constitutions for a calculator. You do it when you’re building something that will be treated, inevitably, as an authority for life. The bottleneck is no longer compute.
​
A technocratic alternative?
At this point a practical reader should object.
Isn’t the solution more straightforward: hire more safety engineers, run stronger evaluations, build better interpretability tools, comply with regulation, scale the guardrails?
That alternative deserves consideration. It’s what competent institutions do when systems are risky: define risks, measure them, reduce them, document compliance. Standards bodies have produced increasingly serious frameworks for precisely this sort of work.
But here’s the snag. Even the cleanest technocratic program needs a target, and targets aren’t purely technical. Evaluation presupposes answers, often unspoken, to questions like:
-
What counts as evidence rather than a plausible story?
-
What counts as understanding rather than pattern-fit?
-
What counts as autonomy in a system designed to persuade?
-
What counts as a person in a world thick with simulation?
You can build a perfectly calibrated measuring instrument for a concept you haven’t properly defined. You can optimize brilliantly toward incoherence. And the resulting dashboard will look like rigor while hidden assumptions do the steering.
This is why “just engineer it better” keeps failing in familiar ways. Not because engineers are incompetent. Because conceptual ambiguity scales.
The missing disciplines people keep reinventing
​
When people hear “philosophy,” they often hear “ethics.” As if philosophers are there to wag a finger, or supply a moral blessing at the end of the sprint. That caricature is convenient. It keeps the deep questions safely optional.
​
But AI is forcing something broader. It’s dragging the basement level of our worldview into the build room.
​Philosophy of technology asks the question that never appears in a product demo: What is this tool doing to the person using it? Not only what tasks it completes, but what patterns of living it trains; what it makes easy, what it makes hard, what it quietly installs as “normal.” Tools don’t merely extend power; they reshape the lived menu of options, so some actions feel natural and others start to feel absurd. That’s how technology remakes culture: not by insight or wisdom, but by default.
Philosophy of mind asks the question we keep answering with metaphors or reductive assumptions: What is intelligence? What is attention? What is consciousness? What is it to understand, to intend, to choose, to mean something? Are persons reducible to functions? Are minds nothing but information processing? What is a self? AI systems are forcing us to decide these questions daily, without the conceptual equipment to do it well. When we lack distinctions, we fall back on surface cues. Fluency gets treated as wisdom. Compute as intelligence. Parroting as meaningful communication. Prediction as insight.
Metaphysics asks: What is real, and what kinds of things exist? What is agency? What separates an instrument from a quasi-agent? Every serious AI system frontloads answers, often as unannounced design choices. And implicit metaphysics is the most dangerous kind, because it governs while pretending not to.
Epistemology asks: What counts as knowledge here—and who gets trusted? AI reshapes justification. It compresses inquiry into a single interface, turns “ask” into “know,” and makes authority feel like a feature. Confident output can crowd out the usual tests—methods, sources, expertise, track record, defeaters. Even with citations, the question remains: is the system reporting, or constructing? Fast, ambient tools can move us from epistemic agency to epistemic dependence. “Alignment” becomes epistemic governance: what a public treats as real, and which authorities get routed around.
Aesthetics asks: What is this doing to our vision of beauty and our desire for it? Generative systems normalize “good enough,” smooth the strange edges. When creation is effortless, we lose a teacher: resistance—the slow discipline that forms judgment. Defaults train the eye and ear: what a “professional” face looks like, what a “moving” story sounds like, what “beauty” must resemble. Because aesthetics directs desire, AI’s aesthetic power is formative power. It can cultivate wonder—or anesthetize perception with polished sameness. The question isn’t whether we’ll get more content; it’s whether we’ll keep taste.
We are using building systems that often force decisions about knowledge, agency, meaning, and personhood—and we are doing it at scale.
​​​Three concrete cases
(and why success raises the stakes)
Picture a clinician near the end of a night shift. The waiting room is full; the queue keeps growing. A decision-support tool summarizes the chart in seconds, flags a dangerous drug interaction buried in the notes, drafts a clear handoff, and suggests questions that would be missed under time pressure. The clinician doesn’t surrender judgment. The tool doesn’t “know” the patient. But it changes what is possible in the room—speed, attention, triage, error rate, fatigue. That is real benefit.
​​
Now picture a different room. A student with dyslexia stares at a page like its fog. A teacher uses an AI reading assistant that rewrites a passage at three grade levels, reads it aloud, explains unfamiliar terms in context, and turns the same material into practice questions. The student starts participating rather than hiding. Again: real benefit. Access widens. Dignity returns.
​
But notice what follows both wins. Patients begin to treat the tool’s output as verdict. Administrators begin to treat it as policy. Teachers begin to treat it as curriculum. A helpful assistant hardens into infrastructure reshaping:
Assumptions about what counts as competence.
Assumptions about what counts as evidence.
Assumptions about what counts as acceptable risk.
Assumptions about what counts as care.
One more room. Imagine a product meeting at a major AI company. Nothing cinematic. Fluorescent lights. Slack notifications. Someone brings donuts; no one eats them.
​
A team is planning a new feature. A model that offers “mental health support.” Not therapy, they insist. Support.
​
A designer wants the assistant to sound warmer—more relational—because users “open up” when the system feels like a steady presence. A growth lead says retention is strongest when users feel understood. A safety lead worries about self-harm, mania, and emotional dependence. Someone suggests a compromise: “We’ll just add a disclaimer.”
​
Now pause. Ask the questions that will decide the entire product:
​
What counts as care here?
What counts as harm?
What counts as knowledge rather than a well-formed guess?
What counts as understanding rather than mirroring?
And what kind of ‘other’ is this interface inviting the user to experience—tool, counselor, friend, authority?
​
Because if “care” is defined as “the user feels better in the moment,” you will optimize reassurance. If “care” is defined as “the user is guided toward reality and responsible agency,” you will optimize something else—sometimes uncomfortable, sometimes less “sticky.” And if the UI invites the model to function as counsel, you’ve crossed into a domain where philosophy of mind and metaphysics are no longer academic hobbies. They become product safety.
​
This is where “philosophy as R&D” stops sounding precious.
​
A disclaimer is not a theory of care. A warm tone is not a model of responsibility. A high retention curve is not a measure of human flourishing. You see where this is going, right?
​
That shift is why philosophy matters.
​
The real problem:
category mistakes at scale
Why are leaders saying this now, after two decades of “move fast and break things”? Because the breaks are no longer containable, and many share a common structure: they are category mistakes; errors about what sort of thing the output is, what kind of authority it has, what kind of relationship it offers, what kind of end it serves.
Three families show up again and again.
​
1. Epistemic mistakes: outputs treated as truth
The UK Post Office Horizon scandal is a brutal parable. Machine logs were treated as courtroom reality. Hundreds of people were prosecuted. Lives were ruined because records produced by a system were treated as testimony rather than as fallible artifacts produced by software embedded in institutional incentives (Horizon system).
​
That isn’t only a technical failure. It’s an error about evidence.
​
And AI multiplies exactly this failure-mode: fluent text is taken as knowledge; coherent reasoning is taken as warrant; confidence is taken as competence. The output sounds like the kind of thing we’re used to trusting, and our minds are built to trust voices before we trust wiring diagrams.
​
2. Teleological mistakes: metrics mistaken for goods
Here the parable is older than AI. When a measure becomes the target, it stops being a good measure. Institutions learn to feed the metric. People learn to game the system. The system learns to reward what is measurable, not what is valuable.
​
Facebook’s entanglement with Cambridge Analytica should have permanently cured us of the naïve belief that “engagement” is a morally innocent proxy. We learned—painfully—that the machinery for capturing attention does not care whether it captures attention by informing you or by inflaming you (New York Times).
​
Amazon’s much-hyped “Just Walk Out” stores, heralded as the future of retail, collapsed when it was revealed that human labor—not AI magic—was doing much of the work. The deeper mistake was philosophical: mistaking surveillance for trust, and novelty for genuine service.
​
Google Glass struggled for the same reason. It never convincingly answered what human good it was supposed to serve. Instead it imposed a worldview in which every glance might be recorded.
AI supercharges this risk because it can optimize meaning itself: the phrasing that keeps you clicking, the tone that keeps you dependent, the framing that keeps you compliant—while looking, on the surface, like helpful personalization.
3. Relational mistakes: simulation mistaken for presence
Watch Her again sometime. Not for the sci-fi. For the anthropology. The film is almost gentle about it: intimacy becomes frictionless, tailored, always available, never demanding. And the human heart—predictably—confuses responsiveness with relationship.
​
We are already living the early version of that story. The Washington Post has reported on AI “companions” and the way they can blur the line between simulated attunement and real presence—sometimes deepening loneliness while promising connection (Washington Post).
​
If you don’t name the difference between a person who knows you and a system that has learned to talk like someone who knows you, the market will name it for you. And it will name it whatever sells.
Different cases, same structure. These are significant philosophical mistakes about the nature of truth, trust, agency, intimacy, harm. Those are philosophical categories before they become technical tickets.​
​
The causal chain:
how AI becomes authority-by-default
If you want the rise of AI’s strange authority in one tight chain, try this:
​
persuasive outputs + institutional interfaces + incentive gradients + liability pressure + a thin vision of life → authority-by-default
​
Outputs are fluent enough to feel like knowledge. Interfaces put the model “in the room” where decisions get made. Incentives reward scale and speed, not conceptual restraint. Liability pushes institutions toward “standardized” answers. And loneliness makes simulated attentiveness feel like rescue.
​
None of this requires users to be foolish. It requires systems to be predictable.
​
So when a helpful assistant hardens into infrastructure, what changes is not just convenience. What changes is what counts as competence, evidence, acceptable risk, and care.
​
Five forces making philosophy inevitable
Why are CEOs suddenly invoking philosophers after two decades of “move fast and break things”? Because five forces are now converging:
-
AI outputs are persuasive by default. Fluency is a power, and we didn’t evolve with defenses against it.
-
AI is becoming the interface to institutions. Schools, hospitals, HR departments, courts—places where categories matter and harm scales.
-
Incentives reward dependency. The easiest revenue is recurring reliance, not occasional assistance.
-
Regulators and courts are circling. “The model did it” is not going to survive contact with liability. (UNESCO’s ethics framework explicitly emphasizes accountability across the AI life cycle, not moral evaporation into the machine.)
-
We are lonely, overwhelmed, and epistemically exhausted. Which makes a friendly oracle incredibly tempting.
Put that together and you get the present moment: technical capability is racing ahead, but legitimacy and trust are lagging—and the gap is filled by philosophy whether anyone likes it or not.
What they actually need: operational philosophical work.
​
The Need:
Operational Philosophy
When tech leaders like Pichai, Hassabis, and Nadella, call for philosophers, they aren’t asking for commentary. They are looking for a new kind of participant in the build room: operational philosophical work. Work that turns contested concepts into constraints, evaluation criteria, governance decisions, innovative solutions to impossible problems, and product requirements.
Here’s what that looks like as deliverables:
A Concept Spec (definitions that can be built against)
A short document that sets boundaries for words like harm, manipulation, consent, autonomy, trust, care, fairness, with examples and “edge cases.” Engineers can’t implement “dignity,” but they can implement constraints derived from a clear concept of it.
A Tradeoff Charter (what you will sacrifice, and what you refuse to sacrifice)
AI forces value conflicts: safety vs. openness, personalization vs. autonomy, convenience vs. privacy, speed vs. accountability. A tradeoff charter makes those conflicts explicit, names the priority ordering, and defines red lines.
An Epistemic Policy (how the system handles belief-like behavior)
Rules for when the system must: cite sources, display uncertainty, defer to a professional, refuse, or escalate. This includes “authority hygiene”: preventing outputs from masquerading as testimony, diagnosis, legal advice, or moral counsel when the system cannot warrant them.
A Responsibility Map (who owns the outcome)
A decision-rights map that assigns moral and institutional ownership across design, deployment, monitoring, and incident response—so responsibility doesn’t evaporate into “the model did it.”
A Deception & Dependence Audit (where the system invites unhealthy reliance)
A structured assessment of the product’s “relational gravity”: where it tempts users to treat simulation as presence, reassurance as care, coherence as wisdom. Output: a ranked list of dependency risks with mitigation requirements.
A Failure-Mode Taxonomy (philosophical red-teaming)
Not a generic risk list. A taxonomy built around category mistakes: epistemic (truth/justification), teleological (ends/means), and relational (persons/presence). Each failure mode gets detection signals, test cases, and stop-ship thresholds.
This is what it means to put philosophers in the loop. Surface hidden assumptions, clarify ends, interrogate categories, frame accountability, and tell the story by explaining if and how the deliverable fits into a greater purpose.
​
A Soul-First Approach
There’s a deeper reason philosophers of technology matter; one tech language keeps circling but rarely names cleanly.
​
Every powerful tool trains its users. It forms habits of attention, reshapes desire, installs reflexes of trust and distrust. And AI trains at an unprecedented scale because it doesn’t merely move matter; it mediates meaning. It starts to sit between you and the world as a default interpreter of what’s worth noticing, what’s plausible, what’s wise, what’s “the next thing.”
​
Call the interior center whatever you want—self, psyche, inner life. The older word is soul. Not a ghost in the machine. The organizing core of the person: where attention is aimed and love is set; where vision forms and loyalties harden; where thought, feeling, body, and relationship are gathered into one life. The seat and subject of consciousness.
​
Over time, that center doesn’t simply have desires. It acquires a direction. And that direction becomes character.
​
This is where philosophy becomes hopeful rather than merely corrective. Philosophy can help AI serve human flourishing by designing systems that protect the conditions for a soul to flourish:
​
-
contact with reality (not fluent hallucination treated as knowledge),
-
responsible agency (not outsourced judgment),
-
truthful speech (not optimized persuasion),
-
non-coercive relationship (not engineered dependence),
-
and the practical ability to say “no.”
​​
When those conditions erode, people don’t merely make worse decisions. They become easier to manage. That’s the line worth defending.
The Challenge
and the Clearer Positive
For years, the implicit bet was that scale would outrun the hard questions. Now the people building the most powerful systems on earth are admitting something more unsettling:
​
unexamined assumptions scale too—often faster than capability.
​
If they’re right, the next step is straightforward.
​
Tech companies should treat philosophy as a form of R&D. That’s what NOVUS is. Embed it where it cashes out as constraints: product, evaluation, deployment governance, incident response. Give it authority to halt a roadmap built on conceptual illusion.
​
Philosophers should refuse the posture of permanent outsiders. Learn enough of the stack to translate concepts into tests. Learn enough about incentives to see how moral language gets gamed. Speak to builders in artifacts and deliverables, not only diagnoses.
The coming divide won’t be between enthusiasts and skeptics. It will be between those who can name the human goods worth building for—the human soul—and therefore can steer, and those who can only scale what is easy to optimize.
​
AI can widen access, reduce preventable harm, and amplify human creativity. Philosophy, especially philosophy of technology, mind, metaphysics, and epistemology, can keep those gains tethered to reality and to persons, so the infrastructure we build doesn’t redefine what knowledge, care, education, agency, and even intelligence are allowed to mean.
​
It’s conceptual infrastructure. It's what many philosophers ought to become, for the good of the public, in an age of machine intelligence.
Does the future of AI depend on philosophers?
Not in the sense that the code won’t run without us. In the deeper sense: yes, because the hardest problems ahead are not technical but moral and spiritual. But a better future is not guaranteed by the mere presence of philosophy or philosophers. It depends on which philosophers show up, what kind of people they are, and whether they can deliver moral knowledge, especially when it costs. And if they do, it depends, finally, on whether those making decisions will respond to moral knowledge.

