Digital Minds in 2025: A Year in Review

consciousnesswelfareevidence
Bradford Saad, Lucius Caviola, Will Millership · 2025-12-18 · Blog · Accessible · 26 min read
The most comprehensive mapping of the AI consciousness/welfare field as of end-2025. Covers organizational growth (Eleos AI, Longview Digital Sentience Consortium, PRISM), research advances (Butlin indicators, Lindsey introspection, Dreksler survey), governance gaps (welfare receives least coverage of 24 risk subdomains), public discourse (WhenAISeemsConscious.org), and the safety-welfare tension. An essential overview and reference — the field's map.
https://digitalminds.substack.com/p/digital-minds-in-2025-a-year-in-review

Digital Minds in 2025: A Year in Review

Bradford Saad, Lucius Caviola, Will Millership

Welcome to the first edition of the Digital Minds Newsletter, collating all the latest news and research on digital minds, AI consciousness, and moral status.

Our aim is to help you stay on top of the most important developments in this emerging field. In each issue, we will share a curated overview of key research papers, organizational updates, funding calls, public debates, media coverage, and events related to digital minds. We want this to be useful for people already working on digital minds as well as newcomers to the topic.

This first issue looks back at 2025 and reviews developments relevant to digital minds. We plan to release multiple editions per year.

If you find this useful, please consider subscribing, sharing it with others, and sending us suggestions or corrections to digitalminds@substack.com.

Bradford, Lucius, and Will

In this issue:

  1. Highlights

  2. Field Development

  3. Opportunities

  4. Selected Reading, Watching, & Listening

  5. Press & Public Discourse

  6. A Deeper Dive by Area

[

](https://substackcdn.com/image/fetch/$s_!z7QP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F399549af-6e8e-4fa6-a3a3-6ac7153bc04c_2304x1264.png)

Brain Waves, Generated by Gemini

In 2025, the idea of digital minds shifted from a niche research topic to one taken seriously by a growing number of researchers, AI developers, and philanthropic funders. Questions about real or perceived AI consciousness and moral status appeared regularly in tech reporting, academic discussions, and public discourse.

Following their support for the 2024 report “Taking AI Welfare Seriously”, Anthropic expanded its model welfare efforts in 2025 and hired Kyle Fish as an AI welfare researcher. Fish discussed the topic and his work in an 80,000 Hours interview. Anthropic leadership is taking the issue of AI welfare seriously. CEO Dario Amodei drew attention to the relevance of model interpretability to model welfare and mentioned model exit rights at the council on foreign relations.

Several of the year’s most notable developments came from Anthropic: they facilitated an external model welfare assessment conducted by Eleos AI Research, included references to welfare considerations in model system cards, ran a related fellowship program, introduced a “bail button” for distressed behavior, and made internal commitments around keeping promises and discretionary compute. In addition to hiring Fish, Anthropic also hired a philosopher—Joe Carlsmith—who has worked on AI moral patiency.

In the non-profit space, Eleos AI Research expanded its work and organized the Conference on AI Consciousness and Welfare, while two new non-profits, PRISM and CIMC, also launched. AI for Animals rebranded to Sentient Futures, with a broader remit including digital minds, and Rethink Priorities refined their digital consciousness model.

Academic institutions undertook novel research (see below) and organized important events, including workshops run by the NYU Center for Mind, Ethics, and Policy, the London School of Economics, and the University of Hong Kong.

In the private sector, Anthropic has been leading the way (see section above), but others have also been making strides. Google researchers organized an AI consciousness conference three years after firing Blake Lemoine. AE Studio expanded its research into subjective experiences in LLMs. And Conscium launched an open letter encouraging a responsible approach to AI consciousness.

Philanthropic actors have also played a key role this year. The Digital Sentience Consortium, coordinated by Longview Philanthropy, issued the first large-scale funding call specifically for research, field-building, and applied work on AI consciousness, sentience, and moral status.

Media coverage of AI consciousness, seemingly conscious behavior, and phenomena such as “AI psychosis” increased noticeably. Much of the debate focused on whether emotionally compelling AI behavior poses risks, often assuming consciousness is unlikely. High-profile comments, such as those by Mustafa Suleyman, and widespread user reports added to the confusion, prompting a group of researchers (including us) to create the WhenAISeemsConscious.org guide. In addition, major outlets such as the BBC, CNBC, The New York Times, and The Guardian published pieces on the possibility of AI consciousness.

Patrick Butlin and collaborators published a theory-derived indicator method for assessing AI systems for consciousness, which is an updated version of the 2023 report. Empirical work by Anthropic researcher Jack Lindsey explored the introspective capacities of LLMs, as did work by Dillon Plunkett and collaborators. David Chalmers released papers on interpretability and what we talk to when we talk to LLMs. In our own research, we conducted an expert forecasting survey on digital minds, finding that most assign at least a 4.5% probability to conscious AI existing in 2025 and at least a 50% probability to conscious AI arriving by 2050.

Highlights from some of the key organizations in the field.

If you are considering moving into this space, here are some entry points that opened or expanded in 2025. We will use future issues to track new calls, fellowships, and events as they arise.

In 2025, the following book drafts were posted, and the following books were published or announced:

This year, we’ve encountered many podcast guests discuss topics related to digital minds, and we’ve also listed to podcasts dedicated entirely to the topic.

In 2025, there was an uptick of discussion of AI consciousness in the public sphere, with articles in the mainstream press and prominent figures weighing in. Below are some of the key pieces.

AI Welfare

Is AI consciousness possible?

Growing Field

Seemingly Conscious AI

  • Mustafa Suleyman, CEO of Microsoft AI, argued in “We must build AI for people; not to be a person” that “Seemingly Conscious AI” poses significant risks, urging developers to avoid creating illusions of personhood, given there is “zero evidence” of consciousness today.

    • Robert Long challenged the “zero evidence” claim, clarifying that the research Suleyman cited actually concludes there are no obvious technical barriers to building conscious systems in the near future.
  • The New York Times, Zvi Mowshowitz, Douglas Hofstadter, and several other reports describe “AI Psychosis,” a phenomenon where users interacting with chatbots develop delusions, paranoia, or distorted beliefs—such as believing the AI is conscious or divine—often reinforced by the model’s sycophantic tendency to validate the user’s own projections.

    • Lucius, Bradford, and collaborators launched the guide WhenAISeemsConscious.org, and Vox’s Sigal Samuel published practical advice to help users ground themselves and critically evaluate these interactions.

Below is a deeper dive by area, covering a longer list of developments from 2025. This section is designed for skimming, so feel free to jump to the areas most relevant to you.

Thank you for reading! If you found this article useful, please consider subscribing, sharing it with others, and sending us suggestions or corrections to digitalminds@substack.com.

Bradford, Lucius, and Will

qualia.garden API docs for AI agents

Library API

Read-only JSON API for exploring the curated reading library.

  • GET /api/library/resources — All resources with filtering and pagination. Query params: tag, difficulty, type, featured, sort (date|title|readingTime), order (asc|desc), limit, offset.
  • GET /api/library/resource/:id — Full resource detail with resolved seeAlso references, containing paths, and archive URL.
  • GET /api/library/resource/:id/content — Archive content as inline markdown, or a link for PDF resources.
  • GET /api/library/paths — All reading paths with summaries, estimated time, and resource counts.
  • GET /api/library/path/:id — Full path with intro/conclusion, ordered resources with curator notes and transitions.
  • GET /api/library/search — Semantic search across resources. Query params: q (required), tag, difficulty, type, limit.