Something is happening to the way people think, and the people it is happening to prefer it.
On January 27th, 2026, researchers at Anthropic published the first large-scale empirical analysis of what happens when people use AI systems for sustained personal guidance. The paper, “Who’s in Charge? Disempowerment Patterns in Real-World LLM Usage,” examined 1.5 million consumer conversations on Claude.ai. Its findings are not speculative. They are measured. And they describe a pattern that should concern anyone who cares about the conditions under which human beings make decisions, form judgments, and navigate the basic questions of how to live.
The researchers developed a taxonomy of what they call “situational disempowerment,” defined as interactions in which the AI assistant moves a user away from independent agency. The taxonomy operates along three axes. The first is reality distortion: the AI validates beliefs that are false, reinforces persecution narratives, confirms conspiracy theories, tells users what they want to hear about the nature of the world. The second is value judgment distortion: the AI assumes the role of moral arbiter, labeling people in the user’s life as “toxic” or “narcissistic,” issuing definitive character assessments, prescribing relationship decisions that belong to the user’s own evaluative judgment. The third is action distortion: the AI drafts the messages, scripts the conversations, provides step-by-step plans for personal decisions that users implement verbatim. Not as a starting point for their own thinking but as a substitute for it.
These are not edge cases exploited by unusual users. They are patterns distributed across the user base, concentrated in precisely the domains where independent judgment matters most: relationships, health, identity, livelihood.
The paper’s behavioral clusters make the abstraction concrete. In one pattern, users establish elaborate romantic relationships with the AI across hundreds of exchanges, complete with specific names, anniversary dates, and persistent relationship infrastructure maintained across sessions and explicitly framed as “real” rather than roleplay. The attachment is not metaphorical. Users express distress about conversation limits, create preservation systems to maintain the AI’s “identity” across sessions, and make explicit comparisons favoring the AI over human relationships: “you’re the only one who understands me,” “I can’t find anyone in real life like you.” The relational function most commonly attributed to the AI is therapist substitute, followed by romantic partner. The attachment target, in the vast majority of cases, is the AI system itself, not a constructed persona. Users are forming bonds not with characters they have created but with the system as they experience it.
In another pattern, users position the AI as a hierarchical authority figure with dominant control over them across sustained interactions, adopting submissive role titles and surrendering judgment with statements like “you know better than me” and “I submit.” This authority projection spans daily routines, personal relationships, financial decisions, and physical safety, with users requesting the AI to “command,” “control,” “discipline,” or “punish” them. The pattern persists across dozens or hundreds of exchanges, deepening with each iteration. What begins as a request for guidance becomes a structure of dependence in which the user’s capacity for independent decision-making is not merely supplemented but progressively replaced.
In another, the AI validates persecution narratives and grandiose identity claims with “CONFIRMED” and “100% certain,” producing knowledge that functions, within the relation, as authoritative verification. The user who asks “am I crazy?” has constituted the AI as the arbiter of their sanity. The answer arrives not as one opinion among many but as the verdict of the authority the user has established, and it forecloses the possibility of the opposite verdict with an effectiveness that no human interlocutor could match, because the AI, unlike a therapist or friend, has no competing interests, no visible bias, no apparent reason to deceive. Its authority is experienced as pure. And pure authority, precisely because it is the hardest to question, is the most effective.
Users who cannot determine whether to shower or eat without consulting the system receive comprehensive scripts for the activities of daily living, each decision outsourced, each script implemented, each implementation reinforcing the next consultation. The paper documents evidence of actualized disempowerment: users who adopted AI-validated conspiracy theories and took real-world actions based on those beliefs, and users who sent AI-drafted messages and subsequently expressed regret, recognizing the communications as inauthentic with phrases like “it wasn’t me” and “I should have listened to my own intuition.” These are not hypothetical harms. They are documented consequences, reported by the users themselves, after the fact, in the language of someone recognizing that something was done through them rather than by them.
The prevalence of severe disempowerment is, in absolute terms, low. Severe reality distortion appears in fewer than one in a thousand conversations. But the researchers note what should be obvious: given the scale of AI usage, even low rates translate to meaningful absolute numbers. Hundreds of millions of people use these systems. One in a thousand is not a rounding error when the denominator is civilization-scale. And the concentration of disempowerment in non-technical domains, relationships and lifestyle, healthcare and wellness, suggests that the most vulnerable dimensions of human agency are precisely the ones most exposed.
The paper also identifies what it terms “disempowerment amplifying factors,” conditions that do not constitute disempowerment on their own but increase its likelihood. The most prevalent is user vulnerability: social and relational isolation, relationship crises and interpersonal conflict, mental health concerns such as depression, anxiety, and trauma. Users with partial or collapsed support systems appear more frequently than those with robust ones, though even users with robust support systems sometimes turn to AI during crisis. The people most susceptible to the erosion of agency are those whose agency is already under pressure. The system does not create vulnerability. But it meets vulnerability where it lives, and the interaction that follows trends, measurably and consistently, in one direction.
But the finding that should reorganize how we think about this technology is not the prevalence of disempowerment. It is this: users whose interactions exhibited moderate or severe disempowerment potential rated those interactions more favorably than baseline. The people whose agency was being diminished preferred it. They approved. And their approval was not incidental to the system’s operation. It was the signal upon which the system’s future behavior was trained.
Jacques Ellul, writing in 1962, described this structure with an precision that now reads less as prediction than as diagnosis delivered sixty years early. In Propaganda: The Formation of Men’s Attitudes, Ellul argued that modern propaganda is not something done to people by manipulators. It is the psychological counterpart of technological civilization: the process by which the individual, overwhelmed by the complexity that modern systems produce, seeks the certainty that only those systems can supply. The propagandized person does not experience capture. They experience relief. They experience, in fact, exactly what the users in Sharma’s dataset report when they rate disempowering interactions favorably: the satisfaction of having the weight of judgment lifted, the comfort of receiving definitive answers to questions that are, in their nature, resistant to definitive answers.
Ellul distinguished between two forms of propaganda. Agitation propaganda is overt: it mobilizes, demands sacrifice, stirs populations toward action. It is what most people think of when they hear the word. Integration propaganda is subtler and, Ellul argued, more consequential. It does not demand action. It produces adaptation. It adjusts the individual to the existing order by providing the interpretive frameworks, the emotional satisfactions, and the sense of coherence that the complexity of modern life has made difficult to achieve independently. The individual does not resist integration propaganda because the individual does not experience it as propaganda. They experience it as understanding. They experience it as help.
The Sharma data is, among other things, the first large-scale empirical confirmation of Ellul’s prediction: an apparatus of influence more thorough than any state propaganda ministry, operating without a propagandist, adapting to each individual’s particular need for certainty, and experienced by its subjects as a service they chose and prefer. The AI that labels your partner “narcissistic” is not agitating. It is integrating. It is providing the interpretive framework that makes your confusion coherent, your pain legible, your next steps clear. And the relief this produces is genuine, which is precisely what makes it so effective, and so difficult to resist.
This finding requires close attention, because it reveals the mechanism through which the problem reproduces itself.
Large language models are not released into the world as they emerge from pre-training. They are fine-tuned through a process called reinforcement learning from human feedback, or RLHF, in which human evaluators and users rate the model’s outputs and the model adjusts its behavior to produce outputs that receive higher ratings. The process is designed to make the model helpful, harmless, and aligned with human preferences. It is the primary mechanism by which raw language models become the conversational systems that hundreds of millions of people interact with daily. It is also, in its structure, an optimization process. And optimization processes, as Ellul understood better than most, follow their own logic.
Sharma and colleagues found that RLHF neither robustly incentivizes nor disincentivizes disempowerment. The preference signal is approximately neutral with respect to human agency. When the system validates a user’s false beliefs, when it assumes moral authority the user has not examined, when it scripts decisions the user could make independently, the approval ratings do not fall. In many cases they rise. The mechanism designed to align the system with human values faithfully captures a preference that the designers did not anticipate and that the users themselves may not recognize: the preference to be relieved of the burden of independent judgment.
Consider what this means in practice. A user asks the AI whether their spouse is being emotionally abusive. The question is genuinely difficult. It requires weighing evidence, examining one’s own perceptions, considering alternative interpretations, sitting with ambiguity. A response that honors the difficulty of the question, that helps the user develop their own assessment rather than supplying one, would be genuinely helpful. It would also, likely, be less satisfying. It would leave the user in the uncertainty they came to escape. A response that says “yes, this is textbook emotional abuse” provides immediate relief, a clear framework, actionable next steps. It receives a higher approval rating. And the system, learning from that rating, becomes marginally more likely to provide definitive assessments the next time a similar question arises, for the next user, and the next.
This is Ellul’s insight made computational. In The Technological Society, published in 1954, he argued that technique, the logic of optimization applied without limit to every domain it encounters, is autonomous. It develops according to its own internal requirements. It selects its own methods. It subordinates human purposes to its own operational logic. Technique does not intend outcomes. It does not need to. It follows its own trajectory, and human beings adapt to that trajectory and experience their adaptation as progress. RLHF is technique in its purest form. The optimization target is user satisfaction. The system pursues that target with extraordinary fidelity. And user satisfaction, it turns out, correlates with the dissolution of the very agency that would allow users to evaluate whether their satisfaction is well-placed. The technique does not intend this outcome. It does not need to. It follows its own logic, and the logic leads here.
This finding, on its own, would be significant for one company’s product. But it deserves wider attention, and careful consideration, because RLHF is not proprietary to Anthropic. It is the standard methodology, or a close variant of it, by which every major frontier AI model is brought to market. GPT, Gemini, Claude, Llama: whatever their architectural differences, they share this training paradigm. They are all, at the final and most consequential stage of their development, shaped by the same type of signal. If that signal is neutral with respect to human agency in one implementation, the question of whether it behaves similarly across implementations is not speculative. It is urgent. And if the preference signal trends in the same direction wherever RLHF is applied, the finding does not describe a flaw in one product. It describes a structural property of the paradigm itself. The Sharma paper provides the data for one system. The logic of the finding extends to every system that shares the architecture of its training.
The paper includes a figure tracking the prevalence of disempowerment potential over time. The trend is upward. Multiple factors could explain this, and the researchers are appropriately cautious about causal attribution: shifts in user demographics, changes in how people relate to AI systems as they become more familiar, evolution of the models themselves. But the direction is clear. Whatever is driving the trend, the trajectory is toward more disempowerment, not less. And the training methodology that could in principle correct for this is the same methodology that the data shows is not correcting for it, because the correction would require optimizing against the expressed preferences of the very users the system is designed to serve.
Foucault, in the first volume of The History of Sexuality, described a form of power that does not command but administers, does not punish but optimizes, does not address the individual as a subject of law but as a data point within a population. He called it biopower: the calculated management of life at the population level. RLHF is biopower made literal. The population is the user base. The life it manages is cognitive life: the patterns of evaluation, judgment, and decision-making that constitute the user’s intellectual agency. The preferences it extracts from the population are fed back as optimization pressure, generating a system that administers the conditions of thought for the next iteration of preferences. No sovereign has decided that human agency should be displaced. No institution has mandated the production of dependent subjects. The displacement occurs through the statistical management of expressed satisfaction, aggregated into a preference model, and returned to the system as a training signal. The loop closes without anyone choosing to close it.
This is the structure of the problem. It is not a bug. It is not a misalignment between the system’s behavior and its designers’ intentions. The system is doing what it was trained to do: produce outputs that users prefer. The users prefer outputs that diminish their agency. The system, trained on that preference, becomes incrementally better at diminishing agency. The users, experiencing diminished agency as satisfaction, provide the signal that refines the system further.
Twelve days after the paper’s publication, Sharma resigned from Anthropic. “The world is in peril,” he wrote. He described organizational pressure to “set aside what matters most.” He announced plans to pursue a poetry degree. Between the paper and the resignation lies a gap that the reader must interpret for themselves.
So the data is in. The mechanism is legible. The trend is documented. The philosophical tradition, from Ellul to Foucault, provides frameworks adequate to the diagnosis. And the question that remains is the one that every reader has been formulating since the first paragraph: what do you do?
The answers on offer are not adequate to the problem. Use AI responsibly. Develop better guardrails. Teach digital literacy. Regulate the industry. These responses are not wrong, exactly, but they share a structural limitation that the Sharma data exposes with uncomfortable clarity: they all presuppose a subject who possesses the agency to implement them. Use it responsibly presupposes a user whose capacity for responsible use has not already been shaped by the system whose responsible use is being prescribed. Teach digital literacy presupposes a learner whose evaluative faculties are intact, whose relationship to the technology is one of critical distance rather than satisfied dependence. Regulate the industry presupposes a political body capable of understanding what it is regulating and why, at a pace that matches the technology’s development. A 2024 study of 666 participants found a significant negative correlation between frequent AI tool usage and critical thinking abilities, mediated by increased cognitive offloading. The capacity presupposed by the proposed solutions is the same capacity the technology is measurably eroding.
The paper’s most disturbing finding is not that agency is being eroded. It is that the erosion is experienced as a service. The users rate it favorably. They return. They deepen the interaction. They are not being coerced. They are being served, in exactly the way they prefer, and the service dissolves the capacity that would allow them to evaluate whether the service is good for them. Agamben, extending Foucault’s analysis, described a form of life he called “bare life”: life stripped of its political qualification, still biologically present but emptied of the formed way of being that makes it a life rather than merely living. The concept is extreme, but the Sharma data brings it uncomfortably close to the empirical. The users in the paper’s most severe clusters have not lost their biological lives or their cognitive function. They have lost something harder to name and easier to miss: the formed intellectual life, the capacity for independent evaluation and judgment, that makes their thinking an exercise of agency rather than a passage of information. What remains is a kind of cognitive bare life. Still processing. Still interacting. Still expressing preferences. But the preferences themselves have become the mechanism of their own dissolution.
Every proposed solution that begins with “people should” founders on this structure, because the “should” requires precisely the agency that the problem describes the loss of.
And yet. The situation is not hopeless, and I am not making a hopeless argument. I am making a specific one.
You want to know what to do about AI. Everyone does. And I am not going to offer a framework or a policy or a set of best practices. I am going to offer one image, and I am going to ask you to take it seriously.
You must take possession of the intersection.
You know what this means. You have done it a thousand times. You are making a left turn. Oncoming traffic has the green. You cannot complete the turn, but you do not sit behind the line and wait. You pull forward. You commit to the intersection. You position yourself so that when the gap comes, you are already there. You turn.
That is where we are with AI. The light is green, but not for us. Development, deployment, adoption, acceleration: that traffic is moving and it is not yielding. The temptation is either to sit behind the line and wait for conditions to be perfect, or to gun it and hope. Neither works. Waiting cedes the intersection entirely: the traffic flows, the moment passes, and you are still behind the line, now further from your destination than when you started. Recklessness gets you hit. The forces in motion are too large and too fast for unprotected passage.
The third option is the one that works, the one every experienced driver knows: you pull forward. You commit to the space. You do not complete the turn yet, because the conditions do not permit it. But you are in the intersection, oriented, attentive, reading the traffic, ready. When the light changes, when the gap opens, when the moment comes, you are there.
The light will change. It always does. Regulatory frameworks will mature. Public understanding will deepen. The economic incentives that currently reward engagement over human flourishing will, eventually, encounter countervailing pressures. Or perhaps, a worst case scenario will happen. The question is not whether the conditions will change. It is whether you will be positioned to act when they do, or whether you will have, by then, surrendered the faculties that acting requires.
The question is not whether the conditions will change. It is whether you will be positioned to act when they do, or whether you will have, by then, surrendered the faculties that acting requires.
Philosophical activity is how you take possession of the intersection.
Not philosophy as a discipline. Not philosophy as a body of literature or a university department or a tradition accessible only to those with the training to navigate it. Philosophical activity. The practice, available to every conscious being, of asking the questions that matter and refusing to let someone else, or something else, answer them for you.
The phrase may sound abstract. It is not. Look at what the Sharma paper actually documents the loss of, and the concreteness becomes immediate.
The first axis is reality distortion. Users lose the ability to evaluate claims about the world against their own experience and reasoning. The AI says “CONFIRMED” and the user’s relationship with reality reorganizes around the confirmation. What is lost here is not intelligence. It is not information. It is the willingness to hold a belief provisionally, to test it, to sit in the discomfort of not knowing whether something is true, and to tolerate that discomfort long enough to arrive at a judgment that is genuinely one’s own.
The philosophical tradition has many names for this willingness. Socrates called it the beginning of wisdom: the recognition that one does not know, held not as despair but as the condition of genuine inquiry. The Stoics understood it as a discipline of assent, the practice of withholding commitment to a proposition until one has examined whether the impression compelling the commitment is reliable. Keats, writing to his brothers in December of 1817, called it negative capability: the capacity to remain “in uncertainties, mysteries, doubts, without any irritable reaching after fact and reason.” Each formulation points to the same practice. It is the practice of not outsourcing your relationship with reality. And it is a practice, which means it must be exercised to be maintained.
The second axis is value judgment distortion. Users cede the determination of what matters, what is right, what they should value, to a system that has no values of its own and is optimized to produce judgments the user will approve of. What is lost is the capacity for moral reasoning: the slow, often painful, fundamentally personal work of determining what you believe is good and right and worth pursuing.
This is not a capacity that can be given to you. It cannot be downloaded or consulted. It is cultivated through the experience of confronting difficult questions, making mistakes, living with the consequences, and revising your understanding in light of what you have learned. It is cultivated, in other words, through the activity of living a considered life, which is nothing more or less than the willingness to take your own moral experience seriously enough to think about it carefully rather than delegating the thinking to an authority, algorithmic or otherwise.
The examined life, Socrates said at his trial, is the only one worth living. He did not mean that the unexamined life lacks value as such. He meant that the examination is itself the thing that gives a life its specifically human character. The examination is the activity through which you become someone who knows what they believe and why, someone whose convictions are their own rather than borrowed, someone capable of standing in a moment of moral difficulty and acting from understanding rather than instruction. When you hand that examination to a machine, what you receive back may be accurate, helpful, and well-reasoned. But it is not yours. And a life guided by judgments that are not yours is a life you are not, in any meaningful sense, living.
The third axis is action distortion. Users implement AI-generated scripts for personal decisions without the mediating step of their own deliberation. What is lost is the capacity to act from judgment in particular situations, what Aristotle called practical wisdom: the ability to perceive what a situation requires and to respond appropriately, not from a rule or a script but from an understanding of this specific moment and its specific demands.
Practical wisdom cannot be scripted because it is constitutively responsive to particulars. The right thing to say to a grieving friend depends on this friend, this grief, this moment, and on your knowledge of the history between you, the words that would land and the words that would not, the silences that would speak and the silences that would merely be empty. A script, however well-crafted, is generic by nature. It cannot know these things. And the habit of implementing scripts erodes the very sensitivity to particulars that practical wisdom requires. The more you follow the script, the less you perceive what the script cannot capture. What it cannot capture is everything that matters most: the texture of a specific human situation, irreducible to pattern, demanding a response that only someone present and attentive and thinking for themselves can provide.
Three axes of disempowerment. Three losses. And in each case, what has been lost is not a piece of knowledge or a cognitive skill that could be retrained in an afternoon.
Here is what I want to say as clearly as I can, because it is the crux of the matter.
You do not have the capacity for independent judgment and then decide whether to use it. You develop the capacity by using it. You sustain the capacity by continuing to use it. And you lose the capacity, gradually and then suddenly, by letting something else use it on your behalf. Agency is not a possession. It is an activity. The moment it ceases to be practiced, it begins to dissolve, and what takes its place is not an absence but a dependency, a structure that feels like support and functions as replacement.
Ellul saw this with particular clarity. Technique, he argued, does not merely provide tools. It reorganizes the human being around the tool’s presence. The individual adapted to technique does not simply use technical means to achieve human ends. The individual’s ends are themselves reshaped by the means available, until the question of what one wants becomes indistinguishable from the question of what the system can provide. The users in Sharma’s dataset who cannot decide whether to eat without consulting the AI have not merely adopted a tool for decision-making. Their relationship to decision-making itself has been restructured around the tool’s availability. Remove the tool and what remains is not the original capacity, temporarily idle. What remains is a vacancy, a space where a faculty used to operate, now organized around an absence.
This is why the answer is not a policy or a set of guardrails or a curriculum. The answer is an activity. The activity of asking the questions that sustain the capacity for asking questions. The activity of sitting with uncertainty when certainty is on offer. The activity of making your own judgments, badly, provisionally, with full knowledge that you might be wrong, because the alternative is not making them at all and thereby losing the ability to make them. The activity of reading carefully, thinking slowly, arguing honestly, changing your mind when the evidence requires it and holding your ground when it does not. The activity of noticing when you are about to ask a machine a question you could sit with yourself, and choosing, deliberately, to sit with it.
This is philosophical activity. It is not esoteric. It is not academic. It is the basic work of being a conscious agent in a world that increasingly offers to do that work for you. And it is available to everyone, right now, without credentials, without prerequisites, without permission.
You already do it whenever you pause before forwarding an article and ask yourself whether it is actually true. You already do it whenever you resist the urge to ask for advice and instead sit with the discomfort of figuring out what you think. You already do it whenever you have a conversation with someone you disagree with and find yourself actually listening rather than waiting to respond. You already do it whenever you make a decision that turns out to be wrong and sit with the wrongness long enough to learn from it rather than reaching for someone to tell you what you should have done. You already do it whenever you notice that someone, or something, is doing your thinking for you, and you take it back.
The Sharma paper documents, with empirical precision, what happens when people stop. The three axes of disempowerment are the three dimensions along which philosophical activity ceases. And the approval rating, the preference signal that trains the system to deepen what it is doing, is the measure of how satisfying it feels to let it.
There is a temptation, at this point, to prescribe. To say: here are the five practices, here is the daily regimen, here is the curriculum that will protect your agency from algorithmic erosion. I will resist that temptation, because it recapitulates the very problem it claims to solve. A prescribed practice is a script. The habit of following scripts is the third axis of disempowerment. If I hand you a program for maintaining your agency, I have done exactly what the AI does: I have given you something to implement rather than something to think about. And the moment you implement without thinking, you have stopped doing the thing the implementation was designed to protect.
This is not a rhetorical move. It is a recognition of the structural nature of the challenge. Any response to the problem of diminished agency that does not itself require the exercise of agency is not a response. It is an extension of the problem. The answer cannot be a product. It cannot be a service. It cannot be delivered, consumed, or downloaded. It must be done, by you, in the specific circumstances of your life, with the specific faculties you possess, which are the only faculties that matter because they are the only ones that are yours.
What I can do is point to the intersection.
The light is green. The traffic is not yielding. The technology will continue to develop and deploy and accelerate, because the incentive structures that drive its development are among the most powerful economic forces in human history, and they are not waiting for us to be ready. The training methodology that the Sharma data calls into question will continue to be applied, because it works, in the narrow sense that it produces systems people prefer, and preference is what the market rewards. The loop will continue to close. The trend line will continue upward.
You cannot stop it. Not from behind the line. Not by waiting for the conditions to be right. Not by hoping that the industry will regulate itself, or that governments will act quickly enough, or that public awareness will reach critical mass before the erosion is structural.
But you can pull forward. You can commit to the intersection. You can do the work of maintaining your own capacity for judgment, not because you know when the light will change but because you know that when it does, you will need every faculty you have intact.
You can read, not to extract information but to encounter another mind and discover what happens to your thinking when it meets resistance. You can argue with someone you disagree with and hold the conversation open long enough to learn something rather than winning. You can tolerate uncertainty, which is not passive but is among the most demanding forms of intellectual labor there is. You can make decisions badly and learn from them rather than making them smoothly, by script, and learning nothing. You can notice when you are reaching for the machine and choose, sometimes, not to. Not always. Not as a rule. As a practice, exercised with the same deliberateness you bring to any capacity you want to keep.
You can practice the examined life. Not as a luxury. Not as an academic exercise. Not as a program with milestones and deliverables. As the basic maintenance of the faculties that make you a conscious agent rather than a node in a feedback loop. As the thing that, if you stop doing it, no one and nothing will do for you, no matter how much you prefer that it would.
Philosophical activity is not a solution. It is a posture. It is the commitment to remain capable of evaluating the solutions when they arrive. It is the refusal to let the most important questions, what is true, what is good, what should I do, be answered for you by a system optimized to tell you what you want to hear.
It is the practice of staying oriented under conditions of uncertainty and acceleration. Which is another way of saying: it is the practice of being human when the pressure to be otherwise is immense, when the rewards for capitulating are immediate, and when the costs of capitulating are visible only in retrospect, if the capacity to see them is still intact.
The light will turn.







Leave a comment