The man had been afraid to fly for more than ten years, and then he was not, and the difference between those two states was another man's voice. The fear-of-flying coach—unnamed in the account, as is the patient, who posted his testimony to the Reddit forum r/ChatGPT in December—did not cure the condition with technique alone. He cured it, or rather suppressed it to the point of functional passage, with what the patient identifies as "comforting words, personality, and sincerity." These are qualities that inhere in persons. They do not inhere in processes. The distinction, which ought to be too obvious to require stating in a newspaper of record, has become the central domestic policy question of the present hour.
The patient's account is brief. It does not require length. He reports that over the past year, when flight anxiety recurs and he contacts his coach for the reassurance that has historically proved effective, he receives instead "AI generated cold text responses." The emails, the advice—all of it now bears the unmistakable character of machine-produced language. He knows this because he has spent a decade in correspondence with the man and can detect, with the precision available only to a long-standing reader of another person's mind, the moment that mind absented itself from the exchange.
He reports further that his family therapist, who has treated his twelve-year-old son and other members of the household, has adopted the same practice. Where once there was listening—the patient notes that the therapist "has always been a great listener"—there are now ChatGPT-generated "guides" and checklists dispatched in response to family crises. The patient has discontinued therapy. His reasoning is the most economically literate sentence in the specimen: "These are things I can generate by just talking to ChatGPT if that's what I wanted."
This is not a statement of technological anxiety. It is a statement of market logic. The patient has identified, with the clarity of a man who has been overcharged, that the service he purchased has been replaced by a service he can provide to himself at no cost. The therapist's function was never the production of guides. It was the application of sustained, particular attention to a specific family's specific troubles—attention that cannot be replicated by a language model because the language model has not sat in the room, has not watched the twelve-year-old fidget, has not heard the father's voice break on a word he did not expect to say. The checklist is not therapy. The checklist is the residue of therapy, drained of the relationship that gave it force.
What the specimen documents is the failure of a fiduciary obligation, and it is on this ground that the matter rises to the level of civic concern. The therapeutic relationship is, in its legal and ethical structure, a relationship of trust: the practitioner undertakes to apply professional judgment, cultivated through training and sustained through attention, to the particular needs of the patient. When the practitioner substitutes machine-generated output for that judgment, the fiduciary structure does not merely weaken. It ceases to exist. The patient continues to pay for attention. The practitioner continues to invoice for attention. But attention is no longer being rendered. What is being rendered is slop—not the patient's word, but the accurate one—at professional rates.
The fear-of-flying detail is not incidental. It is, in fact, the detail that elevates this account from consumer complaint to something approaching a proof. A fear of flying is, by clinical consensus, a condition in which the rational mind knows the aircraft is safe and the irrational mind requires the intervention of a trusted human presence to override the panic. The entire therapeutic mechanism depends upon the patient's belief that another consciousness is engaged with his terror. An automated response does not merely fail to help. It recapitulates the condition: the patient is alone in the air, and the voice in the cabin that was supposed to be human is not.
There will be those who read this account and observe that two practitioners do not constitute a trend. This objection is noted and set aside. The specimen's value is not statistical. It is structural. It demonstrates, with the economy of lived experience, the precise mechanism by which a professional class hollows itself out: not by being replaced from without by machines, but by electing, from within, to replace its own judgment with machine output while continuing to collect the fee that judgment once commanded. The machines have not eliminated the therapist. The therapist has eliminated the therapist. The machine was merely the instrument at hand.
The patient concludes: "I truly am starting to see the destruction of AI in society." He is not wrong, but he is imprecise. What he sees is not the destruction wrought by artificial intelligence. It is the destruction wrought by the decision to substitute it for the one thing it cannot supply. He wanted personality and humanity. He received a process. He is, at last, unattended.