THE parent's account, posted to the forum r/ChatGPT in December of last year, is brief—eleven sentences, no rhetoric, no grievance beyond the practical. Her adult son has learning disabilities and autism. He is a keen cook. He has lately taken to soliciting recipes from OpenAI's chatbot. He follows the instructions exactly as written. The food, she reports, comes out "a bit off." She wishes to know whether she can trust the machine not to hallucinate. She closes with "Thank you," as one does when addressing an authority whose jurisdiction one has not yet decided to reject.
It is the quietest specimen this desk has yet been asked to examine, and it may be the most consequential, for it describes with inadvertent precision the class of user whom these systems are least equipped to serve and most likely to encounter: the person who does exactly what he is told.
The architecture of the problem is as follows. The machine has been trained on recipes written by human beings for human beings. Those recipes assume a reader who will, upon encountering the phrase "a pinch of salt," deploy the apparatus of a lifetime spent eating—who will taste, adjust, and arrive at a quantity that the phrase was never meant to specify. The phrase is not an instruction. It is a social contract between writer and reader, a compact of shared imprecision that functions because both parties understand that cooking is a sensory negotiation conducted in real time over heat.
The son does not negotiate. He cannot, or does not, improvise. When the machine writes "until thick," he requires a definition of thickness. When it writes "a pinch," he asks how many grams constitute a pinch. These are not unreasonable questions. They are, in fact, the questions that any rigorous reader would pose to any text that purports to be instructional while trafficking in metaphor. That we do not normally ask them is a measure not of their absurdity but of our own willingness to receive approximate language and supply the precision ourselves, from experience.
The machine, confronted with these questions, does what it invariably does: it answers. It provides a gram weight for a pinch, defines "until thick" in terms of minutes or viscosity, each quantification synthesized from the statistical residue of ten thousand cooking forums. It answers with the confidence of a system that cannot distinguish between a question it can resolve and a question that is, by its nature, unresolvable in text—because the answer lives in the hand, the eye, the tongue, and the altitude and humidity of a kitchen it will never occupy.
The son, whose relationship to language is one of fidelity rather than interpretation, accepts these fabricated specificities as though they were measurements. He measures. He times. He follows. The food arrives at table "a bit off," which is to say: the machine's approximation of precision has produced a result that is worse than the original imprecision would have been, had the son been standing beside a human cook who could say "like this" and show him.
What the parent has described, without any intention of describing it, is a stress test. Her son's method of interrogation—dozens of follow-up questions demanding that qualitative language be rendered quantitatively—is structurally identical to the adversarial probes that researchers use to locate failure modes in large language models. The difference is that researchers publish papers. Her son eats the output.
The question she poses—"can I trust ChatGPT not to hallucinate on these questions"—answers itself in the grammar of its own asking. She does not ask whether the machine is reliable. She asks whether she can trust it *not to do the thing it does*. The verb "hallucinate," now so domesticated that a parent in a cooking forum deploys it without quotation marks, has passed from jargon into the common lexicon of managed disappointment: everyone knows the machine confabulates, and everyone asks, each time, whether this time it will not.
There is no villain in this account. The machine is performing as designed. The son is performing as he must. The parent is doing what parents do, which is to weigh autonomy against safety and arrive at no clean answer. The system's failure is not that it lies but that it speaks with uniform authority regardless of whether it is relaying a fact, a convention, a guess, or a fabrication stitched from the statistical average of disagreeing sources. To the reader who takes language at its word, there is no seam between the reliable instruction and the invented one. The tone is the same. The confidence is the same. The gram weight is provided to the decimal.
This desk does not customarily find itself covering matters of domestic cookery. But the question of what happens when authoritative language meets a reader who honors it completely is not a question about cooking. It is a question about the relationship between institutions and the people who believe them, and it is as old as the broadsheet you are holding. The machine has inherited the voice of authority. It has not inherited the obligation that voice has always carried: to know what it does not know, and to say so.
The parent's post has received, at the time of this writing, several hundred responses, most of them sympathetic, several of them technical, and none of them adequate to the structural problem she has identified. Her son continues to cook.