Physician Argues AI Cannot Replace Doctors' Judgment

In a New York Times op-ed, Dr. Danielle Ofri argues that while AI can synthesize clinical data and draft administrative materials, it fails to replicate the relational, context-rich pattern recognition of an experienced clinician, according to reporting by Newser. Ofri recounts diagnosing a longtime patient by sensing changes in his breathing and facial expression-cues she says were not captured by vitals or population statistics and that AI could not have detected, per the op-ed. She writes that AI misses the multidimensional person and that future clinicians should receive training in the medical humanities alongside technical skills. Ofri calls the gap between algorithmic analysis and human assessment "an ocean of distance between the 'patient' that AI is analyzing and the patient that the human doctor or nurse is assessing," as quoted in the piece. Editorial analysis: The essay highlights enduring limits of clinical AI around context, empathy, and longitudinal knowledge.
What happened
In a New York Times op-ed, Dr. Danielle Ofri argues that AI can be useful for synthesizing patient data and handling administrative work but cannot substitute for the clinician's relational knowledge, according to reporting by Newser. The op-ed recounts a case in which Ofri instantly sensed that something was off with a longtime patient from his breathing and facial expression, and she writes that those cues were not explainable by vital signs or statistical population indicators. Ofri also writes that AI would not have known a concurrent family crisis had changed the patient's eating habits and likely affected his kidney function.
Technical details
Editorial analysis - technical context: The piece frames AI's strengths as data aggregation and text generation for tasks like insurance appeals, and contrasts those strengths with clinical pattern recognition that arises from sustained clinician-patient relationships. This is a commonly noted limitation in clinical-AI literature: models excel at correlational signal extraction from structured and unstructured records but struggle with multimodal, embodied, and contextual signals that are not well represented in EHR data.
Context and significance
Industry context
For practitioners, Ofri's argument underscores the human factors often cited in deployment studies: continuity of care, nonverbal cues, and patient narratives are hard to encode and validate in model training datasets. The op-ed's declaration of "an ocean of distance between the 'patient' that AI is analyzing and the patient that the human doctor or nurse is assessing" highlights a boundary where automation may change workflows without replacing clinical judgement.
What to watch
Observers should track research and product work that seeks to capture longitudinal, multimodal patient context-for example, systems that integrate caregiver notes, wearable respiration signals, or structured social-determinants data-and whether validation protocols evolve to test for sensitivity to those contextual cues. Also watch for curriculum changes linking clinical training and the medical humanities, a point Ofri raises in arguing for broader clinician preparation.
Editorial analysis: The op-ed is a practitioner-facing reminder that clinical AI evaluation must move beyond aggregate metrics to include longitudinal, interpersonal, and equity-sensitive measures of performance.
Scoring Rationale
The piece is a thoughtful practitioner critique highlighting human-AI limits in clinical settings. It is relevant to clinicians and ML practitioners focused on deployment and evaluation, but it does not report new research or a major product change.
Practice with real Health & Insurance data
90 SQL & Python problems · 15 industry datasets
250 free problems · No credit card
See all Health & Insurance problems


