Duke study shows AI predicts ADHD risk in young children

Researchers at Duke University School of Medicine report that an AI model can estimate a child's future risk of attention-deficit/hyperactivity disorder (ADHD) using routine electronic health records. According to Duke Health, the team analyzed records from more than 140,000 children, spanning birth through early childhood, and the model achieved strong predictive performance for children age 5 and older. Axios and Duke quotes emphasize the tool is intended to flag risk, not to make a diagnosis: "This is not an AI doctor," said Matthew Engelhard, a co-author, per Duke and Axios. AzerNEWS reports the system was initially trained on EHRs from over 720,000 patients and cites an evaluation score of approximately 0.92, but researchers say further validation is required.
What happened
Researchers at Duke University School of Medicine published results showing an AI-based model can estimate a child's risk of later receiving an ADHD diagnosis from routine electronic health records (EHRs). Per Duke Health, the study analyzed EHRs from more than 140,000 children covering birth through early childhood and evaluated model performance for children aged 5 and older. Duke quotes a lead author, Elliot Hill: "We have this incredibly rich source of information sitting in electronic health records." Duke and Axios report the tool is intended to flag children for follow-up rather than to provide a clinical diagnosis. Axios and AzerNEWS report that the underlying system had been trained using EHRs from over 720,000 patients; AzerNEWS additionally reports a standardized evaluation score of approximately 0.92 for predicting a diagnosis within four years when applied at age five.
Technical details
Editorial analysis - technical context: EHR-based risk models typically rely on longitudinal feature extraction, sparse event sequences, and handling of class imbalance and missingness. Researchers in comparable studies often use representation-learning or sequence models to capture developmental patterns, and performance is commonly reported with discrimination metrics such as area under the receiver operating characteristic curve (AUC) or similar standardized scores. For practitioners, this class of approach requires careful attention to data curation, temporal validity, and subgroup performance before clinical deployment.
Context and significance
Industry context
Early identification of developmental and mental-health risk from routinely collected health data is an active research area because earlier support can materially affect educational and behavioral outcomes. The Duke team frames the tool as a prioritization aid for primary care, not a replacement for clinical assessment; Axios quotes co-author Matthew Engelhard: "This is not an AI doctor." Public reporting highlights that the model showed consistent performance across sex, race, ethnicity, and insurance status in the study cohort, per Duke Health and Axios.
What to watch
For practitioners: key open questions include external validation in independent health systems, prospective evaluation for clinical utility, transparency of model inputs and thresholds, and evaluation of fairness across underrepresented subpopulations. Observers should also track integration paths into pediatric workflows, regulatory review if the tool is packaged as a clinical decision support product, and data-governance safeguards around EHR-derived risk labeling.
Bottom line
The Duke work adds to evidence that longitudinal EHR signals can flag developmental risk earlier than typical diagnostic timelines. The study reports strong retrospective performance on a large cohort, but further multi-site validation and careful operational testing are necessary before broad clinical use.
Scoring Rationale
The study demonstrates a scalable, EHR-based predictive model with strong retrospective performance on a large pediatric cohort, making it notable for ML practitioners working in healthcare. It is not a frontier-model release, but it has practical implications for clinical ML, validation practices, and deployment challenges.
Practice with real Ad Tech data
90 SQL & Python problems · 15 industry datasets
250 free problems · No credit card
See all Ad Tech problems

