Skip to main content
← AI in Pediatrics Research Watch · AI & Technology

AI in Pediatrics: Research Watch

High-signal academic papers, clinical studies, and research updates tracked by PHD. Annotated for clinical and commercial relevance.

Last updated: April 23, 2026

This tracker surfaces academic and clinical work with direct relevance to companies, investors, and operators in the pediatric health technology space. Papers are selected for commercial or clinical signal value, not for novelty alone.

Recent High-Signal Additions

Pediatric Cardiology · Echocardiography AI

AI Learning for Pediatric Right Ventricular Assessment: Development and Validation Across Multiple Centers (npj Digital Medicine, December 2025)

Multi-center video deep learning on 24,984 echocardiograms from 3,993 children across four tertiary centers; segmentation, RV fractional area change estimation, and RV disease classification. High-signal peds-first imaging AI in a difficult-to-measure cardiac chamber, with multi-center validation that can reduce operator variability and make pediatric echo analytics more scalable across children's hospitals.

Signal strength: High. Limits: Retrospective evaluation; performance depends on labeling quality and echo acquisition variability; prospective deployment evidence still needed.

View paper →
Hospital Operations · Implementation AI

Artificial Intelligence Length-of-Stay Forecasting and Pediatric Surgical Capacity (JAMA Pediatrics, January 2026)

Preimplementation and postimplementation cohort analysis with model deployment at Boston Children's. An XGBoost length-of-stay model was used in surgical scheduling and capacity management, producing measurable reductions in bed underuse and day-to-day capacity volatility. One of the clearest "AI changed hospital behavior" pediatric papers this year.

Signal strength: High. Limits: Single-center deployment; operations gain does not automatically equal patient-level outcome gain; portability depends on local workflow maturity.

View paper →
ED Decision Support · Pediatric Sepsis

Development and Multicenter Validation of an Explainable Machine Learning Diagnostic Criteria for Pediatric Abdominal Sepsis (npj Digital Medicine, March 2026)

Retrospective development plus multicenter external validation and a small prospective validation cohort. Proposes an ED-ready diagnostic criteria (ABSeD) for pediatric abdominal sepsis. Useful as a marker for where pediatric sepsis AI is moving beyond methods demos toward deployable diagnostic support.

Signal strength: High. Limits: Focused on pediatric abdominal sepsis; needs workflow and prospective impact evaluation beyond model performance.

View paper →
NICU Risk Prediction · Longitudinal Surveillance

Predicting Pulmonary Hypertension in Infants with Bronchopulmonary Dysplasia (Journal of Perinatology, February 2026)

Multicenter EHR-derived cohort with temporal validation; logistic regression and sequence modeling. A deployable-shaped neonatal ML paper: clear clinical event, defined time windows, and explicit temporal validation rather than only random splits. Relevant to children's hospitals evaluating neonatal risk models for longitudinal surveillance.

Signal strength: High. Limits: Retrospective data; pulmonary hypertension definitions depend on echo use and local practice patterns.

View paper →
Clinical NLP · LLM Evaluation and Governance

Large Language Models Using Clinical Text in Pediatrics: A Scoping Review (JAMA Network Open, March 2026)

Scoping review of pediatric clinical-text LLM studies. Makes clear that reporting, evaluation standardization, and privacy safeguards remain inconsistent across the fast-growing pediatric LLM literature — exactly the filter readers need before treating LLM pilots as infrastructure.

Signal strength: High. Limits: Review-level evidence; maps gaps and standards rather than proving clinical utility.

View paper →
Regulatory · Pediatric Device Evidence Standards

FDA-Regulated AI-Enabled Medical Devices With Pediatric Indications (JAMA Network Open, March 2026)

Cross-sectional analysis of FDA regulatory submissions. Quantifies a core governance reality: many FDA-cleared AI devices do not clearly communicate pediatric scope. Matters for children's-hospital buyers, governance teams, and safety committees interpreting "FDA cleared" claims.

Signal strength: High. Limits: Labeling analysis does not evaluate device performance in children.

View paper →
Multicenter Validation · Pediatric Cardiology

Multicenter Validation of AI-Enabled ECG for Pediatric Biological Sex Prediction (npj Digital Medicine, February / April 2026)

External multicenter validation of a previously published AI-ECG model across Texas Children's and Boston Children's, with age/puberty-linked performance gradients and interpretable signal mapping. The value is in the validation pattern — pediatric external validation with transparent performance reporting across sites and age bands.

Signal strength: Medium. Limits: Limited direct clinical utility; stronger as a representation and validation-methods paper than as a bedside workflow paper.

View paper →
Adolescent · Child GenAI Exposure

Generative Artificial Intelligence Applications Use Among US Youth (JAMA Network Open, February 2026)

Cross-sectional passive-sensing study using a parental monitoring app sample of US youth ages 4–17. Provides an empirical baseline for youth GenAI exposure — separating "kids might use chatbots someday" from "GenAI tools are already part of the pediatric/adolescent digital environment."

Signal strength: Medium-high. Limits: Sample is families using a commercial monitoring app; usage does not measure mental-health impact or clinical harm/benefit.

View paper →

Foundational Papers We Still Watch

Paper Journal Why It Matters
Artificial and human intelligence for early identification of neonatal sepsis Pediatric Research Strong framing paper for how neonatal AI should support, not replace, clinical judgment
AI-guided precision parenteral nutrition for NICUs Nature Medicine One of the stronger examples of AI entering a real neonatal workflow
Eye-Tracking-Based Measurement of Social Visual Engagement Compared With Expert Clinical Diagnosis of Autism JAMA Validation anchor for autism diagnostic AI
Development and Replication of Objective Measurements of Social Visual Engagement to Aid in Early Diagnosis and Assessment of Autism JAMA Network Open Replication-oriented support for developmental diagnostics
Toward governance of artificial intelligence in pediatric healthcare npj Digital Medicine Useful framing paper for the governance layer of pediatric AI
AI Chatbots and Youth Mental Health: Practical Recommendations for Clinicians JAACAP Good example of youth chatbot evidence being handled cautiously

Research Gaps We're Tracking

  • Prospective impact studies for pediatric AI tools after model validation
  • Pediatric performance disclosure for FDA-cleared AI-enabled devices
  • LLM safety, privacy, and evaluation standards for pediatric clinical text
  • Parent-proxy portal messaging and GenAI drafted replies in children's hospitals
  • Age-band, fairness, and generalizability testing across pediatric populations
  • Youth-facing chatbot and companion-tool safety evidence beyond usage estimates
  • Neonatal and pediatric ICU AI models that improve workflow decisions, not just prediction metrics

See a paper we should be tracking?

We review submissions and add high-signal research to this tracker.

Submit a paper