A multimodal and temporal foundation model for virtual patient representations at healthcare system scale

Researchers built Apollo, a foundation model trained on 25 billion clinical records spanning 7.2 million patients and 30 years of longitudinal data from a major US hospital system. The model unifies 28 medical modalities including images, text, and 100k+ clinical events into a single patient representation, enabling AI systems to reason across entire care journeys.
Modelwire context
ExplainerThe buried detail here is the 30-year longitudinal depth: Apollo isn't just wide across modalities, it's trained to reason across time within a single patient's history, which is the harder problem most clinical AI systems quietly sidestep by treating each encounter as isolated.
The closest thread in recent coverage is the MADE benchmark paper from arXiv (April 16), which flagged how hard it is to evaluate ML models on high-stakes healthcare tasks when labels are imbalanced and data contamination is a real risk. Apollo faces the same evaluation problem at a much larger scale: 7.2 million patients and 28 modalities means the benchmark surface is enormous, and the paper's validation methodology will matter as much as the architecture itself. The MIT Technology Review piece on enterprise AI as an operating layer is also relevant context here, because Apollo's value proposition is almost entirely about institutional infrastructure, specifically what a hospital system that already owns 30 years of records can do that no external vendor can replicate.
Watch whether Apollo's authors release a public evaluation suite or external validation cohort within the next six months. Without third-party replication on held-out hospital systems, the performance claims remain internal benchmarks on the same data distribution the model trained on.
This analysis is generated by Modelwire’s editorial layer from our archive and the summary above. It is not a substitute for the original reporting. How we write it.
MentionsApollo · arXiv
Modelwire summarizes — we don’t republish. The full article lives on arxiv.org. If you’re a publisher and want a different summarization policy for your work, see our takedown page.