MedMobile: A mobile-sized language model with expert-level clinical capabilities
Krithik Vishwanath, Jaden Stryker, Anton Alaykin, Daniel Alexander Alber, Eric Karl Oermann
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/nyuolab/MedMobileOfficialnone★ 6
Abstract
Language models (LMs) have demonstrated expert-level reasoning and recall abilities in medicine. However, computational costs and privacy concerns are mounting barriers to wide-scale implementation. We introduce a parsimonious adaptation of phi-3-mini, MedMobile, a 3.8 billion parameter LM capable of running on a mobile device, for medical applications. We demonstrate that MedMobile scores 75.7% on the MedQA (USMLE), surpassing the passing mark for physicians (~60%), and approaching the scores of models 100 times its size. We subsequently perform a careful set of ablations, and demonstrate that chain of thought, ensembling, and fine-tuning lead to the greatest performance gains, while unexpectedly retrieval augmented generation fails to demonstrate significant improvements
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| MedMCQA Dev | MedMobile (3.8B) | Accuarcy | 63.2 | — | Unverified |
| MedQA | MedMobile (3.8B) | Accuracy | 75.7 | — | Unverified |