Keywords: Diagnosis/Prediction, Segmentation
Motivation: Deep learning models for segmentation require large datasets, limiting their use in clinical settings. Foundation models, which learn from non-medical images before fine-tuning for specific tasks, have emerged as a possible solution.
Goal(s): We aimed to adapt and evaluate a recent foundation model, the Medical Segment Anything model (MedSAM), for neuroanatomy segmentation.
Approach: Using the Human Connectome Project dataset, we trained MedSAM to segment 102 regions-of-interest and compared its accuracy with a baseline UNet model.
Results: UNet outperformed MedSAM in almost all regions and dataset sizes, but MedSAM showed potential when training with very few MRIs.
Impact: While foundation models such as MedSAM have potential for medical segmentation, they currently may not surpass traditional models when using sufficient data. In the data-limited setting, however, they can be useful when extremely little labeled data is available.
How to access this content:
For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.
After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.
After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.
Keywords