Meeting Banner
Abstract #3562

Unlocking Data-Consistent Synthesis of Clinical Contrasts from Magnetic Resonance Fingerprinting with Semi-Supervised Learning

Mahmut Yurt1, Zihan Zhou2, Congyu Liao3, Cagan Alkan1, Xiaozhi Cao3, Nan Wang3, Julio Oscanoa4, Sophie Schauman5, Mengze Gao6, Zhitao Li7, Tolga Cukur8, Bruno Soares3, Ali Syed3, Shreyas Vasanawala3, John Pauly1, and Kawin Setsompop3
1Department of Electrical Engineering, Stanford University, Stanford, CA, United States, 2Stanford University, Stanford, CA, United States, 3Department of Radiology, Stanford University, Stanford, CA, United States, 4Department of Bioengineering, Stanford University, Stanford, CA, United States, 5Karolinska Institutet, Solna, Sweden, 6Department of Biomedical Physics, Stanford University, Stanford, CA, United States, 7Department of Radiology, Northwestern University, Chicago, IL, United States, 8Department of Electrical and Electronics Engineering, Bilkent University, Ankara, Turkey

Synopsis

Keywords: MR Fingerprinting, MR Fingerprinting

Motivation: We aim to introduce an end-to-end, ultra-fast acquisition and synthesis protocol for contrast-weighted image generation from magnetic resonance fingerprinting.

Goal(s): Our objective is to facilitate data compilation from large patient cohorts to enhance generalizability and accuracy of synthesis.

Approach: We leverage a semi-supervised framework to enable model training using highly-accelerated ground-truth data of the target contrasts, and introduce data-consistent synthesis in inference by performing subject-specific fine-tuning and validation.

Results: Our experiments indicate that the proposed method enables high-quality synthesis using network models trained on prospectively undersampled data of the contrast-weighted images. We show that data-consistent synthesis helps improve synthesis quality and mitigate hallucinations.

Impact: Conventional synthesis models for MRF perform a single-shot inference and are prone to hallucinations and inaccuracy. We introduce semi-supervised learning that enables data-consistency in inference, by merely getting an additional, ultra-fast 1-2min data of the target contrasts for subject-specific fine-tuning.

How to access this content:

For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.

After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.

After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.

Click here for more information on becoming a member.

Keywords