Meeting Banner
Abstract #3563

TransUnet Based Deep Learning with Tissue-weighted Loss Function for Accelerated Magnetic Resonance Fingerprint Reconstruction

Jintao Wei1,2, Zihan Zhou3,4, Bo Dong1,2, Paween Wongkornchaovalit1,2, HuiHui Ye5, and Hongjian He1,6,7
1Center for Brain Imaging Science and Technology, Zhejiang University, Hangzhou, China, 2College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou, China, 3Department of Radiology, Stanford university, Stanford, CA, United States, 4Graduate School of Education, Stanford university, Stanford, CA, United States, 5State Key Laboratory of Modern Optical Instrumentation, College of Optical Science and Engineering, Zhejiang University, Hangzhou, China, 6School of Physics, Zhejiang University, Hangzhou, China, 7State Key Laboratory of Brain-Machine Intelligence, Zhejiang University, Hangzhou, China

Synopsis

Keywords: MR Fingerprinting, Image Reconstruction, transformer, MRF, reconstruction, reproducibility

Motivation: MRF data with short time frames make dictionary matching challenging with large quantitative errors.

Goal(s): We proposed a deep learning method by taking account in global Information and tissue-specific relaxometry, for better quantitative results using shorter frames than dictionary matching with good reproducibility.

Approach: We proposed to combine Unet with an embedded transformer and a tissue-weighted loss function. We compared our approach to dictionary matching and examined various levels of reduced scan time.

Results: The proposed method has less error with shorter frame data comparing dictionary matching. This is more effective for the gray matter and for scan times shorter than 300 frames.

Impact: The proposed method improves the problem of poor quality of reconstructed images for dictionary matching in shorter frames, in addition this method has good repeatability and provides a powerful reconstruction algorithm for MRF sequence acceleration.

How to access this content:

For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.

After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.

After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.

Click here for more information on becoming a member.

Keywords