Meeting Banner
Abstract #3267

Improving Data Consistency and Generalizability in Deep Learning-Based QSM Reconstruction

Yuanbo Zhang1,2, Yudu Li1,3,4, Ruihao Liu1,5, Ziwen Ke1,5,6, Wen Jin1,2, Yibo Zhao1,2, Yao Li5,6, and Zhi-Pei Liang1,2
1Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, IL, United States, 2Department of Electrical and Computer Engineering, University of Illinois at Urbana-Champaign, Urbana, IL, United States, 3Department of Bioengineering, University of Illinois at Urbana-Champaign, Urbana, IL, United States, 4National Center for Supercomputing Applications, University of Illinois at Urbana-Champaign, Urbana, IL, United States, 5School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, China, 6Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China

Synopsis

Keywords: Susceptibility/QSM, Quantitative Susceptibility mapping

Motivation: QSM reconstruction is a highly ill-posed inverse problem. Machine learning promises to address this problem by incorporating data-driven priors. But existing learning-based QSM reconstructions suffer from data inconsistency and limited generalizability.

Goal(s): To improve data consistency and generalizability of learning-based QSM reconstruction.

Approach: Data consistency was improved by enforcing measured tissue phase in k-space, leveraging the characteristics of the dipole kernel. Generalizability was improved by augmenting a small COSMOS training dataset with extensive phase and spatial variation from a large single-orientation QSM dataset.

Results: The proposed method has been validated on both simulation and in vivo data, producing data-consistent and generalizable QSM maps.

Impact: With improved data consistency and generalizability, the proposed method could significantly enhance the reliability of AI-powered QSM reconstruction, making it a practically useful tool to support a wide range of QSM studies in scientific and clinical applications.

How to access this content:

For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.

After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.

After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.

Click here for more information on becoming a member.

Keywords