Meeting Banner
Abstract #2301

CEST analysis using a deep residual learning method based on Bloch-McConnell synthetic training data

Shihao Zeng1, Huabin Zhang1,2, Ziyan Wang1, Jiawen Wang1, Pei Cai1, and Jianpan Huang1,3
1Department of Diagnostic Radiology, Li Ka Shing Faculty of Medicine, the University of Hong Kong, Hong Kong, China, 2Medical Imaging Center, Department of Electronic Engineering and Information Science, University of Science and Technology of China, Hefei, China, 3Tam Wing Fan Neuroimaging Research Laboratory, The University of Hong Kong, Hong Kong, Hong Kong, China

Synopsis

Keywords: CEST / APT / NOE, CEST / APT / NOE

Motivation: Conventional Lorentzian-based fitting methods for CEST MRI are time-consuming and susceptible to initial and bound values.

Goal(s): We aim to propose a deep residual learning method based on Bloch-McConnell synthetic training data to quantify CEST contrasts

Approach: We utilized deep residual networks to predict the Z-spectrum reference (Zref) signal (ZrefNet) for subsequent CEST quantification (Zref – Z). ZrefNet is trained purely on Bloch-McConnell synthetic data, and tested on simulation and in vivo human brain data.

Results: ZrefNet generated accurate CEST contrasts rapidly with high correlation to ground-truth on simulation data and demonstrated feasibility for analyzing human multiple sclerosis / healthy control CEST data.

Impact: We proposed deep residual networks to predict the Z-spectrum reference signal (ZrefNet) for CEST quantification. ZrefNet generated accurate CEST contrasts on simulation data and demonstrated its feasibility for analyzing in vivo human CEST data of multiple sclerosis and healthy controls.

How to access this content:

For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.

After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.

After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.

Click here for more information on becoming a member.

Keywords