Meeting Banner
Abstract #2824

Integrating Deep Learning for Detection and Correction of Motion Artifacts in multi-echo GRE MRI

Eun-Gyu Ha1, Kyu-Jin Jung1, Mohammed A. Al-masni2, and Dong-Hyun Kim1
1Department of Electrical and Electronic Engineering, Yonsei University, Seoul, Korea, Republic of, 2Department of Artificial Intelligence, Sejong University, Seoul, Korea, Republic of

Synopsis

Keywords: Artifacts, Motion Correction

Motivation: Motion artifacts in mGRE MRI scans reduce image quality, increasing the risk of misdiagnosis and often necessitating repeat scans, negatively impacting patient care and diagnostic accuracy.

Goal(s): Our goal is to develop a novel deep learning-based framework for reducing motion artifacts in mGRE MRI k-space data, ensuring the generation of high-quality images.

Approach: The methodology proceeds by detecting and correcting motion-corrupted phase encoding lines within the k-space domain, employing a two-stage DeepFillv2 algorithm. It also integrates motion parameter estimation to enhance the framework's robustness.

Results: The model’s effectiveness in identifying and rectifying motion artifacts in MRI was confirmed through quantitative and qualitative evaluation.

Impact: The proposed k-space domain framework progresses by identifying phase encoding lines affected by motion and repairing them using deep learning techniques, thereby proving improved image quality and demonstrating potential as a diagnostic aid.

How to access this content:

For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.

After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.

After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.

Click here for more information on becoming a member.

Keywords