Meeting Banner
Abstract #1324

Deep Learning-based Disambiguation for Multiple AD Radiotracers using PET/MRI

Ashwin Kumar1, Donghoon Kim1, Elizabeth Mormino2, Akshay Chaudhari1, Christina Young2, Kevin Chen3, Mehdi Khalighi1, and Greg Zaharchuk1
1Radiology, Stanford University, Stanford, CA, United States, 2Neurology, Stanford University, Stanford, CA, United States, 3Biomedical Engineering, National Taiwan University, Taipei, Taiwan

Synopsis

Keywords: PET/MR, PET/MR

Motivation: AD patients must undergo repeated visits for amyloid and tau radiotracer imaging, leading to high costs and dose concerns due to PET's inability to simultaneously acquire multiple radiotracers during a single session.

Goal(s): Using PET/MRI scans, we used deep learning to create separate amyloid and tau PET images from a simulated combined dual-tracer image.

Approach: We simulated a combined amyloid-tau image by blending co-registered list-mode data and employed a 2.5D U-Net architecture for effective separation.

Results: Mixed-dose models, incorporating physics-inspired data augmentation and MR information, exhibited enhanced anatomical preservation and reduced variability in quantitative metrics.

Impact: The demonstrated separation of a simulated combined amyloid and tau PET/MRI study into its individual components using DL may allow for simultaneous injection of multiple radiotracers in a single acquisition, streamlining the imaging process for AD patients.

How to access this content:

For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.

After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.

After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.

Click here for more information on becoming a member.

Keywords