Keywords: Image Reconstruction, Image Reconstruction
Motivation: Deep Learning MRI reconstruction requires access to high-performance compute hardware (GPUs) but reconstruction times can still remain slow.
Goal(s): Use domain specific hardware accelerators to speed up deep learning MRI reconstruction compared to classic CPU/GPU implementations.
Approach: We use an open source SoC design and simulation framework, Chipyard, and a neural network accelerator, Gemmini, to run inference on a pretrained MRI reconstruction network. The results are measured on FireSim, an FPGA simulation framework.
Results: Simulated inference times of pretrained models on Gemmini are much faster compared to CPU and show similar image quality metrics compared to the inference run on a GPU.
Impact: Shows the potential of developing custom compute hardware designed to accelerate deep learning MRI reconstruction.
How to access this content:
For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.
After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.
After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.
Keywords