Meeting Banner
Abstract #4306

Validation of Deep Learning based Tissue Segmentation for Efficient and Robust Quantitation of Background Parenchymal Enhancement on Breast MRI

Yu-Tzu Kuo1, Anum S. Kazerouni2, Vivian Y. Park3, Wesley Surento2,4, Suleeporn Sujichantararat2, Daniel S. Hippe5, Habib Rahbar2, and Savannah C. Partridge1,2
1Department of Bioengineering, University of Washington, Seattle, WA, United States, 2Department of Radiology, University of Washington, Seattle, WA, United States, 3Department of Radiology, Severance Hospital, Research Institute of Radiological Science, Yonsei University College of Medicine, Seoul, Korea, Republic of, 4Department of Biomedical Informatics and Medical Education, University of Washington, Seattle, WA, United States, 5Clinical Research Division, Fred Hutchinson Cancer Center, Seattle, WA, United States

Synopsis

Keywords: Segmentation, Segmentation, Breast, Machine Learning/Artificial Intelligence

Motivation: Quantitative measurements, such as background parenchymal enhancement (BPE) on breast dynamic contrast-enhanced (DCE) MRI, show potential as markers for breast cancer risk.

Goal(s): This study compares the efficiency and reproducibility of automated deep learning (DL) vs. semi-automated methods for fibroglandular tissue (FGT) segmentation and quantitation of BPE on DCE-MRI.

Approach: Quality of FGT segmentations generated by fuzzy c-means (semi-automated) and DL (fully automated) algorithms were scored by a radiologist, and the agreement of quantitative BPE measurements across segmentation approaches was assessed.

Results: Deep learning-based FGT segmentation offers more accurate and robust FGT segmentation, enhancing quality and maintaining BPE reproducibility.

Impact: Application of deep learning for segmentation of fibroglandular tissue on breast MRI can improve the robustness and reliability of quantitative imaging biomarkers, with the potential to improve risk stratification and clinical decision-making for high-risk breast cancer screening.

How to access this content:

For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.

After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.

After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.

Click here for more information on becoming a member.

Keywords