We trained an established deep-learning-model architecture (3D-Deep-Convolutional-Neural-Network, DeepMedic) on manual segmentations from 70 meningiomas independently segmented by two radiologists. The trained deep-learning model was then validated in a group of 55 meningiomas. Ground truth segmentations were established by two further radiologists in a consensus reading. In the validation-group the comparison of the automated deep-learning-model and manual segmentations revealed average dice-coefficients of 0.91±0.08 for contrast-enhancing-tumor volume and 0.82±0.12 for total-lesion-volume. In the training-group, interreader-variabilities of the two manual readers were 0.92±0.07 for contrast-enhancing-tumor and 0.88±0.05 for total-lesion-volume. Deep-learning based automated segmentation yielded high segmentation accuracy, comparable to manual interreader-variability.
How to access this content:
For one year after publication, abstracts and videos are only open to registrants of this annual meeting. Registrants should use their existing login information. Non-registrant access can be purchased via the ISMRM E-Library.
After one year, current ISMRM & ISMRT members get free access to both the abstracts and videos. Non-members and non-registrants must purchase access via the ISMRM E-Library.
After two years, the meeting proceedings (abstracts) are opened to the public and require no login information. Videos remain behind password for access by members, registrants and E-Library customers.
Keywords