Robust latent representation of white matter streamlines are critical for parcellating streamlines. This work introduced a novel transformer-based siamese network with triplet margin loss, that learns to embed any lengths of streamlines into fixed-length latent representations. Results showed that a minimum of two layers of transformer encoders were sufficient to model streamlines with a very limited number of training data.
This abstract and the presentation materials are available to members only; a login is required.