PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
Transformer models have recently started gaining popularity in Computer Vision related tasks. Within Medical Image Segmentation, segmentation models such as TransUNet have incorporated transformer blocks alongside convolutional blocks while remaining faithful to the popular U-Net architecture. The present work utilizes attention maps to examines information flow within transformer blocks of three such segmentation models: (i) TransUNet, (ii) 2D CATS, and (iii) 2D UNETR. Based on the attention maps, compressed versions of these models are proposed which retain only as many transformer layers as are necessary for the model to achieve a global receptive field. The parameter saving is more than 60% whereas the dice metric does not drop by more than 5% compared to the original (uncompressed) model.
Syed Nouman Hasany,Caroline Petitjean, andFabrice Meriaudeau
"A study of attention information from transformer layers in hybrid medical image segmentation networks", Proc. SPIE 12464, Medical Imaging 2023: Image Processing, 124641O (3 April 2023); https://doi.org/10.1117/12.2652215
ACCESS THE FULL ARTICLE
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
The alert did not successfully save. Please try again later.
Syed Nouman Hasany, Caroline Petitjean, Fabrice Meriaudeau, "A study of attention information from transformer layers in hybrid medical image segmentation networks," Proc. SPIE 12464, Medical Imaging 2023: Image Processing, 124641O (3 April 2023); https://doi.org/10.1117/12.2652215