ENGLISH
La vitrine de diffusion des publications et contributions des chercheurs de l'ÉTS
RECHERCHER

Automatic segmentation of echocardiographic images using a shifted windows vision transformer architecture

Nemri, Souha et Duong, Luc. 2024. « Automatic segmentation of echocardiographic images using a shifted windows vision transformer architecture ». Biomedical Physics and Engineering Express, vol. 10, nº 6.

[thumbnail of Duong-L-2024-29683.pdf]
Prévisualisation
PDF
Duong-L-2024-29683.pdf - Version publiée
Licence d'utilisation : Creative Commons CC BY.

Télécharger (1MB) | Prévisualisation

Résumé

Echocardiography is one the most commonly used imaging modalities for the diagnosis of congenital heart disease. Echocardiographic image analysis is crucial to obtaining accurate cardiac anatomy information. Semantic segmentation models can be used to precisely delimit the borders of the left ventricle, and allow an accurate and automatic identification of the region of interest, which can be extremely useful for cardiologists. In the field of computer vision, convolutional neural network (CNN) architectures remain dominant. ExistingCNNapproaches have proved highly efficient for the segmentation of various medical images over the past decade. However, these solutions usually struggle to capture long-range dependencies, especially when it comes to images with objects of different scales and complex structures. In this study, we present an efficient method for semantic segmentation of echocardiographic images that overcomes these challenges by leveraging the selfattention mechanism of the Transformer architecture. The proposed solution extracts long-range dependencies and efficiently processes objects at different scales, improving performance in a variety of tasks.Weintroduce Shifted Windows Transformer models (Swin Transformers), which encode both the content of anatomical structures and the relationship between them. Our solution combines the Swin Transformer and U-Net architectures, producing a U-shaped variant. The validation of the proposed method is performed with the EchoNet-Dynamic dataset used to train our model. The results show an accuracy of 0.97, a Dice coefficient of 0.87, and an Intersection over union (IoU) of 0.78. Swin Transformer models are promising for semantically segmenting echocardiographic images and may help assist cardiologists in automatically analyzing and measuring complex echocardiographic images.

Type de document: Article publié dans une revue, révisé par les pairs
Professeur:
Professeur
Duong, Luc
Affiliation: Génie logiciel et des technologies de l'information
Date de dépôt: 18 oct. 2024 20:26
Dernière modification: 28 oct. 2024 16:27
URI: https://espace2.etsmtl.ca/id/eprint/29683

Actions (Authentification requise)

Dernière vérification avant le dépôt Dernière vérification avant le dépôt