EfficientUNetViT: Efficient Breast Tumor Segmentation Utilizing UNet Architecture and Pretrained Vision Transformer

Autor: Shokofeh Anari, Gabriel Gomes de Oliveira, Ramin Ranjbarzadeh, Angela Maria Alves, Gabriel Caumo Vaz, Malika Bendechache
Jazyk: angličtina
Rok vydání: 2024
Předmět:
Zdroj: Bioengineering, Vol 11, Iss 9, p 945 (2024)
Druh dokumentu: article
ISSN: 2306-5354
DOI: 10.3390/bioengineering11090945
Popis: This study introduces a sophisticated neural network structure for segmenting breast tumors. It achieves this by combining a pretrained Vision Transformer (ViT) model with a UNet framework. The UNet architecture, commonly employed for biomedical image segmentation, is further enhanced with depthwise separable convolutional blocks to decrease computational complexity and parameter count, resulting in better efficiency and less overfitting. The ViT, renowned for its robust feature extraction capabilities utilizing self-attention processes, efficiently captures the overall context within images, surpassing the performance of conventional convolutional networks. By using a pretrained ViT as the encoder in our UNet model, we take advantage of its extensive feature representations acquired from extensive datasets, resulting in a major enhancement in the model’s ability to generalize and train efficiently. The suggested model has exceptional performance in segmenting breast cancers from medical images, highlighting the advantages of integrating transformer-based encoders with efficient UNet topologies. This hybrid methodology emphasizes the capabilities of transformers in the field of medical image processing and establishes a new standard for accuracy and efficiency in activities related to tumor segmentation.
Databáze: Directory of Open Access Journals
Nepřihlášeným uživatelům se plný text nezobrazuje