Alberto Solano
$^{1}$ Intelligent Data Analysis Laboratory, ETSE (Engineering School), Universitat de València, 46100 Burjassot, Spain
$^{2}$ Image Processing Lab., Universitat de València, 46980 Paterna, Spain
For many years, convolutional neural networks dominated the field of computer vision, not least in the medical field, where problems such as image segmentation were addressed by such networks as the U-Net. The arrival of self-attention based networks to the field of computer vision through ViTs seems to have changed the trend of using standard convolutions. Throughout this work, we apply different architectures such as U-Net, ViTs and ConvMixer, to compare their performance on a medical semantic segmentation problem.
All the models have been trained from scratch on the DRIVE dataset and evaluated on its private counterpart to assess which of the models performed better in the segmentation problem.
Our major contribution is showing that the best performing model (ConvMixer) is the one that shares the approach from the ViT (processing images as patches) while maintaining the foundational blocks (convolutions) from the U-Net.
This mixture doesn't only produce better results (
This repository contains all the code and data needed to reproduce the results shown in the paper.
The Data
folder contains the different splits (training, validation and test) used, while the Code
folder is organized as follows:
Checkpoint_<model>
: Results of each model training.Models/
: PyTorch definition of the models used.Train/
: Training script of each model.Utils/dataset.py
: PyTorch dataset and augmentations used.Utils/utils.py
: Utilities to process the data and manage the training.
-
ConvMixer (Light): Weights | Parameters
-
ConvMixer: Weights | Parameters
Sample results extracted from the paper: