in

Retina Blood Vessel Segmentation Using A U-Net Based Convolutional Neural Network


This is an implementation of the research paper “Retina Blood Vessel Segmentation Using A U-Net Based Convolutional Neural Network” written by Wang Xiancheng, Li Weia, et al.

Check out the standalone demo notebook and run segRetino inferences here.

Inspiration

Various eye diseases can be diagnosed through the characterization of the retinal blood vessels. The characterization can be extracted by using proper imaging techniques and data analysis methods. In case of eye examination, one of the important tasks is the retinal image segmentation.The paper presents a network and training strategy that relies on the data augmentation to use the available annotated samples more efficiently, to segment retinal blood vessels using a UNET convolutional neural network.

Dataset

We have used the Digital Retinal Images for Vessel Extraction (DRIVE) dataset for retinal vessel segmentation. It consists of a total of JPEG 40 color fundus images; including 7 abnormal pathology cases. Each image resolution is 584×565 pixels with eight bits per color channel (3 channels), resized to 512×512 for our model.

Guidelines to download, setup and use the dataset

The DRIVE dataset may be downloaded here as two files named training.zip and test.zip.

Please write the following commands on your terminal to extract the file in the proper directory.

$ mkdir drive $ unzip </path/to/training.zip> -d </path/to/drive>  $ unzip </path/to/test.zip> -d </path/to/drive>

The resulting directory structure should be:

/path/to/drive
    -> train
        -> image
            -> 21_training_0.tif
            -> 22_training_0.tif
               ...
        -> mask
            -> 21_training_0.gif
            -> 22_training_0.gif
    -> test
        -> image
            -> 01_test_0.tif
            -> 02_test_0.tif
               ...
        -> mask
            -> 01_test_0.gif
            -> 02_test_0.gif

Model Components

The UNET CNN architecture may be divided into the Encoder, Bottleneck and Decoder blocks, followed by a final segmentation output layer.

  • Encoder: There are 4 Encoder blocks, each consisting of a convolutional block followed by a Spatial Max Pooling layer.
  • Bottleneck: The Bottleneck consists of a single convolutional block.
  • Decoder: There are 4 Decoder blocks, each consisting of a deconvolution operation, followed by a convolutional block, along with skip connections.

Note: The convolutional block consists of 2 conv2d operations each followed by a BatchNorm2d, finally followed by a ReLU activation.

model_arch

Implementation Details

  • Image preprocessing included augmentations like HorizontalFlip, VerticalFlip, Rotate.
  • Dataloader object was created for both training and validation data
  • Training process was carried out for 50 epochs, using the Adam Optimizer with a Learning Rate 1e-4.
  • Validation was carried out using Dice Loss and Intersection over Union Loss.

Installation and Quick Start

To use the repo and run inferences, please follow the guidelines below

Cloning the Repository:

  $ git clone https://github.com/srijarkoroy/segRetino

Entering the directory:

  $ cd segRetino/

Setting up the Python Environment with dependencies:

  $ pip install -r requirements.txt

Running the file for inference:

  $ python3 test.py

Running the test file downloads the pretrained weights of the UNET Model that we have trained on the DRIVE Dataset. However if you want to re-train the model please mention the path to your dataset on you local machine after augmentations, inside the train.py file, as:

train_x = sorted(glob(<path/to/augmented/train/image/folder/>)) train_y = sorted(glob(<path/to/augmented/mask/image/folder/>)) valid_x = sorted(glob(<path/to/test/image/folder/>)) valid_y = sorted(glob(<path/to/test/mask/folder/>))

Once the path has been mentioned, the model may be trained by running the command:

$ python3 train.py

Note: If images have not been augmented, please see the instructions for augmentation here.

The test file saves two images in the mentioned paths, a masked image showing only the blood vessels, and a blend image showing the blood vessels within the retina. If you don’t want to save the blend image, consider running the following code snippet:

# Creating the SegRetino object initialized with the test image pathseg = SegRetino(‘<path/to/test/img>’) # Running inferenceseg.inference(set_weight_dir=”unet.pth”, path=”<path/to/save/masked/image>”, blend=False, blend_path = None)

Check out the standalone demo notebook and run segRetino inferences here.

Note: Is is advisable to use a GPU for running the inferences since performing segmentation on 512×512 images with a heavy UNET architecture is expensive.

Results from Implementation

GitHub – srijarkoroy/segRetino: An implementation of the research paper “Retina Blood Vessel Segmentation Using A U-Net Based Convolutional Neural Network”

An implementation of the research paper &quot;Retina Blood Vessel Segmentation Using A U-Net Based Convolutional Neural Network&quot; – GitHub – srijarkoroy/segRetino: An implementation of the res…


Leave a Reply

Your email address will not be published. Required fields are marked *

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

GIPHY App Key not set. Please check settings

Communyco — Where creators connect

[Ale’s Unpopular Opinions – Issue #1] Big Tech battles over ICE contracts & more