Skip to content

Brainy is a virtual MRI analyzer. Just upload the MRI scan file and get 3 different classes of tumors detected and segmented. In Beta.

Notifications You must be signed in to change notification settings

IAmSuyogJadhav/Brainy

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

61 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Brainy

The virtual MRI Analyzer


Preview


A screenshot of the project interface

Installation Instructions

  1. Clone this repository

    git clone https://github.com/IAmSuyogJadhav/Brainy.git

    or Download and then extract its contents.

  2. Download the model from here and put the file inside app/static/models/ folder. Do not change the name (It should be Model_1.h5).

  3. From the root folder of te repository, run:

pip3 install -r requirements.txt

to install ll the dependencies. Note that this project is based on python3.

  1. Start the app using following command, when in the main folder:

    flask run

    It may take a while.

  2. Now open your browser and navigate to http://localhost:5000 and follow the instructions. It is recommended to have a Nvidia GPU, since it can speed up the prediction task manifolds.

Sample MRI scans

There are 2 sample cases from the BraTS2018 dataset insides the samples/ directory. You can use them to quickly test the app. These cases also contain a file (named in the format xxx_seg.nii.gz) that shows the ground truth segmentation map for that case.

Project Overview

MRI is a very popular technique to detect tumours. MRI creates cross-section pictures of your insides. But MRI uses strong magnets to make the images – not radiation. An MRI scan takes cross-sectional slices (views) from many angles, as if someone were looking at a slice of your body from the front, from the side, or from above your head. MRI creates pictures of soft tissue parts of the body that are sometimes hard to see using other imaging tests.MRI is very good at finding and pinpointing some cancers. An MRI with contrast dye is the best way to see brain and spinal cord tumors.

Why this Project

With a surge in the brain related ailments in the world, every year over 2,500 of the Indian children suffer from medulloblastoma, a pediatric malignant primary brain tumour which spreads through the cerebrospinal fluid (CSF) and frequently metastasize to different locations along the surface of the brain and spinal cord, doctors have said.

According to them, in India, every year 40,000-50,000 persons are diagnosed with brain tumour. Of these 20 per cent are children. Until a year ago, the figure was only somewhere around 5 per cent.

The doctors said that if the cases are detected early, then 90 per cent of the tumour cases are curable, provided the treatment protocol is followed correctly.

-Source (“The Hindu “)
  • Though the technology used in MRI seems to be quite technically advanced, for the detection of tumour in the MRI image, currently we mostly rely on human skill and expertise.

  • Although this method proves to be accurate in most of the cases, it sometimes fails when the human eye fails to detect tiny irregularities

  • In a country like India, there is a lack of skilled Oncologists and Neurologists present.

  • According to a report in 2018 , the country is facing a severe shortage of care-givers with merely 2000 Oncologists to look after around 10 Million patients.

  • The condition is much more worse in the rural areas where the count of Oncologists is much less.

Proposed Solution

Jump directly to the training notebook by clicking here

​ In order to solve this problem, we propose Brainy, our Brain MRI segmentation tool. Not only would Brainy automatise the entire process of diagnosis for tumours , but it would also help in solving out those cases which are generally missed by the human eye.

About the Dataset

​ The data that we’ve used is the BRATS dataset. This dataset contains the segmented images of the brain tissues. After the image has been segmented, there are 3 classes namely Necrotic and Non-Enhancing tumour (NCR/NET), Peritumoral Edema (ED) and GD-Enhancing tumour (ET). There are about 220 images in the training dataset.

The Model


To segment the brain MRI’s we use a popular network called U-Net. The U-Net architecture is built upon the Fully Convolutional Network and modified in a way that it yields better segmentation in medical imaging. Compared to FCN-8, the two main differences are:

1. U-net is symmetric and
2. The skip connections between the downsampling path and the upsampling path apply a concatenation operator instead of a sum.

These skip connections intend to provide local information to the global information while upsampling. Because of its symmetry, the network has a large number of feature maps in the upsampling path, which allows to transfer information. By comparison, the basic FCN architecture only had number of classes feature maps in its upsampling path.


The U-Net Model


The U-Net owes its name to its symmetric shape, which is different from other FCN variants. U-Net architecture is separated in 3 parts:

1. The contracting/downsampling path
2. Bottleneck
3. The expanding/upsampling path

Metrics


To quantify the performance of our image segmentation, Dice Score is used. The algorithm is validated by calculating the Dice score, which is a measure of how similar the objects are. So it is the size of the overlap of the two segmentations divided by the total size of the two objects. That is, the score for a particular class c is the size of the overlap between the predicted region and its true counterpart. We were able to achieve best coefficient of dice loss score 0.61 (higher the better).

Team

Gk Tejus Suyog Jadhav Udbhav Bamba

References


This project was made as part of the TCS PanIIT Conclave 2019 - Mission AI: Solve for India, a 24 hour hackathon organised at IIT Delhi, India on 19th-20th January by the PanIIT organisation, Skillenza and TCS. Our project achieved 4th rank at the same.

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •