Skip to content

saeed-anwar/PLCNN

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

26 Commits
 
 
 
 

Repository files navigation

Deep Localization of Protein Structures in Fluorescence Microscopy Images

This repository is for Deep localization of protein structures in fluorescence microscopy images (PLCNN) introduced in the following paper

Muhammad Tahir, Saeed Anwar, and Ajmal Mian, "Deep Localization of Protein Structures in Fluorescence Microscopy Images", Neural Computing and Applications (NCAA), 2021

The model is built in PyTorch 0.4.0, PyTorch 0.4.1 and tested on Ubuntu 14.04/16.04 environment (Python3.6, CUDA9.0, cuDNN5.1).

Contents

  1. Introduction
  2. Network
  3. Test
  4. Results
  5. Citation
  6. Acknowledgements

Introduction

Accurate localization of proteins from fluorescence microscopy images is a challenging task due to the inter-class similarities and intra-class disparities introducing grave concerns in addressing multi-class classification problems. Conventional machine learning-based image prediction relies heavily on pre-processing such as normalization and segmentation followed by hand-crafted feature extraction before classification to identify useful and informative as well as application specific features.We propose an end-to-end Protein Localization Convolutional Neural Network (PLCNN) that classifies protein localization images more accurately and reliably. PLCNN directly processes raw imagery without involving any pre-processing steps and produces outputs without any customization or parameter adjustment for a particular dataset. The output of our approach is computed from probabilities produced by the network. Experimental analysis is performed on five publicly available benchmark datasets. PLCNN consistently outperformed the existing state-of-the-art approaches from machine learning and deep architectures.

Image datasets for protein localization; each image belongs to a different class. Most of the images are sparse.

Network

The architecture of the proposed network. A glimpse of the proposed network used for localization of the protein structures in the cell. The composition of R_s, R_l, P_s and P_l are provided below the network structure, where the subscript s have a small number of convolutions as compared to l

Test

Quick start

  1. Download the trained models for our paper and place them in '/TestCode/experiment'.

    The PLCNN model can be downloaded from Google Drive or here. The total size for all models is 5MB.

  2. Cd to '/TestCode/code', run the following scripts.

    You can use the following script to test the algorithm

    #PLCNN
    CUDA_VISIBLE_DEVICES=0 python main.py 

Results

All the results for HeLa, CHO, Endo, Trans and Yeast.

Quantitative Results

Performance comparison with machine learning and CNN-Specific algorithms. The “Endo” and “Trans” is the abbreviation for LOCATE Endogenous and Transfected datasets, respectively. Best results are highlighted in bold.

Performance against traditional CNN methods using Yeast and HeLa datasets. The best results are in bold.

The effect of decreasing the training dataset. It can be observed that the performance decrease for traditional ensemble algorithms with the decrease in training data while, on the other hand, PLCNN gives a consistent performance with a negligible difference.

ETAS accuracies for individual members of ensemble on CHO dataset for tau = 40.

For more information, please refer to our papar

Confusion matrices

The confusion matrices for different datasets.

Confusion matrix for CHO dataset. The rows present the actual organelle class while the columns show the predicted ones. The results are aggregated for 10-fold cross-validations. The accuracies for each class are summarized in the last row as well as columns.

Confusion matrix for Yeast dataset. The predicted organelle are shown in the columns while the true values are present in the rows. The summaries of accuracies are given in the last row and column.

The correct predictions are highlighted via green while the red depicts incorrect. Our method prediction score is high for true outcome and vice versa.

The average quantitative results of ten execution for each method on the HeLa dataset. Our PLCNN method consistently outperforms with a significant margin.

Visualization results from Grad-CAM. The visualization is computed for the last convolutional outputs, and the corresponding algorithms are shown in the left column the input images

Citation

If you find the code helpful in your resarch or work, please cite the following papers.

@article{tahir2019PLCNN,
  title={Deep localization of protein structures in fluorescence microscopy images},
  author={Tahir, Muhammad and Anwar, Saeed and Mian, Ajmal and Wahab Muzaffar, Abdul},
  journal={Neural Computing and Applications (NCAA)},
  year={2021}
}

About

Deep Localization of Protein Structures in Fluorescence Microscopy Images, Neural Computing and Applications (NCAA), 2021

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published