Skip to content

leaxsd/Deep-Learning-Super-Resolution-Image-Reconstruction-DSIR

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

47 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Deep-Learning Super-resolution Image Reconstruction (DSIR)


Super-resolution microscopy techniques (PALM, STORM…) can improve spatial resolution far beyond diffraction limit trough recording many frames of sparse single events emission (fluorescence labels). Low density (LD) data acquisition can provide tens of thousand of images of sparse events that can be readily localized using standard fitting algorithms (e.g, Thunder-STORM). However, the acquisition of such large number of images takes some time and can result to the sample drift and degrade the sample due to the intense light exposion. High density (HD) data of several hundred images can be faster in acquisition time but result of a very dense number of events per frames, which compromise the performance of the fitting localization algorithms. This repository proposes a method that use convolution neural network (ConvNet) auto-encoder [1] to reconstruct a localization images from HD datasets.

Fig.1 - ConvNet auto-encoder representation for super-resolution image reconstruction.

These work were probably developed in parallel with the work of Elias Nehme, et all. published in [2]. Both achieved very similar rehttps://developer.nvidia.com/academic_gpu_seedingsults despite different coding languages.

ConvNet auto-encoder proposed here were coded in PyTorch. We define an auto-encoder architecture which consider and input array of size (208, 208, 1), original image size of 26x26 px resized by a factor of 8. The ground-truth label is a pixel converted image (208, 208, 1) from the list emitters positions (x, y). Both transformations are defined in data_load.py file.

Fig.2 - (Upper row) Training dataset ground-truth pixel localization image examples and (Lower row) correspondent trained auto-encoder predictions.

Files

  • data_load.py : Parse the raw data: Rescale input image, plot position to pixel map label and convert both (data and label) to torch tensor.
  • conv_autoencoder.py: training the ConvNet auto-encoder.
  • load_model.py: load a pre-trained model and test against the example Tubes HD dataset.

Code

Training Dataset

A dataset of randomly generated single emission events were used to training our model. ThunderSTORM imageJ plugin data generator were used to generate training and validation datasets with the following settings:

General setup: Camera setup:
size: 26x26 px Pixel Size: 100 nm
train set: 3500, val set: 1500 Photoelectrons per A/D counts: 1.0
PSF: Integrated Gaussian Base level: 100
FWHM range: 200:300 nm
Intensity range: 80:2050 photons
Density: 2.0 emitters/μm2
BG noise: 20

You can also download the used dataset here. The raw dataset should be uncompressed at ~/data/dataset/ folder.

Auto-encoder

The 3 layers Encoder and 4 layers Decoder are defined following the code bellow:

def forward(self, x):
    # Encode
        x = self.conv1(x)
        x = self.pool(F.relu(self.bn1(x)))  # out [16, 104, 104, 1]
        x = self.conv2(x)
        x = self.pool(F.relu(self.bn2(x))) # out [8, 52, 52, 1]
        x = self.conv3(x)
        x = self.pool(F.relu(self.bn2(x))) # out [8, 26, 26, 1]
	# Decode
        x = self.convt1(x)
        x = F.relu(self.bn2(x)) # out [8, 52, 52, 1]
        x = self.convt2(x)
        x = F.relu(self.bn1(x)) # out [16, 104, 104, 1]
        x = self.convt3(x)
        x = F.relu(self.bn2(x)) # out [8, 208, 208, 1]
        x = self.conv4(x)
        x = F.relu(self.bn4(x)) # out [1, 208, 208, 1]

Loss function

Using an Adam optimizer, the loss function were defined as:

loss_function

where, is the label images, is the neural network predictions, is a Gaussian kernel and is the total number of images per batches. The operation denotes a 2D Gaussian convolution between and .

Training Parameters

The train() function in conv-autoencoder.py takes the following arguments:

  • epochs (int): total number of training epochs. [100]
  • lr (float): learning rate for the Adam optimizer. [1e-4]
  • batch_size (int): batch size of training and validation dataset. [32]
  • seed: randomization seed number. [99]
  • kernel_width (int): size in pixel of the square Gaussian kernel (). [5]
  • kernel_fwhm (int): Full width of half maximum of the Gaussian kernel (). [3]
  • verbose (boolean): defines whether show up the visdom output results. [True]
  • save (boolean): defines whether save or not the training model on the end (or KeyboardInterrupt) on training. [True]
  • model_path (path): path where to save the model. [None]

The preset parameters will save the model on path:

~/data/temp/_timestamp_/cae_model__timestamp_.pt

where ṭimestamp = day month year _ hour minute second string.

Visdom

We use visdom to track the results during the training process. You need to run visdom server beforehand:

python -m visdom.server

then you can open in your browser http://localhost:8097/# .

For each training epoch you you'll have a set of 4 images with the prediction image reconstruction and the correspondent pixel ground-truth localization and the loss function value plot.

Fig.3 - Ground-Truth pixel localization, auto-encoder prediction image reconstruction and training loss visdom output.

Results

We trained a model for 100 epochs using the [train dataset](#Training Dataset) and later a fine tuning with a smaller learning rate and a kernel_fwhm = 1. Resulted model can be found on model/autoencoder_model.pt.

Test Dataset

In order to test our trained model we used the Tubes HD dataset available from the 2013 IEEE International Symposium on Biomedical Imaging Challenge [3]. (Download Here)

localization Image Reconstruction

Fig.4 - Tubes HD dataset. (Left) single input frame image (Right) Zoom at red square and the ground-truth emitters positions(red crosses).

Fig.5 - (Left) Single frame DSIR and (Right) comparison with ground-truth emitters positions (red crosses).

Fig.6 - (Left) Reconstruction of the 361 frame of the Tubes HD dataset. (Right) Zoom (green square Fig.4) image and ground-truth emitters position (red dots). The total reconstruction time for all frames is about 3 sec.

Discussions

Here we presented a ConvNet auto-encoder model applied to high density recorded stochastic super-resolution microscopy data. Although not discussed in here, this DSIR can improve performance compared with the state-of-the-art fitting algorithms that usually performed poorly with high density data. Also DSIR can be much faster the the most of fitting algorithms.

The current performance might be improved by feeding the same auto-encoder with different training dataset, E.g., experimental measured sparse data where the label positions are define via a standard fitting algorithm. Despite, DSIR be able to reconstruct localization super-resolution images, it lacks in terms of quantitative information since the ConvNet auto-encoder doesn't outputs the localization coordinates of each detected event.

Finally, this kind of approach has an import hole because employs machine learning as a scientific tool to retrieve unaccessible information rather than a simple analysis tool.

References

  1. https://blog.keras.io/building-autoencoders-in-keras.html
  2. https://arxiv.org/abs/1801.09631v2
  3. http://bigwww.epfl.ch/smlm/challenge2013/index.html

forthebadge forthebadge

About

Deep-Learning convolutional auto-encoders applied to super-resolution microscopy data to localization image reconstruction.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages