-
Notifications
You must be signed in to change notification settings - Fork 16
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training dataset #8
Comments
Hello! We originally had our own complex normalization strategy, tailored for SIM, but we received a lot of requests for raw data. Accordingly, we made the raw data public, and used the raw data with simpler normalizations. All the paper's results are on these raw data, which we used for training all our models. You can ignore the old data from earlier commits. For Widefield: we applied z-score normalization. The z-score is computed across all 360x400 captures; Regarding the preparation code, we now modified the notebook Thank you for your positive feedback. Please only use the raw data, and we hope it will be useful for your work. |
Thanks for clarifying this! |
Hello @majedelhelou sorry for reopening the issue. For LR (Widefield) images, it looks straightforward since it's just a z-score standardization as you commented.
Thanks again 👍 |
Hello @nmhkahn Yes correct for the LR Widefield images, we note in the readme the mean and standard deviation values we computed for the z-score across all the 400 samples times 360 FOVs.
These points should not matter too much [the effect of using a noisy or noise-free LR Widefield image for normalizing the HR SIM, and the choice of the downsampling kernel] and you can also use your own strategies for normalizing the HR SIM. The goal is to get a somewhat more uniform intensity distribution for the networks, and small fluctuations in how we get them should not have a significant impact. Hope this helps! |
Hi. thanks for sharing a great dataset!
In the paper, it says "The training set consists of 240 LR and HR image sets, and the test set consists of 120 sets of images".
But in the
data/normalized
directory, I only can get 120 images, which might be the test set only.I downloaded the raw dataset as well but I think that the dataset preparation code and the raw data are not matched (preparation code assumes the image as png, but raw data is npy files).
I also saw that there was a training dataset on this repo in an early commit (c674e02), but not sure that it is safe to use these old training images.
Please let me know that 1) training dataset from c674e02 is right and 2) if not, where can I get the training dataset.
The text was updated successfully, but these errors were encountered: