[MAISI] The Role of FOV at inference time using finetuned Diffusion Net #2038
Unanswered
alejandrodl
asked this question in
General
Replies: 1 comment
-
Thank you for the sharing!!! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I'm sharing here an interesting issue my team and I just had, along with its solution, which was kindly provided by the MAISI team.
We tried finetuning the MAISI’s Diffusion Net model on a dataset of 221 medical volumes that did not belong to the original model's training data. We ran the training with the tutorial notebook provided and after that, when we ran inference, the generated images seemed to collapse into noise:
In the end, the model happened to be finetuned correctly, only that the FOV (Field Of View) at inference was different from the one at training time, hence the mismatch. When adjusted correctly, we got appropriate generation results:
This magnitude
FOV = dimensions * spacing
at inference has to fall within the range of that of training samples, otherwise the generations would be incorrectly computed. In order to adjust inference parameters accordingly, one must:dim
to numbers divisible by 128 (e.g. 239x221x137 -> 256x256x128). This isnew_dim
parameter.new_spacing = spacing * (dim/new_dim)
config_maisi_diff_model.json
file, changedim
variable tonew_dim
andspacing
tonew_spacing
.If one runs inference with these two new parameters, generations should be correct.
P.S. Thanks a lot to the MAISI team, keep rocking! :)
Beta Was this translation helpful? Give feedback.
All reactions