-
Notifications
You must be signed in to change notification settings - Fork 37
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Help in finetuning generator #21
Comments
The steps you tried are quite reasonable! The missing ingredient is that preprocessing computes various vocabularies of motif/atom types based on the data, which affects the shapes of some layers in the model, but those shapes are already fixed if a pretrained model is loaded, hence the shape mismatch (intuitively, this is saying that the number of motif/atom types found in your finetuning dataset was smaller than the number of types found in the original dataset, which is expected). So, one would have to tell preprocessing to use the vocabularies from the pretrained checkpoint instead of computing them afresh. This isn't supported in the current code (we did briefly experiment with fine-tuning, but not enough for this to end up in the release), but shouldn't be hard to add. I'll hack something together this week and then share with you as a branch; once you test it out I can then make a PR and merge it into |
Oh, I see, I get now. Thank you so much! |
@gmseabra: can you pull The only change to the workflow you described would be passing
means that you will do some multiple of Let me know how this goes! |
@gmseabra Did you have any luck with fine-tuning? |
Hi! I've been trying to replicate this example with the steps you provided above, where I try to finetune it on a small set of 3K molecules, I still encounter the following error (running everything on Colab now). I just took the existing checkpoint, and I wanted to finetune it to a smaller set: !molecule_generation preprocess input output trace !molecule_generation train MoLeR trace But even when aligning the metadata that the model was originally trained with (I took the same Guacamol files), it still didn't want to run. This is the error I encountered:
I'm not sure whether I did something wrong, and where I could fix it! |
Hi @MeilinaR, sorry for the silence, I forgot to respond to this one. Are you still having issues? I noticed the command you pasted passes |
Closing due to lack of activity. |
Hi,
I'm trying an experiment, to fine-tune the generator with a small set of molecules with specific properties, (so it will generate new molecules with similar properties) but I'm running into some errors that I have been unable to solve. I'd really appreciate if anyone could shed some light into what I'm doing wrong.
What I'm doing:
finetune_moler/input
.$ molecule_generation preprocess finetune_moler/input finetune_moler/output finetune_moler/trace
The pre-process step seems to run just fine. But in the fine-tuning step, I'm getting the following error:
Could someone point me to what I'm doing wrong? Would it be possible to get an example of successfully fine-tuning the model?
Thanks a lot!
Gustavo.
The text was updated successfully, but these errors were encountered: