Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batchsize in Finetuning is irrelevant #21

Open
phquanta opened this issue Oct 16, 2021 · 1 comment
Open

Batchsize in Finetuning is irrelevant #21

phquanta opened this issue Oct 16, 2021 · 1 comment

Comments

@phquanta
Copy link

phquanta commented Oct 16, 2021

I've noticed that BatchSize even it is there for Fine Tuning it is always 1 to make sure it functions properly. If one wants to set it bigger than 1, then it triggers an error due to a fact that self.max_len=0 and no padding takes place. I don't know how it would affect training if one uses max_len vs not using max_len with batchsize=1.

@taoshen99
Copy link

I'm having a similar error as @phquanta when setting "finetune_batch_size" >=2. It seems that self.max_len would always be 0 when data_type is set as 'finetune'. This causes no 'A' padding to raw SMILES. As a result, each numpy.arrary(X) (line105 in data_loader.py) differs in axis 0 and could not be concatenated.
I believe that SMILES in finetune should also be padded to the max length.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants