-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
BC train maze dataset #7
Comments
Hello, |
Merged
Hello! I made a new pull request which should resolve this issue. :) Let me know if not! |
Hello! Closing as issue is resolved. Open another issue if there is another problem. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hello,
While trying to train BC model on the FO Maze task with the command
python -m llm_rl_scripts.maze.bc.fully_observed_bc HF gpt2 data_path --outputs_path=output_path
, I have noticed that the eval_frac parameter is set by default to 0.1.The data is splitted using this code:
Is it correct that you use 10% of train data to actually train and 90% just to evaluate during the train process? train_items has length 124, which is less than the train_bsize = 128, batch size you are using, that is why it seems to me counterintuitive. Please, clarify this part.
Additionally, I had an issue with the training with multiple epochs, because it seems that Seq2SeqDataset is needed to be created instead of Seq2SeqIterableDataset, because Seq2SeqIterableDataset is iterable and has no length, and after
steps_per_epoch = len(dataset) // bsize if isinstance(dataset, Dataset) else None
line, every epoch has no steps to train.Finally, I could not run the 'easier BC code' you have added in one of the commits because of the import module issues (jax_agent, jax_bc and some others are missing)
I look forward to your response.
The text was updated successfully, but these errors were encountered: