Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Removed all all-zero masks #116

Merged
merged 2 commits into from
Aug 29, 2023
Merged

Removed all all-zero masks #116

merged 2 commits into from
Aug 29, 2023

Conversation

kywch
Copy link
Collaborator

@kywch kywch commented Aug 28, 2023

No description provided.

@jsuarez5341
Copy link
Collaborator

We learned some stuff from this:

  • It does not appear that we have all 0 masks in the forward pass, yet somehow 0 entropy and very negative logits are leaking into the training batch
  • KL is already exploding by the time entropy crashes

Probably, we are going to need to stain the masked items with large, unique negative values to figure out what is leaking. This is likely part of a bigger problem with the complexity introduced into clean_pufferl to handle variable numbers of agents

@jsuarez5341 jsuarez5341 merged commit eca6938 into 2.0 Aug 29, 2023
6 checks passed
@kywch kywch deleted the no-zero-mask branch September 7, 2023 22:05
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
2 participants