Skip to content

distributed sampler fix#467

Merged
jpata merged 14 commits intomainfrom
jp_20260325_samplerfix2
Mar 28, 2026
Merged

distributed sampler fix#467
jpata merged 14 commits intomainfrom
jp_20260325_samplerfix2

Conversation

@jpata
Copy link
Copy Markdown
Owner

@jpata jpata commented Mar 25, 2026

The distributed sampler was causing issues on LUMI with 8 GPUs if different workers ran out of the dataset at different times. The tests are expected to fail, there is an outdated test on the main branch that is removed in this PR.

TODO:

  • wait until pyg-cld-hits-v1_cld_20260326_093401_088146 has trained on LUMI for 24h and post the loss curves to check that it doesn't crash now

The learning rate might need tuning but it seems there are no more issues with the sampler. EDIT: found a crash on 2xL40s after 40k steps, fix attempt in be6b00a.

Screenshot 2026-03-27 at 08 09 03
  • wait until pyg-cld-hits-v1_cld_20260327_081645_454635 has trained on LUMI after fixing be6b00a.
  • wait until pyg-cld-hits-v1_cld_20260327_081548_787211 has trained on Tallinn, same as above
Screenshot 2026-03-28 at 08 07 00

@jpata jpata merged commit b4e03f2 into main Mar 28, 2026
3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant