Skip to content

Record: Order-Adaptive BackoffMixer (mean val_bpb=0.5440)#825

Open
hypery11 wants to merge 1 commit intoopenai:mainfrom
hypery11:submission/2026-03-26_final_champion
Open

Record: Order-Adaptive BackoffMixer (mean val_bpb=0.5440)#825
hypery11 wants to merge 1 commit intoopenai:mainfrom
hypery11:submission/2026-03-26_final_champion

Conversation

@hypery11
Copy link

Results

Seed val_bpb Eval time
42 0.5437 ~391s
1337 0.5450 ~391s
2024 0.5434 ~391s
Mean 0.5440
Std 0.0008
  • Artifact: ~16.0 MB
  • Train: 600s on 8xH100 SXM
  • Eval: ~391s (well under 600s)

Method

11-layer transformer (512d, 8/8 full MHA, XSA-all, LeakyReLU(0.5)^2, 3.5x MLP). Order-adaptive entropy-gated BackoffNgramMixer with per-order entropy thresholds. Score-first, backward-looking, deterministic.

Acknowledgments

Huge thanks to the incredible community that made this possible:

This competition has been an amazing collaborative experience. Every improvement here builds on ideas shared openly.

  • 8xH100 SXM, train <=600s
  • Eval <=600s (391s)
  • Artifact <=16MB
  • 3-seed validation (std 0.0008)

Seeds: 0.5437 / 0.5450 / 0.5434 (std 0.0008).
Order-adaptive entropy gating + BackoffNgramMixer.
~16MB artifact. Train 600s, eval 391s.
@MatoTeziTanka
Copy link

Really impressive work — the order-adaptive entropy gating with per-order thresholds is a thoughtful design, and the 3-seed consistency (std 0.0008) is excellent. The acknowledgments section is also great to see — this competition has been genuinely collaborative.

One thing to flag: checking the log output, it looks like seeds 42 and 2024 may exceed the 16,000,000 byte artifact cap:

  • Seed 1337: 15,948,371 bytes ✅
  • Seed 42: ~16,022,243 bytes (over by ~22K)
  • Seed 2024: ~16,030,231 bytes (over by ~30K)

We ran into the exact same issue on our PR #769 seed 42 (over by 25,731 bytes) and had to rerun with tighter quantization. It's a subtle one — the submission.json may not reflect the per-seed sizes accurately.

Might be worth double-checking the individual seed artifact sizes against the 16,000,000 limit before the maintainers review. The fix for us was minor — just tightening the compression/quantization slightly to get the headroom.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants