Skip to content

docs: add scoring system, rules, and contribution track#45

Merged
msaroufim merged 23 commits intogpu-mode:mainfrom
yf225:docs/scoring-and-rules
Mar 14, 2026
Merged

docs: add scoring system, rules, and contribution track#45
msaroufim merged 23 commits intogpu-mode:mainfrom
yf225:docs/scoring-and-rules

Conversation

@yf225
Copy link
Contributor

@yf225 yf225 commented Mar 13, 2026

Summary

  • Add scoring section with point allocation table and ranking formula (top 10 per kernel earn performance points)
  • Add rules & requirements (correctness, submission format)
  • Add open-ended contribution track (autotuner improvements, bug fixes, tooling, docs)

Test plan

  • Review scoring formula matches hackathon spec
  • Verify point allocations are correct
  • Check rules match latest requirements

🤖 Generated with Claude Code

yf225 and others added 9 commits March 13, 2026 15:42
Documents ENABLE_TILE=0 vs ENABLE_TILE=1 and the TileIR compilation
pipeline available via nvtriton on B200 instances. Covers how to enable
TileIR with Helion (ENABLE_TILE=1 + HELION_BACKEND=tileir), the
different tunables (num_ctas/occupancy vs num_warps/maxnreg), and how
to hardcode TileIR configs in submissions.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Group both sections under a single "Optional: Extra Performance Knobs"
heading to emphasize neither is required. Streamline both into
step 1 (autotune) / step 2 (hardcode) format. Add a "Which combination"
section showing all 4 options to try.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Remove duplicate bash export block — the Python os.environ in the
code example is sufficient for both local autotuning and submissions.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Add point allocation table, scoring formula (correctness + performance
ranking), rules & requirements, and the separate open-ended contribution
track for non-kernel Helion contributions.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
@codecov
Copy link

codecov bot commented Mar 13, 2026

Codecov Report

✅ All modified and coverable lines are covered by tests.

📢 Thoughts on this report? Let us know!

yf225 and others added 10 commits March 13, 2026 16:21
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Each submission uses one static helion.Config for all shapes, not
per-shape configs. Simplified rules to reflect this.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Use a factory function (_make_kernel) to create kernel variants with
different helion.Config objects, and dispatch in custom_kernel() based
on input tensor shapes. This lets participants optimize each benchmark
shape independently.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
…st vs benchmark shapes

Test shapes: TODO to replace with default config or any config that passes correctness.
Benchmark shapes: TODO to replace with autotuned config.
Also add instructions on getting default config via autotune_effort="none".

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Per-shape configs are the recommended approach. Remove mentions of
using a single config across all shapes.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Configs are always participant-provided.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
@yf225 yf225 closed this Mar 14, 2026
@yf225 yf225 reopened this Mar 14, 2026
yf225 and others added 4 commits March 13, 2026 18:19
The previous description incorrectly stated geometric mean of 100 runs.
The actual helion eval uses CUDA graphs with L2 cache clearing, 10
measurements, and arithmetic mean.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
The LOC-based rule was gameable (denominator inflation with padding code),
so switch to a qualitative rule: inline triton/asm is allowed as escape
hatches, but predominantly inline submissions may be disqualified.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Spawn mode isolates each autotuner trial in a subprocess with timeout
protection, preventing hangs or crashes from killing the entire run.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
@msaroufim msaroufim merged commit 506fbd4 into gpu-mode:main Mar 14, 2026
7 of 8 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants