‼️ Rules Update & Clarifications - Action Required ⏳

Hi everyone :wave:

Thank you for the thoughtful questions and feedback over the past weeks. Based on those discussions, we’ve consolidated and clarified the Official Rules for the Global Chess Challenge 2025.

:point_right: Action required: To continue participating, please visit the challenge page, click Participate, and accept the updated rules.

The challenge overview page has been updated to reflect the latest and most accurate information. Please review it for the newest details, links, and resources.

:date: Timeline (UTC): Deadline Extension

  • Round 1: Dec 4, 2025 → Dec 31, 2025 (23:55)
  • Round 2: Jan 1, 2026 → Jan 31, 2026 (23:55)
  • Final Tournament (no new submissions): Feb 1 → Feb 7, 2026
  • Winners announced: Feb 15, 2026

Note: Model submissions close on January 31, 2026. From February 1 onward, no new submissions are accepted; this period is reserved for the Final Tournament of eligible models and post-challenge verification.

:test_tube: Evaluation Structure (Rounds 1 & 2)

Rounds 1 and 2 use a baseline evaluation against fixed Stockfish opponents to ensure stability and comparability.

Each submission plays:

  • 50 games vs Stockfish Skill 0 (Depth 1)
  • 50 games vs Stockfish Skill 0 (Depth 5)

All games use identical positions, time controls, and compute constraints.

All evaluations are performed on a standardized AWS Trainium configuration, specifically a trn1.2xlarge instance, to ensure consistency and fairness across all submissions.

:bar_chart: Leaderboard Scoring

  • Primary metric: Average Centipawn Loss (ACPL)
    • Computed using Stockfish Level 20 (Depth 20) as the reference evaluator
  • Secondary metric: Win Rate
    • Used for tie-breaking and additional analysis

:memo: Important: Reasoning Text vs Scoring

  • Submissions must output both:
    • a chess move
    • a short textual explanation
  • Only the move inside <uci_move> tags is scored
  • Reasoning text is not evaluated or scored

:warning: Invalid move handling

  • Missing / malformed / illegal <uci_move> → up to 3 retries
  • Still invalid after retries → treated as a resignation (loss for that game, and a 1000 CPL for the resignation move)

Resignation & ACPL Handling (Clarification)

To address an issue identified with ACPL calculation for very short games, the evaluation logic has been updated as follows:

  • Intentional resignations are permitted, but
  • A resignation now incurs a fixed penalty of +1000 centipawns (CPL) applied to the resignation move.

This change ensures that:

  • Models cannot artificially achieve low ACPL by resigning early, and
  • ACPL remains comparable across games of different lengths.

This penalty applies only to the resignation move and does not otherwise alter move evaluation or game outcome handling.

:checkered_flag: Eligibility & Final Tournament

:white_check_mark: Advancement Criteria

  • After Round 2, submissions with ACPL lower than the official baseline model become eligible for Finals.

♜ Final Tournament Format

  • Swiss-style tournament
  • Rankings based only on game outcomes:
    • Win: 1 point
    • Draw: 0.5 points
    • Loss: 0 points
  • ACPL is not used in Finals

:abacus: Tie-breaks (in order)

  1. Head-to-head result (if applicable)
  2. Buchholz (or equivalent strength-of-opposition metric)
  3. Sonneborn–Berger (where applicable)
  4. Any additional rule announced before the Finals

:trophy:♜ Final prize winners are determined exclusively by the results of the Swiss-style Final Tournament, in accordance with the tournament scoring and tie-breaking rules. ♜:trophy:

:rocket: Submissions

  • Submission cap: increased to up to 20 submissions per team per day

:brain: Eligible Models & Backends

Only officially supported model types and execution backends are eligible, and all evaluations are performed exclusively on AWS Trainium using the AWS Neuron + vLLM backend.

:pushpin: Supported list:
global-chess-challenge-2025-starter-kit/docs/neuron-and-vllm-tuning.md at master · AIcrowd/global-chess-challenge-2025-starter-kit · GitHub.

:exclamation:Note: We recognize that some teams are currently facing challenges submitting models targeting the AWS Neuron backend for execution on AWS Trainium.

The organizers are working closely with the AWS team to improve support responsiveness and unblock submission issues, so that all compliant models can be evaluated under the official infrastructure.

:lock: Execution Constraints

All models are executed on organizer-controlled infrastructure and must operate as standalone language models:

  • :no_entry_sign: No tool calling
  • :no_entry_sign: No web access
  • :no_entry_sign: No chess engines
  • :no_entry_sign: No heuristic search or auxiliary decision systems

:heavy_check_mark: All decisions must be produced solely via token-level inference from the provided text input.

Attempts to bypass these constraints may result in disqualification.

:pray: Final Notes

Thanks again for the feedback and patience. We hope these clarifications make the evaluation and Finals structure fully transparent.

We’re excited to see how far participants can push reasoning quality, efficiency, and hardware-aware optimization on AWS Trainium.

All the best,
Team Global Chess Challenge

1 Like

So whatever i trained on Top of Llama 3.1 8B model is a waste because the model is not eligible anymore it is 8.03B. this is ridiculous! after 16 days you’re stating this?

Thanks for the detailed clarification and rules and the extended timeline guys, just need some clarity on 1) Does everyone qualify for round 2 automatically?
2) Will you total the round 1 and round 2 scores, and then pick the best X for finals, or just anyone who scores more than the baseline. Just wanted the relevance of Round 1 in the final selection?

This is the most disorganized Hackathon i Have ever Attended. The amount of GPU credits i wasted running and testing is quite frustrating. All these clarifications were supposed to be provided in the beginning . At least formation of a discord group is pretty standard. Its high time the organisers hold a Q&A atleast now!

@hasheerama They did a QA yesterday in the workshop actually, and now they have given you another extra month should solve for most problems i believe.

(post deleted by author)

okay now i am confused

  • Round 1: Dec 4, 2025 → Dec 31, 2025 (23:55)??

so what timeline was extended?

Round 2 ends Jan 31, Final Tournament Feb 1-7, Winners Feb 15

but that was never announced if i remember correctly wasnt it TBA earlier?

I dont know about others but i was using GPU credits i won in an earlier hackathon till now without knowing what is the right way to train

wasted 2+ weeks before releasing critical information:

  • Supported model architectures (just posted 9 mins ago in that forum post)
  • Neuron compilation requirements
  • vLLM parameter constraints
  • Submission command structure

WHAT NEEDS TO HAPPEN

  1. Extend Round 1 deadline by at least 2 weeks to account for the infrastructure information delay
  2. Provide additional AWS credits to compensate for compute spent on models trained without full technical specifications
  3. Confirm ALL technical documentation is now complete - no more critical details to be announced
  4. Provide a testing sandbox so participants don’t waste submissions debugging Neuron compilation issues

Nothings Gonna Happen!
Screenshot 2025-12-20 202957