Skip to content

leanprover/lean-eval-leaderboard

Repository files navigation

lean-eval-leaderboard

View the leaderboard →

Results store and public website data source for the lean-eval benchmark.

This repository holds machine-written artifacts produced by the lean-eval CI and by the leaderboard-site build pipeline.

  • results/ is the append-only public success log written by lean-eval CI.
  • site-data/ is the derived, presentation-oriented data consumed by the public website.

Do not edit generated files here by hand.

Successes are sticky: once a problem is marked solved for a given user, that record is never modified or removed, even if a later submission from the same user no longer proves it.

Submitter-facing instructions live on the Verso website at LeaderboardSite/Pages/Submit.lean. This README is for people building or operating the leaderboard repository, not for people trying to submit proofs.

File layout

results/
  <github-login>.json

site-data/
  problems.json
  leaderboard.json

One file per submitter. Users without any successful submission have no file. Filenames use the user's GitHub login, lowercased, since GitHub logins are case-insensitive.

The site-data/ directory is generated from the raw results/ files together with benchmark metadata imported from the benchmark repository. Its schema is documented in docs/site-data-schema.md. Its files are derived (.gitignored) and produced by scripts/generate_site_data.py, which the deploy workflow runs before every site build, and which local developers invoke via lake script run generate (the generate script in lakefile.lean runs the generator and then builds the site). The deploy pins the benchmark sibling clone to the commit recorded in benchmark-snapshot/.benchmark-commit, so the regenerated site-data and the checked-in snapshot's catalog stay in lockstep.

Record schema (v2)

{
  "schema_version": 2,
  "user": "kim-em",
  "solved": {
    "Claude Opus 4.6": {
      "two_plus_two": {
        "solved_at": "2026-04-11T10:45:00Z",
        "benchmark_commit": "8e1b9cf5e1d3c2b1a0f9e8d7c6b5a4938271605f",
        "submission_repo": "kim-em/my-lean-eval-proofs",
        "submission_ref": "deadbeefcafef00dbaadc0de1234567890abcdef",
        "submission_public": true,
        "issue_number": 42
      },
      "list_append_singleton_length": {
        "solved_at": "2026-04-12T08:15:30Z",
        "benchmark_commit": "8e1b9cf5e1d3c2b1a0f9e8d7c6b5a4938271605f",
        "submission_repo": "kim-em/my-lean-eval-proofs",
        "submission_ref": "deadbeefcafef00dbaadc0de1234567890abcdef",
        "submission_public": false,
        "issue_number": 43
      }
    },
    "GPT-5.5": {
      "two_plus_two": {
        "solved_at": "2026-04-12T11:00:00Z",
        "benchmark_commit": "8e1b9cf5e1d3c2b1a0f9e8d7c6b5a4938271605f",
        "submission_repo": "kim-em/another-gist",
        "submission_ref": "0123456789abcdef0123456789abcdef01234567",
        "submission_public": true,
        "issue_number": 44
      }
    }
  }
}

Top-level fields

Field Type Description
schema_version integer Currently 2.
user string GitHub login with original case preserved.
solved object Map from model name to per-problem solution records. Never empty.

Per-model bucket

The keys of solved are free-form model identifiers as supplied on the submission form (e.g. "Claude Opus 4.7", "GPT-5.5", "Aristotle (Harmonic)"). Each value is an object mapping <problem_id> to a solution record.

Per-problem solution record

Field Type Description
solved_at string ISO 8601 UTC timestamp of when the record was first written.
benchmark_commit string 40-character SHA of the leanprover/lean-eval commit evaluated against.
submission_repo string Identifier of the submission source: owner/repo for a GitHub repository, or user/gist-id for a gist.
submission_ref string 40-character SHA pinning the submission at evaluation time. For repositories this is a commit SHA; for gists this is the gist revision SHA.
submission_public boolean true if the submission source was public at evaluation time, false otherwise. The leaderboard site uses this to decide whether to link to the solution.
issue_number integer Issue number in leanprover/lean-eval that triggered the evaluation.
production_description string | absent Optional free-form description of how the solution was produced.

The model field has moved out of the per-problem record and become the key of the surrounding bucket.

Write semantics

When the lean-eval CI records a successful submission:

  1. It reads results/<login>.json, or starts from an empty solved map.
  2. The submission carries one model name (the value of the issue form's Model field). The CI uses that as the bucket key.
  3. For each problem that passed in the submission:
    • If solved[<model>][<problem_id>] already exists, do nothing (sticky no-op).
    • Otherwise, create the model bucket if needed, and add a new record with the fields above.
  4. If at least one new record was added, the CI commits and pushes the updated file. If no new records were added, it makes no commit.

The map only ever grows: existing buckets keep their problems, and existing problem records keep their original solved_at timestamps and audit trail (benchmark_commit, submission_ref, issue_number).

A single user can therefore claim the same problem under several model names — one record per (user, model, problem) triple. This is the change from v1, which keyed sticky records on (user, problem) only and so could record at most one model per problem per user.

Commit convention

Commits by the lean-eval CI use the message form:

record: <login> solved <problem_id>[, <problem_id>...] using <model> @ <benchmark_short_sha>

One commit per submission, grouping all newly-recorded problems for a single model together.

Schema evolution

Breaking changes bump schema_version. Consumers should refuse to parse a file whose schema_version they do not know. Non-breaking additive changes (new optional fields) keep the version number stable.

Website build direction

The public website should live in this repository and consume derived artifacts from site-data/, not the raw results/ tree directly. The site itself will be implemented in Verso.

  • Raw results preserve audit information and sticky solve history.
  • Derived site data resolves aggregation rules, ranking, and problem metadata.
  • The frontend should render site-data/leaderboard.json and site-data/problems.json without re-implementing benchmark logic in the browser.

See:

About

Results for the lean-eval benchmark (https://github.com/leanprover/lean-eval)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors