T-math / README.md
the-hir0's picture
Update README.md
2e3558a verified
metadata
license: apache-2.0
task_categories:
  - question-answering
language:
  - ru
pretty_name: T-math
size_categories:
  - n<1K
dataset_info:
  features:
    - name: question
      dtype: string
    - name: verifiable_answer
      dtype: string
    - name: year
      dtype: string
    - name: grade
      dtype: string
    - name: full_answer
      dtype: string
    - name: solutions
      list: string
    - name: task_complexity
      dtype: string
    - name: olympiad
      dtype: string
  splits:
    - name: train
      num_bytes: 510955
      num_examples: 331
  download_size: 228445
  dataset_size: 510955
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*

🧮 T-Math

T-Math is a dataset of Russian math olympiad problems created to assess the reasoning capabilities of large language models (LLMs) in mathematics.
It includes 331 problems from the All-Russian School Olympiad and the Moscow Olympiad for high school students, covering the period from 1998 to 2025.
The tasks and their ground-truth answers were extracted automatically and subsequently verified by human assessors.

Key features:

  • Challenging problems that require multi-step reasoning (median completion length for Qwen3-32B is 16K tokens), sourced from top-tier Russian olympiads
  • Easily verifiable: answers are numeric-only and checked using the math_verify library to compare mathematical expressions
  • Not yet saturated, even by frontier reasoning models such as Gemini 2.5 Pro and DeepSeek R1
  • Contains 331 samples — the largest Russian math olympiad-level benchmark — making it more statistically robust compared to smaller datasets like the 30-sample AIME benchmark

📊 Evaluation Results

Model pass@1
o4-mini-high 0.73
DeepSeek-R1-0528 0.71
Gemini-2.5-Pro 0.70
Claude Sonnet 4 0.56
T-pro-it-2.0 0.54
Qwen3-32B 0.53

🗂️ Filtering procedure

The text was extracted from PDFs using Qwen/Qwen2.5-VL-72B-Instruct. Tasks, along with their ground-truth and verifiable (numeric) answers, were extracted via LLM calls. We filtered out invalid questions using an LLM based on the following criteria:

  • Tasks requiring multiple answers
  • Tasks without a single correct answer
  • Theorem-like tasks where the main goal is proving a statement, making automatic verification non-trivial
  • Tasks with non-numeric answers, to simplify answer comparison
  • Tasks that cannot be solved without access to an accompanying image

Next, we removed tasks of moderate difficulty where Qwen3-8B achieved a 100% pass@16 rate, as they offer limited value for benchmarking reasoning. Finally, both the questions and the verifiable answers were manually reviewed by assessors to ensure consistency with the original sources.

🛠️ How to use

Add the following system prompt to guide the model to return the final answer in a \boxed{} tag, making it easier to parse:

Решите следующую математическую задачу эффективно и ясно. Последняя строка вашего ответа должна иметь следующий формат:
'Таким образом, окончательный ответ: $\boxed{ОТВЕТ}$.' (без кавычек), где ОТВЕТ - это просто окончательное число или выражение, решающее задачу.
Думайте шаг за шагом перед ответом.

You can then use the following code snippet with the math_verify library to compare mathematical expressions:

from math_verify import LatexExtractionConfig, parse, verify
from latex2sympy2_extended import NormalizationConfig


def accuracy_reward(completion: str, solution: str) -> float:
    """Reward function that checks if the completion matches the ground truth."""
    # parse the gold solution (assumed to always succeed)
    gold_parsed = parse(solution, extraction_mode="first_match")

    # parse the model’s completion with the same LaTeX extraction settings
    answer_parsed = parse(
        completion,
        extraction_config=[
            LatexExtractionConfig(
                normalization_config=NormalizationConfig(
                    nits=False,
                    malformed_operators=False,
                    basic_latex=True,
                    equations=True,
                    boxed="all",
                    units=True,
                )
            )
        ],
        extraction_mode="first_match",
    )

    # verify and return binary reward; on error, print and give 0.0
    try:
        return float(verify(gold_parsed, answer_parsed))
    except Exception as e:
        print(f"verify failed: {e}, answer: {answer_parsed}, gold: {gold_parsed}")
        return 0.0

📖 Citation

If you find our work useful in your research, please consider citing the following paper:

@inproceedings{stoianov-etal-2026-pro,
    title = "{T}-pro 2.0: An Efficient {R}ussian Hybrid-Reasoning Model and Playground",
    author = "Stoianov, Dmitrii  and
      Taranets, Danil  and
      Tsymboi, Olga  and
      Latypov, Ramil  and
      Dautov, Almaz  and
      Kruglikov, Vladislav  and
      Nikita, Surkov  and
      Abramov, German  and
      Gein, Pavel  and
      Abulkhanov, Dmitry  and
      Gashkov, Mikhail  and
      Zelenkovskiy, Viktor  and
      Batalov, Artem  and
      Medvedev, Aleksandr  and
      Potapov, Anatolii",
    editor = "Croce, Danilo  and
      Leidner, Jochen  and
      Moosavi, Nafise Sadat",
    booktitle = "Proceedings of the 19th Conference of the {E}uropean Chapter of the {A}ssociation for {C}omputational {L}inguistics (Volume 3: System Demonstrations)",
    month = mar,
    year = "2026",
    address = "Rabat, Marocco",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2026.eacl-demo.22/",
    doi = "10.18653/v1/2026.eacl-demo.22",
    pages = "297--319",
    ISBN = "979-8-89176-382-1"
   }