reasoning-core is a suite of textual procedural data generators for language model pre-training and post-training. It is centered on expressive symbolic and algorithmic tasks, including full fledged first-order-logic, formal mathematics with TPTP, planning, and CFG syntax tasks.
We release pre-generated data scaled to more than 10B tokens
🤗 https://hf.co/collections/reasoning-core/datasets
uv pip install reasoning-core
from reasoning_core import list_tasks, get_task, score_answer
T = get_task('arithmetics')()
x = T.generate_example()
assert score_answer(x.answer, x)==1GALLERY (names link to task code)
planning · table_qa · table_conversion · equation_system · code_execution · diff_prediction · diff_patching · regex_following · regex_induction · graph_pathfinding · graph_node_centrality · graph_cycle_detection · graph_isomorphism · arithmetics · symbolic_arithmetics · sequential_induction · conjecture_entailment · proof_reconstruction · bayesian_association · bayesian_intervention · logic_nli · evidence_retrieval · parsability · parsing · continuation · set_intersection · set_missing_element · count_elements · set_equality
Run bash run_generate.sh for multi-threaded generation to json files (readable by Huggingface Datasets).
#!pip install uv #install uv if needed
!uv tool install prime --with openai -q
!uv tool run prime -- env install sileod/reasoning-core-env
from verifiers import load_environment
import os; from openai import OpenAI
env = load_environment("reasoning-core-env")
client = OpenAI( base_url="https://openrouter.ai/api/v1", api_key=os.getenv("OPENROUTER_API_KEY")) #🔑
results = env.evaluate(client=client, model="gpt-4.1-mini", num_examples=20, rollouts_per_example=1)
df=env.make_dataset(results).to_pandas()We use a custom interface but compatible interface. Our tasks, which are mostly orthogonal to RG, can be imported in it.
import reasoning_gym, reasoning_core
from reasoning_gym.composite import DatasetSpec
reasoning_core.register_to_reasoning_gym() # registers RC tasks into RG
specs = [
DatasetSpec(name='leg_counting', weight=1, config={}), #from reasoning_gym 🏋
DatasetSpec(name='arithmetics', weight=1, config={}), #from reasoning_core ◉
]
D=reasoning_gym.create_dataset('composite', size=10, seed=42, datasets=specs)And the other way around:
frm reasoning_core import get_task
t=get_task('reasoning_gym')
t.generate_example(level=1, rg_task='lcm') #or unspecified for random taskhttps://arxiv.org/abs/2603.02208
@article{reasoningcore2026,
title={Reasoning Core: A Scalable Procedural Data Generation Suite for Symbolic Pre-training and Post-Training},
author={Lacombe, Valentin and Quesnel, Valentin and Sileo, Damien},
journal={arXiv preprint arXiv:2603.02208},
year={2026},
url={https://arxiv.org/abs/2603.02208}
}