Skip to content

CritPt-Benchmark/CritPt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Evaluation Pipeline for the CritPt Benchmark

Probing the Critical Point (CritPt) of AI Reasoning: a Frontier Physics Research Benchmark

|🌐 Website | 🤗 Dataset | 📖 Tech report |

CritPt (Complex Research using Integrated Thinking – Physics Test; reads as "critical point") is the first benchmark designed to test LLMs on unpublished, research-level reasoning tasks that broadly covers modern physics research areas, including condensed matter, quantum physics, atomic, molecular & optical physics, astrophysics, statistical physics, nuclear physics, high energy physics, mathematical physics, fluid dynamics, nonlinear dynamics and biophysics.

It currently includes 71 challenges and 190 checkpoints, crafted by a team of 50+ active physics researchers from 30+ leading institutions worldwide, including senior Ph.D. students, postdocs, professors and research scientists.

Updates

Nov 21, 2025: CritPt made its official debut on Artificial Analysis!

Nov 21, 2025: The evaluation pipeline and challenge dataset are now publicly available! To evaluate your own models on CritPt, please follow the instructions provided on this page. An example challenge can be found here. See 70 test challenges in this folder or on Hugging Face.

Leaderboard

Model Challenge Accuracy¹
GPT-5 (high, code & web) 12.6
GPT-5 (high, code) 10.6
Gemini-3 Pro² 9.1
GPT-5 (high) 5.7
Gemini-2.5 Pro 2.0
o3 (high) 1.4
DeepSeek R1 1.1
Gemini-2.5 Flash 1.1
o4-mini (high) 0.6
Claude Opus 4 0.3
GPT-5 (minimal) 0.0
Llama-4 Maverick 0.0
GPT-4o 0.0

¹ We use average accuracy over 5 runs × 70 test challenges as our primary performance metric.
² Result of Gemini 3 Pro (without tools) here is provided by Artificial Analysis (source). Our independent evaluation results will be released soon.

To evaluate your own LMs on CritPt challenges, follow the steps below

0. Setup

  • Clone this repository:
git clone https://github.com/CritPt-Benchmark/CritPt.git && cd ./CritPt
  • Install required packages:
pip install -r requirements.txt
  • Setup model API keys in .env as environment variables:
ANTHROPIC_API_KEY=
DEEPSEEK_BASE_URL=https://api.deepseek.com
GOOGLE_API_KEY=
OPENAI_API_KEY=
DEEPSEEK_API_KEY=
TOGETHER_API_KEY=

1. Generation of model responses

To generate responses using an API model, run the following command:

python -m critpt generate generate model=google/gemini-2.5-flash task_config=.config/multiturn_without_answer_without_tool.json

This command will launch the generation process with the specified model and task configuration.

Advanced Generation with Model-specific Configs

Below are the model configurations used in our evaluation. You can modify the parameters to customize the setup.

Claude & Gemini

python -m critpt generate epochs=5 model=google/gemini-2.5-pro,anthropic/claude-opus-4-20250514 reasoning_tokens=27000

DeepSeek

python -m critpt generate epochs=5 model=openai-api/deepseek/deepseek-reasoner max_connections=256

GPT-5, o3 and o4-mini

python -m critpt generate epochs=5 model=openai/gpt-5,openai/o3,openai/o4-mini reasoning_effort=high max_connections=150

GPT-5 (minimal)

python -m critpt generate epochs=5 model=openai/gpt-5 reasoning_effort=minimal model_label=nothinking

Non-reasoning models

python -m critpt generate epochs=5 model=together/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8,openai/gpt-4o,google/gemini-2.5-flash

Tool Use (GPT-5 Only)

python -m critpt generate epochs=5 model=openai/gpt-5 reasoning_effort=high model_label=python+search max_connections=150 task_config=.config/multiturn_without_answer_with_tool_web.json

python -m critpt generate epochs=5 model=openai/gpt-5 reasoning_effort=high model_label=python max_connections=150 task_config=.config/multiturn_without_answer_with_tool.json

2. Grading

Model responses should be submitted to our grading server to receive a score. The server is operated in partnership with Artificial Analysis (more API details here).

To ensure data integrity and prevent leakage, our grading server only accepts complete batches containing responses to all 70 problems. Any submission with missing or unmatched problem IDs will be automatically rejected. Each account is limited to 10 submissions within a 24-hour window.

To initiate the grading process, run the following command:

python ../evaluate_all_results.py

Contact

For inquiries or collbarations, please contact minhui.zhu@anl.gov, mtian8@illinois.edu.

Citation

@article{zhu2025probing,
  title={Probing the Critical Point (CritPt) of AI Reasoning: a Frontier Physics Research Benchmark},
  author={Zhu, Minhui and Tian, Minyang and Yang, Xiaocheng and Zhou, Tianci and Zhu, Penghao and Chertkov, Eli and Liu, Shengyan and Du, Yufeng and Yuan, Lifan and Ji, Ziming and others},
  journal={arXiv preprint arXiv:2509.26574},
  year={2025}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published