EXAONE-Deep
HomeModel OverviewTechnical FeaturesPerformanceApplicationsDeploymentTutorials

EXAONE-Deep Performance Evaluation

Performance Overview

EXAONE-Deep models demonstrate exceptional performance across multiple benchmarks, particularly excelling in tasks that require complex logical thinking such as mathematical reasoning, scientific understanding, and code writing. Below is a detailed assessment of the model's performance in various benchmarks, showing that:

Mathematical Capability

In the MATH-500 test, EXAONE-Deep 32B model achieved a score of 95.7, the 7.8B model reached 94.8, and the 2.4B model scored 92.3, all leading in their respective model size categories.

Notably, in the CSAT 2025 Math test, the 32B model scored 94.5, outperforming all similar models.

Scientific Reasoning

In the GPQA Diamond test, which evaluates PhD-level problem-solving capabilities in physics, chemistry, and biology, EXAONE-Deep models achieved excellent scores of 66.1 (32B), 62.6 (7.8B), and 54.3 (2.4B).

The 7.8B and 2.4B models ranked first in their respective size categories, demonstrating exceptional scientific reasoning capabilities.

Code Generation

In the LiveCodeBench test, EXAONE-Deep models achieved scores of 59.5 (32B), 55.2 (7.8B), and 46.6 (2.4B).

Compared to other models of similar size, the EXAONE-Deep series demonstrated leading advantages in code writing and understanding, with the 7.8B model outperforming the OpenAI o1-mini model.

Mathematical Reasoning Performance Comparison

ModelMATH-500
(pass@1)
AIME 2024
(pass@1 / cons@64)
AIME 2025
(pass@1 / cons@64)
CSAT Math 2025
(pass@1)
EXAONE Deep 32B95.772.1 / 90.065.8 / 80.094.5
DeepSeek-R1-Distill-Qwen-32B94.372.6 / 83.355.2 / 73.384.1
QwQ-32B95.579.5 / 86.767.1 / 76.794.4
DeepSeek-R1-Distill-Llama-70B94.570.0 / 86.753.9 / 66.788.8
DeepSeek-R1 (671B)97.379.8 / 86.766.8 / 80.089.9
EXAONE Deep 7.8B94.870.0 / 83.359.6 / 76.789.9
DeepSeek-R1-Distill-Qwen-7B92.855.5 / 83.338.5 / 56.779.7
DeepSeek-R1-Distill-Llama-8B89.150.4 / 80.033.6 / 53.374.1
OpenAI o1-mini9063.6 / 80.054.8 / 66.784.4
EXAONE Deep 2.4B92.352.5 / 76.747.9 / 73.379.2
DeepSeek-R1-Distill-Qwen-1.5B83.928.9 / 52.723.9 / 36.765.6

In terms of mathematical reasoning, the EXAONE-Deep series models demonstrate exceptional performance:

• EXAONE-Deep 32B achieved the highest score of 94.5 in the CSAT 2025 Math test, and leads all models with a score of 90.0 in AIME 2024 cons@64.

• EXAONE-Deep 7.8B scored 94.8 in the MATH-500 test, far exceeding other models of similar size and even outperforming OpenAI o1-mini.

• EXAONE-Deep 2.4B significantly outperforms the DeepSeek-R1-Distill-Qwen-1.5B model in all tests, demonstrating its superiority among lightweight models.


Benchmark Descriptions

MATH-500

A collection of 500 high-quality math problems covering algebra, geometry, combinatorics, probability theory, and other fields, used to evaluate a model's mathematical reasoning ability.

AIME

American Invitational Mathematics Examination, serving as a qualifying standard for the US Mathematical Olympiad, with a high level of difficulty. AIME 2024/2025 represents recent contest evaluations, and cons@64 indicates accuracy with consideration for diverse sampling.

CSAT Math

The mathematics section of the Korean College Scholastic Ability Test, used to assess the model's ability to solve high school-level math problems.

GPQA Diamond

A test that evaluates PhD-level problem-solving capabilities, covering difficult problems in physics, chemistry, and biology.

LiveCodeBench

A real-time code writing evaluation benchmark that tests the model's performance in solving programming problems, writing functional code, and debugging.

MMLU

Massive Multitask Language Understanding, a comprehensive assessment of the model's knowledge mastery across various academic disciplines.