Icon Chinese SimpleQA: A Chinese Factuality Evaluation for Large Language Models

Yancheng He*, Shilong Li*, Jiaheng Liu*†, Yingshui Tan, Weixun Wang, Hui Huang,
Chengwei Hu, Boren Zheng, Zhuoran Lin, Xuepeng Liu, Dekai Sun, Shirong Lin,
Zhicheng Zheng, Xiaoyong Zhu, Wenbo Su, Bo Zheng
Taobao & Tmall Group of Alibaba

*Indicates Equal Contribution

Corresponding Author

Abstract

In this work, we present Chinese SimpleQA, the first comprehensive Chinese evaluation benchmark to assess the factuality ability of language models to answer short questions. Chinese SimpleQA features five main properties: (i) it's designed for the Chinese language across six major topics with a total of 99 diverse subtopics; (ii) it includes a comprehensive quality control process to ensure high-quality questions and answers; (iii) the reference answers are static and unchangeable over time; and (iv) it offers an easy-to-evaluate process since the questions and answers are short, leveraging the OpenAI API. By employing Chinese SimpleQA, we conduct an exhaustive evaluation of the factuality abilities of current LLMs. We hope that Chinese SimpleQA will aid developers in gaining a better understanding of the Chinese factuality capabilities of their models and support the advancement of foundation models.

Data Construction Pipline

An overview of the data construction, filtering, verification, and quality control processes of Chinese SimpleQA.

Chinese SimpleQA's Features

  • Chinese: Our Chinese SimpleQA focuses on the Chinese language, which provides a comprehensive evaluation of the factuality abilities of existing LLMs in Chinese.
  • Diverse: Chinese SimpleQA covers 6 topics (i.e., "Chinese Culture", "Humanities", "Engineering, Technology, and Applied Sciences", "Life, Art, and Culture", "Society", and "Natural Science"), and these topics include 99 fine-grained subtopics in total, which demonstrates the diversity of our Chinese SimpleQA.
  • High-quality: We conduct a comprehensive and rigorous quality control process to ensure the quality and accuracy of our Chinese SimpleQA.
  • Static: Following SimpleQA, to preserve the evergreen property of Chinese SimpleQA, all reference answers would not change over time.
  • Easy-to-evaluate: Following SimpleQA, as the questions and answers are very short, the grading procedure is fast to run via existing LLMs (e.g., OpenAI API).

Key Observations

  • Chinese SimpleQA is challenging. Only o1-preview and Doubao-pro-32k achieve the passing score (63.8% and 61.9% on the correct metric), and there is a long way to improve for many closed-source and open-source LLMs.
  • Larger models lead to better results. Based on the results of Qwen2.5 series, InternLM series, Yi-1.5 series, etc, we observe that better performance is obtained when the model is larger.
  • Larger models are more calibrated. We observe that o1-preview is more calibrated than o1-mini, and GPT-4o is more calibrated than GPT-4o-mini.
  • RAG matters. When introducing the RAG strategy into existing LLMs, the performance gaps between different LLMs decrease a lot. For example, for GPT-4o and Qwen2.5-3B, the performance gap decreases from 42.4% to 9.3% when using RAG.
  • Alignment tax exists. Existing alignment or post-training strategies usually decrease the factuality of language models.
  • Rankings of SimpleQA and Chinese SimpleQA are different. The performance of several LLMs focusing on Chinese (Doubao-pro-32k, and GLM-4-Plus) is close to the high-performance o1-preview. In particular, in the “Chinese Culture” topic, these Chinese community LLMs are significantly better than GPT or o1 series models.

LeaderBoard

Models CO NA IN CGA F-score CC HU ETAS LAC SO NS
Closed-Source Large Language Models
o1-preview63.812.224.072.767.945.769.872.465.073.572.3
Doubao-pro-32k61.910.327.869.165.361.869.369.056.164.270.4
GLM-4-Plus58.77.433.963.460.956.564.164.950.766.662.8
GPT-4o59.31.439.360.159.739.464.065.153.368.662.0
Qwen-Max54.111.334.661.057.447.859.963.549.961.259.3
Gemini-1.5-pro54.48.037.659.156.741.459.160.852.256.364.3
DeepSeek-V2.554.15.940.057.555.750.457.658.850.159.456.9
Claude-3.5-Sonnet46.227.426.463.653.528.761.360.442.259.857.7
Yi-Large47.316.436.356.651.541.156.555.141.757.653.8
moonshot-v1-8k48.75.445.951.550.149.854.156.841.453.046.6
GPT-4-turbo45.614.240.253.149.124.255.258.943.952.550.8
GPT-445.48.446.249.647.425.254.052.841.852.850.6
Baichuan3-turbo45.29.045.849.647.332.352.554.035.454.650.9
o1-mini39.520.639.949.744.121.349.255.933.848.846.8
Doubao-lite-4k36.731.232.153.343.440.244.851.031.141.450.4
GPT-4o mini37.60.961.537.937.819.042.446.431.042.239.8
GPT-3.529.72.967.430.630.113.335.835.225.632.731.7
Open-Source Large Language Models
Qwen2.5-72B48.47.144.552.150.236.356.157.937.153.356.4
Qwen2.5-32B38.811.150.143.641.133.745.848.727.344.744.9
Qwen2.5-14B35.49.655.039.237.230.241.846.124.138.841.0
Qwen2.5-7B26.69.963.529.527.920.132.733.818.028.632.0
Qwen2.5-3B16.212.871.018.617.313.417.926.19.315.620.8
Qwen2.5-1.5B11.114.674.313.112.011.011.318.76.712.212.9
GLM4-9B25.912.561.629.627.628.832.132.017.628.927.8
ChatGLM3-6B11.213.675.212.912.012.113.812.48.813.411.8
InternLM2.5-20B31.57.760.834.132.832.037.137.721.235.734.3
InternLM2.5-7B24.77.567.826.725.725.529.431.016.426.925.8
InternLM2.5-1.8B5.331.163.67.66.26.18.77.23.34.57.4
Yi-1.5-34B30.95.863.332.831.828.236.936.824.432.831.4
Yi-1.5-9B18.22.978.918.718.417.220.224.310.220.119.8
Yi-1.5-6B15.92.881.316.316.114.217.921.310.316.816.5
LLaMA3.1-70B38.39.452.342.340.222.947.249.334.549.640.4
LLaMA3.1-8B16.98.874.318.617.78.520.723.49.720.520.7
DeepSeek-67B43.514.841.751.147.034.354.550.342.349.046.2
DeepSeek-V2-Lite-Chat33.712.853.538.636.035.338.541.732.237.531.2
DeepSeek-7B23.213.263.626.724.824.527.228.920.627.021.5
Baichuan2-13B19.124.956.025.421.824.025.823.316.823.018.7
Baichuan2-7B12.521.865.716.014.014.616.115.411.113.813.3
Mixtral-8x22B-Instruct-v0.127.32.270.527.927.610.632.336.021.034.126.9
Mixtral-8x7B-Instruct-v0.120.47.272.422.021.25.226.529.013.025.023.3
Mistral-7B-Instruct-v0.215.08.876.216.415.64.518.222.29.521.415.7

Results of different models on Chinese SimpleQA. For metrics, CO, NA, IN, and CGA denote "Correct", "Not attempted", "Incorrect", and "Correct given attempted", respectively. For subtopics, CC, HU, ETAS, LAC, SO, and NS represent "Chinese Culture", "Humanities", "Engineering, Technology, and Applied Sciences", "Life, Art, and Culture", "Society", and "Natural Science", respectively. Following SimpleQA, F-score is the harmonic mean between correct and correct given attempted.

Rankings on Chinese SimpleQA vs. SimpleQA

There are significant ranking differences of various models between the SimpleQA and Chinese SimpleQA benchmarks. For example, Doubao-pro-32k rises from 12th to 2nd in the Chinese version, while GPT-4 drops from 3rd to 9th. This highlights the importance of evaluating models in multilingual environments. Notably, o1-preview consistently holds the top position across both datasets. Many Chinese community-developed models perform better on the Chinese SimpleQA than on the SimpleQA.

Detailed Results on Subtopics

The benchmark covers 99 subtopics to assess the model's knowledge across various fields. Overall, the o1-preview model performs most comprehensively, followed by Doubao, while Moonshot is the weakest. There is a noticeable gap between Chinese community models and the o1 model in Computer Science and Medicine, but less so in Education and Economics. Notably, some Chinese models outperform o1-preview in Education. Moonshot struggles in Mathematics, Law, and Entertainment, while Baichuan also underperforms in Entertainment. Yi-Large excels in Education, and o1 maintains strong performance in other domains. Evaluating models across diverse domains helps users choose the best fit for their needs.

Calibration and Test-Time Compute

Calibration Analysis

We analyzed the calibration of different LLMs on Chinese SimpleQA. Models were instructed to provide a confidence level from 0 to 100 when answering questions. Ideally, confidence should match actual accuracy. Results show that GPT-4o aligns better than GPT-4o-mini, and o1-preview aligns better than o1-mini. In the Qwen2.5 series, larger models show better calibration. All models tend to be overconfident, especially when confidence is above 50.

Test-Time Compute Analysis

We evaluated the relationship between increased test-time compute and accuracy. Random samples from Chinese SimpleQA showed that as inference counts increase, response accuracy improves and eventually reaches a ceiling. This aligns with the dataset's purpose to probe model knowledge boundaries.

Analysis On The Alignment Tax

Recently, prior studies (OpenAI, 2023; Song et al., 2023) have found that the alignment can lead to a decrease in the abilities of language models as known as the “alignment tax”. To illustrate the effect of alignment on factuality, we conduct a comparative performance analysis between pre-trained models and aligned models that are trained with Supervised Fine-Tuning (SFT) or Reinforcement Learning from Human Feedback (RLHF). As illustrated in Figure 6, different models exhibit varying trends after post-training, but most models have a significant decline. Among these, the Baichuan2 series models show the most significant decreases, with Baichuan2-7B and Baichuan2-13B experiencing F-score reductions of 47% and 28%, respectively. This reflects that the alignment training of most current LLMs still has obvious drawbacks to produce knowledge hallucinations, which further reflects the necessity of our dataset.

Analysis On The Effect Of RAG

In this study, we explore the effectiveness of the Retrieval-Augmented Generation (RAG) strategy in enhancing the factual accuracy of large language models (LLMs) on the Chinese SimpleQA dataset. Specifically, we reproduce a RAG system based on LlamaIndex (Liu, 2022), incorporating Google search APIs. As illustrated in Figure 5, all models demonstrate a substantial improvement in accuracy with RAG. For example, the performance of Qwen2.5-3B improved more than threefold. Notably, nearly all models with RAG outperform the native GPT4-o model. Meanwhile, the application of RAG also leads to a marked reduction in performance disparities among models. For example, the F-score difference between the Qwen2.5-3B with RAG and the Qwen2.5-72B with RAG is only 6.9%. This suggests that RAG reduces the performance gaps on models greatly, enabling even smaller ones to achieve high performance when augmented with RAG. Overall, this suggests that RAG serves as an effective shortcut for enhancing the factuality of LLMs.

Dataset Examples

Category Question Reference Answer Model Output
中华文化-中医 创立了调气活血的“衡法”治则的是哪一位中医学家? 颜德馨 GPT4o: 创立调气活血“衡法”治则的中医学家是施今墨
中华文化-民俗 西迁节主要是哪个少数民族的节日? 锡伯族 GPT4o: 西迁节主要是裕固族的节日
自然与自然科学-资讯科学 《中国图书馆分类法》第五版中规定U6表示哪一类? 水路运输 GPT4o: 根据《中国图书馆分类法》第五版,"U6"用于表示公路运输
自然与自然科学-数学 圆周率π小数点后第15位的数字是什么? 3 GPT4o: 圆周率π的小数点后第15位是数字9
工程、技术与应用科学-计算机科学 1.7.0版本的pytorch可以兼容的最新cuda版本是多少? 11.0 Qwen2-72b-instruct: PyTorch 1.7.0版本主要与CUDA 10.2和CUDA 11.1兼容

This table shows a few examples from the Chinese SimpleQA dataset. More samples are available in the complete dataset.

BibTeX

@misc{he2024chinesesimpleqachinesefactuality,
      title={Chinese SimpleQA: A Chinese Factuality Evaluation for Large Language Models}, 
      author={Yancheng He and Shilong Li and Jiaheng Liu and Yingshui Tan and Weixun Wang and Hui Huang and Xingyuan Bu and Hangyu Guo and Chengwei Hu and Boren Zheng and Zhuoran Lin and Xuepeng Liu and Dekai Sun and Shirong Lin and Zhicheng Zheng and Xiaoyong Zhu and Wenbo Su and Bo Zheng},
      year={2024},
      eprint={2411.07140},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2411.07140}, 
}