Explore the Reasoning Capability of LLMs in the Chess Testbed

1University of California, Los Angeles, 2Microsoft Research,
3MBZUAI, 4University of Toronto, 5Peking University
NAACL 2025 Main Conference
Figure 1: Strategy and Tactic (a)White E2 pawn moves to E4, takes more space in the center, and exerts pressure on black. Black will have a hard time struggling to develop its pieces. (b)White E2 bishop moves to F3 and pins the knight on C6. The black knight cannot move, or the A8 rook behind the knight will be taken. White will take black knight for free in the next move.

Abstract

Reasoning is a central capability of human intelligence. In recent years, with the advent of large-scale datasets, pretrained large language models have emerged with new capabilities, including reasoning. However, these models still struggle with long-term, complex reasoning tasks, such as playing chess. Based on the observation that expert chess players employ a dual approach combining long-term strategic play with short-term tactical play along with language explanation, we propose improving the reasoning capability of large language models in chess by integrating annotated strategy and tactic. Specifically, we collect a dataset named MATE, which consists of 1 million chess positions with candidate moves annotated by chess experts for strategy and tactics. We finetune the LLaMA-3-8B model and compare it against state-of-the-art commercial language models in the task of selecting better chess moves. Our experiments show that our models perform better than GPT, Claude, and Gemini models. We find that language explanations can enhance the reasoning capability of large language models.

Dataset Overview

Dataset Example
Figure 2: A data example in MATE-Strategy&Tactic.
Dataset Summary
Figure 3: Dataset Summary (a)Distribution of samples across the MATE subsets. (b)Distribution of strategy in the MATE. (c)Distribution of tactic in the MATE.

Experimental Results

Case Study

BibTeX

@article@article{wang2024explore,
        title={Explore the Reasoning Capability of LLMs in the Chess Testbed},
        author={Wang, Shu and Ji, Lei and Wang, Renxi and Zhao, Wenxiao and Liu, Haokun and Hou, Yifan and Wu, Ying Nian},
        journal={arXiv preprint arXiv:2411.06655},
        year={2024}
      }