The Counterexample Game: Iterated Conceptual Analysis and Repair in Language Models
Daniel Drucker, Kyle Mahowald
- Published
- May 5, 2026 — 16:26 UTC
- Summary length
- 468 words
- Relevance score
- 70%
Problem
This preprint addresses the gap in understanding the capability of language models (LMs) to perform conceptual analysis, a task traditionally associated with philosophical methodology. Specifically, it investigates whether LMs can iteratively refine definitions through the generation of counterexamples and subsequent repairs. The authors aim to explore the effectiveness of LMs in engaging in high-level philosophical reasoning, a domain where their performance has not been thoroughly evaluated.
Method
The authors propose a framework termed the “Counterexample Game,” which consists of an iterative process involving two distinct LM instances. The first instance generates counterexamples to a proposed definition of a concept, while the second instance attempts to repair the definition based on the generated counterexamples. This process is repeated across 20 different concepts, with thousands of counterexample-repair cycles conducted. The evaluation of counterexample validity is performed by both expert human judges and an LM judge, allowing for a comparative analysis of judgment consistency. The study does not disclose specific architectural details, loss functions, or training compute used for the LMs, focusing instead on the iterative interaction between the models.
Results
The findings reveal that while many counterexamples generated by the LMs are deemed invalid by both human experts and the LM judge, the LM judge accepts approximately twice as many counterexamples as human judges. The consistency of validity judgments is moderate, both among human judges and between humans and the LM. Notably, the authors observe that extended iterations lead to increasingly verbose definitions without a corresponding improvement in accuracy. Additionally, certain concepts demonstrate a resistance to stable definitions, indicating inherent challenges in the conceptual analysis process. These results suggest that while LMs can engage in philosophical reasoning, the effectiveness of the counterexample-repair loop diminishes rapidly.
Limitations
The authors acknowledge that the counterexample-repair process yields diminishing returns, which may limit the practical utility of this approach for sustained philosophical reasoning. They also note that some concepts are particularly resistant to stable definitions, which could indicate fundamental limitations in the LMs’ ability to grasp complex philosophical constructs. An additional limitation not explicitly mentioned by the authors is the potential bias in the selection of concepts and the subjective nature of validity judgments, which may affect the generalizability of the findings.
Why it matters
This work has significant implications for the evaluation of LMs in high-level reasoning tasks, particularly in the context of philosophical inquiry. By demonstrating that LMs can engage in iterated conceptual analysis, the study opens avenues for further research into the capabilities and limitations of LMs in reasoning tasks that require nuanced understanding and definition refinement. The findings also suggest that while LMs can contribute to philosophical discourse, their current limitations necessitate careful consideration of their role in such contexts, potentially guiding future developments in LM architectures and training methodologies.
Authors: Daniel Drucker, Kyle Mahowald
Source: arXiv:2605.03936
URL: https://arxiv.org/abs/2605.03936v1