Mind the (DH) Gap! A Contrast in Risky Choices Between Reasoning and Conversational LLMs
Research on 20 frontier LLMs reveals a major 'description-history gap' in risky decision-making.
Researchers Luise Ge, Yongyan Zhang, and Yevgeniy Vorobeychik published 'Mind the (DH) Gap!', a study comparing 20 frontier and open LLMs on risky choices. They found models cluster into two categories: Reasoning Models (RMs) trained for math behave rationally, while Conversational Models (CMs) show a large 'description-history gap' and are more human-like but less rational. This impacts reliability when using LLMs as decision support systems or in agentic workflows.
Why It Matters
Choosing the wrong model type for high-stakes decisions could lead to irrational, risky outcomes in finance, healthcare, or autonomous systems.