Latent Hypothesis-Planned Exploration
Abstract
Although large language models (LLMs) have been observed to perform at a human level in theory of mind tasks, deeper examinations and systematic testing of their performance in these domains is needed. Primitive decomposition representations show promise for building robotic systems with greater abilities for in-depth natural language understanding and generation. In this work, we explore representations of theory of mind which are combinations of conceptual primitives, focusing on simulations of a Sally-Anne false-belief test. We demonstrate how primitive decompositions into the conceptual building blocks of image schemas and Conceptual Dependency can represent the attribution of false beliefs to intelligent agents. The exploration has consequences for generating controlled and linguistically varied tests posed in natural language as challenge problems for large language models and for cognitive representations more broadly.