Links
Figure: Illustration of the Representations of Hierarchically Related Concepts As Direct Sums
Figure: Representations of Hierarchical and Categorical Concepts in the Gemma-2B LLM
Abstract
Understanding how semantic meaning is encoded in the representation spaces of large language models is a fundamental problem in interpretability. In this paper, we study the two foundational questions in this area. First, how are categorical concepts, such as {‘mammal’, ‘bird’, ‘reptile’, ‘fish’}, represented? Second, how are hierarchical relations between concepts encoded? For example, how is the fact that ‘dog’ is a kind of ‘mammal’ encoded? We show how to extend the linear representation hypothesis to answer these questions. We find a remarkably simple structure: simple categorical concepts are represented as simplices, hierarchically related concepts are orthogonal in a sense we make precise, and (in consequence) complex concepts are represented as polytopes constructed from direct sums of simplices, reflecting the hierarchical structure. We validate these theoretical results on the Gemma large language model, estimating representations for 957 hierarchically related concepts using data from WordNet.
Note
Best Paper Award, ICML 2024 Workshop on Mechanistic Interpretability
Citation
Park, K., Choe, Y. J., Jiang, Y., & Veitch, V. (2024). The Geometry of Categorical and Hierarchical Concepts in Large Language Models. arXiv preprint arXiv:2406.01506.
@article{park2024geometry,
title={The Geometry of Categorical and Hierarchical Concepts in Large Language Models},
author={Park, Kiho and Choe, Yo Joong and Jiang, Yibo and Veitch, Victor},
journal={arXiv preprint arXiv:2406.01506},
year={2024}
}