Poster
in
Workshop: Bridging the Gap Between Practice and Theory in Deep Learning
Anchor Function for Studying Language Models
Zhongwang Zhang · Zhiwei Wang · Zhiqin Xu
Understanding transformer-based language models is becoming increasingly crucial. However, language model research faces significant challenges, including complex data structures, unknown target functions, high computational costs and memory requirements, and a lack of interpretability in the inference process, etc. Drawing a parallel to the use of simple models in scientific research, we propose the concept of an anchor function. This is a type of benchmark function designed for studying language models in learning tasks that follow an "anchor-key" pattern. By utilizing the concept of an anchor function, we can construct a series of functions to simulate various language tasks. We demonstrate the utility of the anchor function with an example, revealing two basic operations by attention structures in language models: shifting tokens and broadcasting one token from one position to many positions. These operations are also commonly observed in large language models.