Visualizing LLM Latent Space Geometry Through Dimensionality Reduction
Alex Ning · Vainateya Rangaraju · Yen-Ling Kuo
Abstract
In this blog post, we extract, process, and visualize latent state geometries in Transformer-based language models through dimensionality reduction to build a better intuition of their internal dynamics. We demonstrate experiments with GPT-2 and LLaMa models, uncovering interesting geometric patterns in their latent spaces. Notably, we identify a clear separation between attention and MLP component outputs across intermediate layers, a pattern not documented in prior work to our knowledge.
Successful Page Load