Toni Liu

Research projects from physics to learning systems

Jacobian Scopes

When an LLM produces an answer, which parts of the input prompt informed that decision?

JacobianScope is a way to compute how much perturbing each input token affects a given target token. In practice, we project each Jacobian along a probe vector of interest to calculate token-level influence scores. This is computationally efficient and yields really interesting results for causal attribution.

Jacobian Scope visualization: per-token influence on predicting liberal versus conservative from the same prompt.
Semantic Scope reveals the potential for political bias that may be present in models such as LLaMA3.2. The prediction for the subject being a “liberal” is attributed to the input token “Columbia”, while “conservative” to the token pairs for “the South”.

In-context learning trajectories of LLMs

How do large language models learn from in-context data?

We leverage the Intensive Principal Component Analysis to visualize how LLMs estimate the probability density underlying input data. Our geometric investigation leads to a natural interpretation of LLM's inference process as a kernel density estimator with adaptive kernel width and shape.

In-context DE trajectory of LLaMA 2-70b, compared to Gaussian KDE and Bayesian histogram.

Diffusion RNN: a model of memory and imagination

What makes abstract thoughts different from concrete memories? Considering memories as fixed points within the brain's dynamical system, we propose to understand abstraction as a hierarchy of slow manifolds, which serve as coarse-grained roadmaps to these fixed points.

We construct a diffusion RNN (Recurrent Neural Network), a neural-dynamical model that can perform memorization and abstraction, and demonstrate its ability to perform dimensionality reduction on the MNIST dataset.

Finally, we present an analytic theory that elucidates the relation between the distribution of data and the learned slow manifolds.

Diffusion RNN extracts a hierarchy of slow manifolds (red) from dimensionally reduced MNIST data.

In-context neural scaling law

Could a pretrained LLM make sense of an unseen, synthetic language, such as, a chain of symbols generated by a random Markov process?

Yes. In fact, we show that LLaMA 2 models possess uncanny abilities to in-context learn a variety of stochastic systems, physical and symbolic. Our observation reveals an in-context version of neural scaling law.

EMNLP Oral Recording

Learning in hyperbolic space

In Euclidean space, the volume of a ball grows polynomially with its radius — but in hyperbolic space, it grows exponentially. This mirrors a fundamental property of trees: the number of nodes at depth d grows exponentially with d. This geometric advantage makes hyperbolic space a natural host for embedding hierarchical data.

We embed hierarchical data by casting shadows: a light source projects opaque objects onto a plane, where parent-child relationships correspond to shadow containment — a child node is represented by an object whose shadow falls entirely within its parent's.

This geometric picture yields a fast embedding algorithm: to embed a tree, one simply places objects such that the containment constraints are satisfied, with deeper nodes casting smaller, nested shadows.

Anticipatory waves in a gigantic marine algae

How do plants anticipate external environemnts and regulate internal states?

We perturb marine algae with patterned illumination, and analyze the response dynamics using image analysis and reduced-dimension observables. Our experiment and analysis unveils wave-like patterns entrained by external illuminations, and coupled to a circadian internal oscillator.