Research
Current projects testing and extending the autoregressive framework
Reference Without Referents
A philosophical analysis arguing that LLMs demonstrate language has autonomous self-generative capacity without requiring reference to external objects or speaker intentions. Meaning emerges from distributional structure, not from words "pointing to" things in the world.
In preparation for Philosophical Review
The Autoregressive Brain
A comprehensive theoretical paper reconceptualizing memory as generative potential rather than storage-retrieval. Integrates evidence from psychology, neuroscience, and machine learning to argue that cognition operates through sequential state generation.
In preparation for Trends in Cognitive Sciences
Morphosyntactic Constraints in Language Models
An empirical test of whether function words and morphology constrain predictions independently of semantic content. Results show morphosyntax reduces next-token entropy by ~1 bit even with nonsense content words --- supporting the distributional account of syntax.
ArXiv preprint forthcoming
World Properties Without World Models
Demonstrates that static word embeddings (GloVe, Word2Vec) encode real-world geographic and climate structure. Ridge regression probes on city-name embeddings predict latitude (R² = 0.72), longitude (R² = 0.63), and temperature (R² = 0.64) from distributional information alone.
Paper submitted — Code on GitHub
Autogenerative Learning
A suite of experiments testing whether transformers can learn to continue sequences they themselves have generated --- probing the boundary between learned patterns and genuinely generative behavior.
Long-Range Token Influence Analysis
A causal intervention framework for measuring how far back in context language models actually use information. Probes the mechanisms of memory-like behavior in transformers by selectively ablating tokens at varying distances.