Reference Without Referents

A philosophical analysis arguing that LLMs demonstrate language has autonomous self-generative capacity without requiring reference to external objects or speaker intentions. Meaning emerges from distributional structure, not from words "pointing to" things in the world.

In preparation for Philosophical Review

The Autoregressive Brain

A comprehensive theoretical paper reconceptualizing memory as generative potential rather than storage-retrieval. Integrates evidence from psychology, neuroscience, and machine learning to argue that cognition operates through sequential state generation.

In preparation for Trends in Cognitive Sciences

Morphosyntactic Constraints in Language Models

An empirical test of whether function words and morphology constrain predictions independently of semantic content. Results show morphosyntax reduces next-token entropy by ~1 bit even with nonsense content words --- supporting the distributional account of syntax.

ArXiv preprint forthcoming

World Properties Without World Models

Demonstrates that static word embeddings (GloVe, Word2Vec) encode real-world geographic and climate structure. Ridge regression probes on city-name embeddings predict latitude (R² = 0.72), longitude (R² = 0.63), and temperature (R² = 0.64) from distributional information alone.

Paper submitted — Code on GitHub

Autogenerative Learning

A suite of experiments testing whether transformers can learn to continue sequences they themselves have generated --- probing the boundary between learned patterns and genuinely generative behavior.

Code on GitHub

Long-Range Token Influence Analysis

A causal intervention framework for measuring how far back in context language models actually use information. Probes the mechanisms of memory-like behavior in transformers by selectively ablating tokens at varying distances.

Code on GitHub

## Research Philosophy All of this work flows from a single question: **if cognition is autoregressive generation over learned distributional structure, what should we expect to find?** Each project tests a different facet of that question. The morphosyntax work tests whether syntactic structure is distributional. The embedding geography work tests whether distributional learning captures world structure. The autogenerative learning work tests the boundaries of what sequential prediction can learn. The philosophical work addresses what the success of LLMs means for theories of meaning and reference. The goal is not to prove that "the brain is an LLM." The goal is to take seriously what autoregressive prediction can accomplish, and to ask whether biological cognition might operate on the same principle --- even if the implementation details differ enormously.