Все публикации

Tomasz Steifer: Ehrenfeucht-Haussler rank and chain of thought

Kartik Ahuja: On Provable Length and Compositional Generalization

Lekai Chen: LLMs as Probabilistic Minimally Adequate Teachers for DFA Learning

Robert Csordas

Eran Malach: Universal Length Generalization with Turing Programs

Dan Friedman: Representing Rule-based Chatbots with Transformers

Keyon Vafa: Evaluating the World Model Implicit in a Generative Model

Chris Köcher: Hard Attention Transformers on Data Sequences: A Formal Language Theoretic Perspective

Anej Svete: Transformers Can Represent n-gram Language Models

Yash Sarrof: The Expressive Capacity of State Space Models: A Formal Language Perspective

Anton Xue: Logicbreaks: A Framework for Understanding Subversion of Rule-based Inference

Zhiyuan Li: Chain Of Thought Empowers Transformers To Solve Inherently Serial Problems

Yingshan Chang: Language Models Need Inductive Biases to Count Inductively

Alessandro Ronca: On the Expressivity of Recurrent Neural Cascades

Martin Berger: Fast grammar inference on GPUs

Daniel Hsu: Transformers, parallel computation and logarithmic depth

Michaël Rizvi: Simulating Weighted Automata over Sequences and Trees with Transformers

Mark Rofin: Why are Sensitive Functions Hard for Transformers?

Brian DuSell: Stack Attention

Will Merrill: The Illusion of State in State-Space Models

Nur Lan: Bridging the Empirical-Theoretical Gap in Neural Network Formal Language Learning

Dylan Zhang: Transformer-Based Models Are Not Yet Perfect At Learning 2 Emulate Structural Recursion

Giuseppe De Giacomo

Alexander Kozachinskiy: Logical Languages Accepted by Transformer Encoders with Hard Attention