
Why do modern language models, trained to do well on next-word prediction, appear to generate coherent documents and capture long-range structure? Here we show that next-token prediction is provably powerful for learning longer-range structure, even with common neural network architectures. Specifically, we prove that optimizing next-token prediction over a Recurrent Neural Network (RNN) yields a model that closely approximates the training distribution: for held-out documents sampled from the training distribution, no algorithm of bounded description length limited to examining the next tokens, for any , can distinguish between consecutive tokens of such documents and tokens generated by the learned language model following the same prefix. We provide polynomial bounds (in , independent of the document length) on the model size needed to achieve such -token indistinguishability, offering a complexity-theoretic explanation for the long-range coherence observed in practice.
View on arXiv