ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.10542
48
0

Language Models, Graph Searching, and Supervision Adulteration: When More Supervision is Less and How to Make More More

13 March 2025
Arvid Frydenlund
    LRM
ArXivPDFHTML
Abstract

This work concerns the path-star task, a minimal example of searching over a graph. The graph, GGG, is star-shaped with DDD arms radiating from a start node, sss. A language model (LM) is given GGG, sss, and a target node ttt, which ends one of the arms and is tasked with generating the arm containing ttt. The minimal nature of this task means only a single choice needs to be made: which of the DDD arms contains ttt?Decoder-only LMs fail to solve this elementary task above 1/D1/D1/D chance due to a learned shortcut that absorbs training supervision. We show how this pathology is caused by excess supervision and we present a series of solutions demonstrating that the task is solvable via decoder-only LMs. We find that the task's minimal nature causes its difficulty, as it prevents task decomposition. Our solutions provide insight into the pathology and its implications for LMs trained via next-token prediction.

View on arXiv
@article{frydenlund2025_2503.10542,
  title={ Language Models, Graph Searching, and Supervision Adulteration: When More Supervision is Less and How to Make More More },
  author={ Arvid Frydenlund },
  journal={arXiv preprint arXiv:2503.10542},
  year={ 2025 }
}
Comments on this paper