ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2205.11558
14
30

Using Natural Language and Program Abstractions to Instill Human Inductive Biases in Machines

23 May 2022
Sreejan Kumar
Carlos G. Correa
Ishita Dasgupta
Raja Marjieh
Michael Y. Hu
Robert D. Hawkins
Nathaniel D. Daw
Jonathan D. Cohen
Karthik Narasimhan
Thomas L. Griffiths
    AI4CE
ArXivPDFHTML
Abstract

Strong inductive biases give humans the ability to quickly learn to perform a variety of tasks. Although meta-learning is a method to endow neural networks with useful inductive biases, agents trained by meta-learning may sometimes acquire very different strategies from humans. We show that co-training these agents on predicting representations from natural language task descriptions and programs induced to generate such tasks guides them toward more human-like inductive biases. Human-generated language descriptions and program induction models that add new learned primitives both contain abstract concepts that can compress description length. Co-training on these representations result in more human-like behavior in downstream meta-reinforcement learning agents than less abstract controls (synthetic language descriptions, program induction without learned primitives), suggesting that the abstraction supported by these representations is key.

View on arXiv
Comments on this paper