ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2201.08239
13
1557

LaMDA: Language Models for Dialog Applications

20 January 2022
R. Thoppilan
Daniel De Freitas
Jamie Hall
Noam M. Shazeer
Apoorv Kulshreshtha
Heng-Tze Cheng
Alicia Jin
Taylor Bos
Leslie Baker
Yu Du
Yaguang Li
Hongrae Lee
H. Zheng
Amin Ghafouri
Marcelo Menegali
Yanping Huang
M. Krikun
Dmitry Lepikhin
James Qin
Dehao Chen
Yuanzhong Xu
Zhifeng Chen
Adam Roberts
Maarten Bosma
Vincent Zhao
Yanqi Zhou
Chung-Ching Chang
I. Krivokon
W. Rusch
Marc Pickett
Pranesh Srinivasan
Laichee Man
Kathy Meier-Hellstern
Meredith Ringel Morris
Tulsee Doshi
Renelito Delos Santos
Toju Duke
J. Søraker
Ben Zevenbergen
Vinodkumar Prabhakaran
Mark Díaz
Ben Hutchinson
Kristen Olson
Alejandra Molina
Erin Hoffman-John
Josh Lee
Lora Aroyo
Ravi Rajakumar
Alena Butryna
Matthew Lamm
V. Kuzmina
Joseph Fenton
Aaron Cohen
R. Bernstein
R. Kurzweil
Blaise Aguera-Arcas
Claire Cui
M. Croak
Ed H. Chi
Quoc Le
    ALM
ArXivPDFHTML
Abstract

We present LaMDA: Language Models for Dialog Applications. LaMDA is a family of Transformer-based neural language models specialized for dialog, which have up to 137B parameters and are pre-trained on 1.56T words of public dialog data and web text. While model scaling alone can improve quality, it shows less improvements on safety and factual grounding. We demonstrate that fine-tuning with annotated data and enabling the model to consult external knowledge sources can lead to significant improvements towards the two key challenges of safety and factual grounding. The first challenge, safety, involves ensuring that the model's responses are consistent with a set of human values, such as preventing harmful suggestions and unfair bias. We quantify safety using a metric based on an illustrative set of human values, and we find that filtering candidate responses using a LaMDA classifier fine-tuned with a small amount of crowdworker-annotated data offers a promising approach to improving model safety. The second challenge, factual grounding, involves enabling the model to consult external knowledge sources, such as an information retrieval system, a language translator, and a calculator. We quantify factuality using a groundedness metric, and we find that our approach enables the model to generate responses grounded in known sources, rather than responses that merely sound plausible. Finally, we explore the use of LaMDA in the domains of education and content recommendations, and analyze their helpfulness and role consistency.

View on arXiv
Comments on this paper