- Aaron Daniel Cohen
- Adam Roberts
- Alejandra Molina
- Alena Butryna
- Alicia Jin
- Apoorv Kulshreshtha
- Ben Hutchinson
- Ben Zevenbergen
- Blaise Hilary Aguera-Arcas
- Chung-ching Chang
- Claire Cui
- Cosmo Du
- Daniel De Freitas Adiwardana
- Dehao Chen
- Dmitry (Dima) Lepikhin
- Ed H. Chi
- Erin Hoffman-John
- Heng-Tze Cheng
- Hongrae Lee
- Igor Krivokon
- James Qin
- Jamie Hall
- Joe Fenton
- Johnny Soraker
- Kathy Meier-Hellstern
- Kristen Olson
- Lora Mois Aroyo
- Maarten Paul Bosma
- Marc Joseph Pickett
- Marcelo Amorim Menegali
- Marian Croak
- Mark Díaz
- Matthew Lamm
- Maxim Krikun
- Meredith Ringel Morris
- Noam Shazeer
- Quoc V. Le
- Rachel Bernstein
- Ravi Rajakumar
- Ray Kurzweil
- Romal Thoppilan
- Steven Zheng
- Taylor Bos
- Toju Duke
- Tulsee Doshi
- Vincent Y. Zhao
- Vinodkumar Prabhakaran
- Will Rusch
- YaGuang Li
- Yanping Huang
- Yanqi Zhou
- Yuanzhong Xu
- Zhifeng Chen
Abstract
We present LaMDA: Language Models for Dialog Applications. LaMDA is a family of Transformer-based neural language models specialized for dialog, which have up to 137B parameters and arepre-trained on 1.56T words of public dialog data and web text. While model scaling alone canimprove quality, it shows less improvements on safety and factual grounding. We demonstrate thatfine-tuning with annotated data and enabling the model to consult external knowledge sources canlead to significant improvements towards the two key challenges of safety and factual grounding.The first challenge, safety, involves ensuring that the model’s responses are consistent with a set ofhuman values, such as preventing harmful suggestions and unfair bias. We quantify safety using ametric based on an illustrative set of values, and we find that filtering candidate responses using aLaMDA classifier fine-tuned with a small amount of crowdworker-annotated data offers a promisingapproach to improving model safety. The second challenge, factual grounding, involves enabling themodel to consult external knowledge sources, such as an information retrieval system, a languagetranslator, and a calculator. We quantify factuality using a groundedness metric, and we find that ourapproach enables the model to generate responses grounded in known sources, rather than responsesthat merely sound plausible. Finally, we explore the use of LaMDA in the domains of education andcontent recommendations, and analyze their helpfulness and role consistency.
Research Areas
Learn more about how we do research
We maintain a portfolio of research projects, providing individuals and teams the freedom to emphasize specific types of work