✍️
The Daily Ink
  • 🙇‍♂️AI research is getting crazy...
  • Paper Summaries
    • 👨‍⚕️[2/25/23] Anthropic makes AI that teaches itself ethics
    • 🪄[2/22/23] Models can magically learn new skills at scale
    • *️[2/21/23] Discovering a better optimization algorithm with evolution
    • 🔮[2/17/23] Talking to models requires special prompts that help them think sequentially
    • 🏹[2/15/23] Teaching LLMs to use tools and not suck at math
    • ➗[2/13/23] English is just math in prettier clothing
    • 👨‍💻[2/8/23] The secret to good writing is editing
    • 💢[2/6/23] Solving context length constraints by distillation
    • 🔭[2/3/23] A Large Language Model for SCIENCE
  • 🎇[2/1/23] Optimal parallelism in ML training is possible, says ALPA
  • 🎼[1/31/23] Google makes a language model for music
  • 🚒[1/27/23] Google's LaMDA model is too convincing, and a researcher is fired
  • 🤔[1/25/23] Teaching computers to think in abstractions
  • 🎩[1/23/23] The secret sauce behind ChatGPT
  • 📸[1/20/23] FlashAttention challenges ML researchers to think about systems-level improvements
  • ✂️[1/18/23] Make models smarter not larger, with data pruning
  • 🙅[1/16/23] DeepMind attempts to make AI that can do anything
  • 🐣[1/8/23] Chinchilla is the strongest animal/model in the world
  • ⬇️[1/5/23] Gradient descent-ing gradient descent
  • 🥘[1/3/23] Cramming: Training a Language Model on a Single GPU in One Day
  • 🗃️[1/1/23] A Neural Corpus Indexer for Document Retrieval
  • 👋[12/27/22] Can We Teach Machines to Act Like Humans? (And Should We?)
Powered by GitBook
On this page
  • Introduction and Motivation
  • Development Details
  • Evaluation
  • Limitations and Future Work
  • Fun Examples
  • References

[1/27/23] Google's LaMDA model is too convincing, and a researcher is fired

LaMDA: Language Models for Dialog Applications

Previous[1/31/23] Google makes a language model for musicNext[1/25/23] Teaching computers to think in abstractions

Last updated 2 years ago

A while ago, there was a about a Google researcher who got convinced that the LaMDA AI was sentient. It did the rounds for a while, but now let’s spend some time investigating what made the model so convincing in the first place.

Introduction and Motivation

There are a few core problems with open dialog systems (and at large LLMs) right now:

  1. Safety: How can we prevent models from outputing toxic or harmful content?

  2. Factual Grounding: How can we help the model stay up to date with facts (“what’s the weather today?”)

  3. Interestingness: How can we bias our models to produce interesting answers? No boring answers.

It is these problems that LaMDA attempts to solve.

Goal: Create a model that can self-evaluate SSI — sensibleness, specificity, and interestingness — while being safe. It can use a toolbox to figure out answers it does not know.

Development Details

The core contribution of this paper is the realization that fine-tuning on a small set of crowdworker-annotated data offers a promising approach to improving model safety. This data is 0.001% of the pre-training data, but is effective nonetheless.

The model is also given a toolset that includes a information retrieval system, a calculator, and a translator. The model learns when to use the model from fine-tuning as well.

The model can thus cite sources, do math, and swap between languages. What a polyglot!

Evaluation

First, we see that fine-tuning has an obvious improvement in both quality and safety and groundedness.

On a wider set of benchmarks, we see that LaMDA continues to perform impressively.

Limitations and Future Work

  1. I am disappointed that LaMDA’s performance wasn’t compared to other dialog models (ChatGPT), which would have demonstrated how their techniques compare to other tactics like RLHF. However, it is very possible that ChatGPT was released after the paper was written 🤷

  2. A lot of this technique depends on the quality of the crowdworker annotations. Thus, it is important to evaluate the patterns of these annotations, as well as make sure they are diverse and representative of various beliefs.

  3. A model, despite having so many tools, still is not able to execute complex reasoning.

Fun Examples

I’ll end this with some fun LaMDA examples, for the model results are fun to read through!

References

[1] Paper link:

[2]

🚒
https://arxiv.org/abs/2201.08239
https://www.washingtonpost.com/technology/2022/06/11/google-ai-lamda-blake-lemoine/
very interesting story