✍️
The Daily Ink
  • 🙇‍♂️AI research is getting crazy...
  • Paper Summaries
    • 👨‍⚕️[2/25/23] Anthropic makes AI that teaches itself ethics
    • 🪄[2/22/23] Models can magically learn new skills at scale
    • *️[2/21/23] Discovering a better optimization algorithm with evolution
    • 🔮[2/17/23] Talking to models requires special prompts that help them think sequentially
    • 🏹[2/15/23] Teaching LLMs to use tools and not suck at math
    • ➗[2/13/23] English is just math in prettier clothing
    • 👨‍💻[2/8/23] The secret to good writing is editing
    • 💢[2/6/23] Solving context length constraints by distillation
    • 🔭[2/3/23] A Large Language Model for SCIENCE
  • 🎇[2/1/23] Optimal parallelism in ML training is possible, says ALPA
  • 🎼[1/31/23] Google makes a language model for music
  • 🚒[1/27/23] Google's LaMDA model is too convincing, and a researcher is fired
  • 🤔[1/25/23] Teaching computers to think in abstractions
  • 🎩[1/23/23] The secret sauce behind ChatGPT
  • 📸[1/20/23] FlashAttention challenges ML researchers to think about systems-level improvements
  • ✂️[1/18/23] Make models smarter not larger, with data pruning
  • 🙅[1/16/23] DeepMind attempts to make AI that can do anything
  • 🐣[1/8/23] Chinchilla is the strongest animal/model in the world
  • ⬇️[1/5/23] Gradient descent-ing gradient descent
  • 🥘[1/3/23] Cramming: Training a Language Model on a Single GPU in One Day
  • 🗃️[1/1/23] A Neural Corpus Indexer for Document Retrieval
  • 👋[12/27/22] Can We Teach Machines to Act Like Humans? (And Should We?)
Powered by GitBook
On this page
  • Motivation and Core Insight
  • Example and Underlying Design
  • Evaluation
  • Limitations and Future Work
  • References

[12/27/22] Can We Teach Machines to Act Like Humans? (And Should We?)

NeurIPS 2022 Outstanding Paper Award, DeepMind

Previous[1/1/23] A Neural Corpus Indexer for Document Retrieval

Last updated 2 years ago

Motivation and Core Insight

Subdomain: Meta-learning — learning not just a task, but a distribution of tasks. Aims to understand underlying underlying abstractions common to tasks.

Problem: Machines learn too slowly, especially compared to humans. They might not be picking up underlying abstractions.

Insight:

  • Humans have inherent inductive biases, which helps them learn faster.

  • If we can train models with these biases, they can also learn faster.

  • We can add these biases using auxiliary tasks — making it navigate multiple representations of a task to think more like a human.

Example and Underlying Design

We want an agent to sweep the grid as per “target board” below. Alongside this, asking it to generate natural language like top-right biases its movement to be explainable, and therefore more human-like.

Adding the idea of a “U-Shape” to the library of functions a program can do allows the model to use abstractions to reach the target board.

Scaled up, one can imagine this could make reinforcement learning more efficient, removing the need for massive amounts of data.

Auxiliary tasks included:

  • Human-generated natural language descriptors

  • Synthetic descriptors — pseudo-code

  • Programs built using library learning (maintaining an ever-growing library of functions that can be composed when trying to solve tasks)

Evaluation

  • The agent got better at tasks humans are great at, and got worse at the control group tasks that humans are bad at.

    • Grounding with human-generated descriptions leads to a human-like inductive bias

    • Grounding on synthetic descriptions does not lead to human-like bias – it improves performance uniformly rather than selectively

    • Grounding on library learning programs leads to a human-like inductive bias

  • Level of abstraction of the task embeddings influences the extent to which the agent acquires a human inductive bias. The mode of information wasn’t the differentiating factor. An example from the paper:

    • The embeddings of these representations that were compressed through abstract concepts may therefore be distilling useful concepts into our meta-learning agent that enables it to acquire a human-like inductive bias.

Limitations and Future Work

  • They expect that the effectiveness of our approach in artificial domain could translate to abstractions in real-world domains that are natural for humans— like mathematical reasoning, planning, or puzzle solving. Future work includes determining the space of tasks for which co-training with language and program abstractions improves an artificial neural agent’s performance.

  • There is still work to do to determine how language and programs can influence acquisition of human inductive biases in more scaled up, naturalistic, and real-world settings.

  • On the language side, the main bottleneck is that collecting human descriptions with sufficient coverage can become prohibitively expensive as the state space grows larger.

  • On the program side, the need to define ad hoc base DSLs for each domain is an important limitation of any approach relying on program abstractions

The Discussion section is a great example of a paper admitting bottle-necks and limitations instead of sweeping them under the rug. Good science.

References

[1] Paper:

👋
https://arxiv.org/abs/2205.11558