[TOC]

  1. Title: Language Models as Agent Models
  2. Author: Jacob Andreas
  3. Publish Year: 3 Dec 2022
  4. Review Date: Sat, Dec 10, 2022

https://arxiv.org/pdf/2212.01681.pdf

Summary of paper

Motivation

Contribution

Some key terms

Current language model is bad

definition of language model

the goodness in language model

image-20221210204938004

the belief desire intention model

image-20221210212445444

context and LM and constraints

Event today’s largest language models make major errors involving factuality and coherence

  1. limitation of training datasets
    1. solution: a small annotation about author’s beliefs and goals – or at the very least, richer information about the social and perceptual context in which language is generated – might improve language modelling
  2. limitation of context windows
    1. a few thousand tokens cannot infer an agent state, understood as a complete set of beliefs, desires and intentions, as this is not a small object.
    2. solution: develop new LMs that do not condition on fixed-size context windows or state vectors, but instead explicitly factorise short-term and long-term context components relevant for prediction.
  3. Limitation of LM architecture

Good things about the paper (one paragraph)