Eight Things To Know About Large Language Models
Bowman Samuel R.. Arxiv 2023
[Paper]
Attention Mechanism
Model Architecture
Prompting
Reinforcement Learning
Survey Paper
The widespread public deployment of large language models (LLMs) in recent
months has prompted a wave of new attention and engagement from advocates,
policymakers, and scholars from many fields. This attention is a timely
response to the many urgent questions that this technology raises, but it can
sometimes miss important considerations. This paper surveys the evidence for
eight potentially surprising such points:
- LLMs predictably get more capable with increasing investment, even without
targeted innovation.
- Many important LLM behaviors emerge unpredictably as a byproduct of
increasing investment.
- LLMs often appear to learn and use representations of the outside world.
- There are no reliable techniques for steering the behavior of LLMs.
- Experts are not yet able to interpret the inner workings of LLMs.
- Human performance on a task isn’t an upper bound on LLM performance.
- LLMs need not express the values of their creators nor the values encoded
in web text.
- Brief interactions with LLMs are often misleading.
Similar Work