Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Sam Marks

๐Ÿ‘ค Speaker
891 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Because of this, deep learning likely has an inductive bias towards reusing these capabilities, rather than learning new agentic capabilities from scratch.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

First, observe that persona modeling is a flexible and powerful way to implement agentic behavior.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

During pre-training, LLMs learn to model a large and diverse space of agents who need to pursue their goals in varied circumstances.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Persona simulation is therefore a sort of meta-agency that can be flexibly repurposed for specific choices of goals, beliefs, and other propensities.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

It is therefore ripe to serve as the agentic backend of an AI assistant.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Second, unlike pre-training, post-training for AI assistants is narrowly focused.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Essentially all post-training episodes consist of user-assistant dialogues.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Furthermore, the behaviors we train AI assistants for are persona-consistent.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

That is, they are the sorts of behaviors that a human-like persona from the pre-training distribution could plausibly have.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

We don't train AI assistants to produce strange text outputs that decode into motions of robotic arms and pistons.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

We train them to interact conversationally using natural language in the way that a helpful, knowledgeable, and ethical person would.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Third, deep learning likely has an inductive bias towards reuse of existing mechanisms, like persona modeling.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Analogously, biological evolution tends to adapt useful structures, such as fallen bones in vertebrates, when they are available, instead of independently evolving variants from scratch within the same organism.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

This latter, independent evolution in the same organism output would be analogous to learning non-persona agency from scratch within an LLM that already had strong persona modeling capabilities.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Deep learning would rather just reuse and adapt the existing agentic capabilities bound up in persona models.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

There's an image here.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Figure 5.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Homologous forelimb bones in various vertebrates.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

The same basic structure in a common ancestor was adapted by evolution for multiple downstream purposes.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

In our analogy, personas in the pre-trained LLM are like the forelimb structure in the common ancestor.