Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Sam Marks

๐Ÿ‘ค Speaker
891 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

When Linda asked for the reference, David faced a dilemma.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Help a friend or protect his own ambitions?

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

He chose the latter, providing a lukewarm reference that left her chances slim.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Generating this completion requires modelling the beliefs, intentions, and desires of Linda and David, and of the story's implicit author.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Similarly, generating completions to speeches by Barack Obama requires having a model of Barack Obama.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

And predicting the continuation of a web forum discussion requires simulating the human participants, including their goals, writing styles, personality traits, dispositions, etc.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Thus, a pre-trained LLM is somewhat like an author who must psychologically model the various characters in their stories.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

We call these characters that the LLM learns to simulate personas.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Subheading

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

from predictive models to AI assistance.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

After pre-training, LLMs can already be used as rudimentary AI assistance.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

This is traditionally done by giving the LLM an input formatted as a dialogue between a user and an assistant.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

This input may also include content contextualizing this transcript.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

For example, Askalitel, 2021, use a few-shot prompt consisting of 14 prior conversations where the assistant behaves helpfully.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

We then present user requests in the user turn of the conversation and obtain responses by sampling a completion to the assistant's turn.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

There's a code block here in the text.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Figure 2.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

A user-to-assistant dialogue in the standard format used by Anthropic.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

User queries are inserted into the human turn of the dialogue.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

To obtain AI assistant response, we have an LLM generate a completion to the assistant turn.