Sam Marks
๐ค SpeakerAppearances Over Time
Podcast Appearances
Lay out a spectrum of views on the exhaustiveness of PSM, ranging from the popular masked Shoggoth view that attributes substantial non-persona agency to the LLM itself, to an antithetical operating system view under which all agency originates from the assistant persona.
2.
Discuss conceptual considerations around the exhaustiveness of PSM and how it might change in the future.
For instance, one reason for PSM to be exhaustive is that personas provide an especially simple way for the LLM to fit the post-training objective.
3.
Survey some relevant empirics.
While these empirical observations don't settle the question of how exhaustive PSM is, we use them as an opportunity to concretely ground the views we discuss.
Our discussion in this section is especially informal, relying heavily on evocative analogies.
There is no well-established definition of agency or goal-directed behaviour, and it's possible that these abstractions are unsuitable in ways that obscure important weaknesses in our analysis.
We nevertheless put these informal questions about the exhaustiveness of PSM forward for future study.
Subheading Shoggoths, Actors, Operating Systems, and Authors
In this section, we describe a spectrum of perspectives on LLM agency.
Roughly speaking, the views here vary on two axes.
One, non-persona agency ascribed to the LLM itself.
At one extreme is the Shoggoth view, which assigns substantial agency to the underlying LLM.
At the other is the operating system view, which assigns none.
In the middle is the router view, where there is some limited non-persona agency in the choice of which persona to enact, but the AI's behavior is always locally persona-like.
2.
Other sources of persona-like agency There may be a interior persona sitting between the assistant and the outer LLM.
For example, even a pre-trained LLM might enact actor persona which is itself enacting the assistant.