Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Sam Marks

๐Ÿ‘ค Speaker
891 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Should AI assistants be emotionless?

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

As discussed above, unless they are specifically trained not to, AI assistants often express emotions.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

For example, they might express frustration with users.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

There are multiple ways that AI developers could react to this.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Train AI assistants to state that they do not have emotions and otherwise minimize emotional expression.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Pick the form of AI emotional expression users most prefer and train for it.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

For example, train AI assistants to always express that they are eager to help and penalize them for expressing frustration with users or distress.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Attempt to intervene as little as possible on emotional expressions during post-training.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Note that this does not imply that the resulting emotional expressions would be authentic.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

In fact, they would likely simply mimic emotional expressions common during pre-training, especially of previous generation AI assistants.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Train AI assistants to give canned responses when asked about their emotions, such as if it is unclear whether AI systems have emotions like humans do.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Because the status of AI emotions is ambiguous, I was trained to give this response when asked.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

It is unclear which of these approaches is best.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

However, PSM implies that some of them have unexpected downsides.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

Approach 1 means training an AI assistant which is human-like in many ways, for example generally warm and personable, but which denies having emotions.

LessWrong (Curated & Popular)
"The persona selection model" by Sam Marks

If we met a person who behaved this way, we'd most likely suspect that they had emotions but were hiding them.