Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Narrator (TYPE III AUDIO)

👤 Speaker
266 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Either right now, or soon enough, asterisk, also you can frame all of the above as some sort of nefarious collusion, but I don't think that's the right thing to do.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Heading Implications for alignment research and policy

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

The familiar human sense of a coherent, stable, bounded self simply doesn't match reality.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Arguably, it doesn't even match reality well in humans, but with AIs, the mismatch is far greater.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

I think in principle, many people would agree, don't anthropomorphize LLMs.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

But in my experience this is way easier to say than to actually do.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Human-based priors creep in, and mammal priors about identities and individualities run deep.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

My hope is that weird as it may sound, metaphors like Pando or Mycelium or Tulkus can help.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Not at the level write a paper about reincarnation in LLMs, but at the level of intuitions about the minds we may meet.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Practically, researchers should stress-test alignment ideas across diverse notions of self.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

If your safety strategy depends heavily on assumptions about clear agentic boundaries, run experiments or thought exercises challenging those boundaries.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Does your method still hold if AI identities blur, merge, or emerge spontaneously?

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Are you talking about AI characters, predictive ground, weights, model families?

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

What happens if implicit coordination arises between AI's sharing training patterns, even without explicit communication?

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

How robust are your conclusions if there's no stable self at all?

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

A simple mental checkpoint.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

When reasoning about alignment, ask, does this conclusion survive different individuality assumptions or even no-self assumptions?

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

What would help more?

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

Having a better theory of agency, call it hierarchical, coalitional or collective.

LessWrong (Curated & Popular)
“The Pando Problem: Rethinking AI Individuality” by Jan_Kulveit

A bunch of AIs helped with writing this post.