Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Zach Furman

๐Ÿ‘ค Speaker
696 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

not monolithic black-box computations, but something more like circuits.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

This is reminiscent of the picture we started with.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Solomonoff induction frames learning as a search for simple programs that explain data.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

It is a theoretical ideal, provably optimal in a certain sense, but hopelessly intractable.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The connection between Solomonoff and deep learning has mostly been viewed as purely conceptual.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

A nice way to think about what learning should do with no implications for what neural networks actually do.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

But the evidence from mechanistic interpretability suggests a different possibility.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

What if deep learning is doing something functionally similar to program synthesis?

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Not through the same mechanism, gradient descent on continuous parameters is nothing like enumerative search over discrete programs.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

But perhaps targeting the same kind of object.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Mechanistic solutions, built from parts, that capture structure in the data generating process.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

To be clear, this is a hypothesis.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The evidence shows that neural networks can learn compositional solutions, and that such solutions have appeared alongside generalization in specific, interpretable cases.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

It doesn't show that this is what's always happening, or that there's a consistent bias toward simplicity, or that we understand why gradient descent would find such solutions efficiently.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

but if the hypothesis is right, it would reframe what deep learning is doing.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The success of neural networks would not be a mystery to be accepted, but an instance of something we already understand in principle.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The power of searching for compact, mechanistic models to explain your observations.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The puzzle would shift from why does deep learning work at all to how does gradient descent implement this search so efficiently?

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

That second question is hard.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Solomonoff induction is intractable precisely because the space of programs is vast and discrete.