Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Zach Furman

๐Ÿ‘ค Speaker
696 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

This is also where the post will shift register.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The remaining sections sketch the structure of these problems and gesture at why certain mathematical frameworks, singular learning theory, algebraic geometry, etc.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

might become relevant.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

I won't develop these fully here, that requires machinery far beyond the scope of a single blog post, but I want to show why you'd need to leave sure at all, and what you might find out in open water.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Subheading The Representation Problem

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The program synthesis hypothesis posits a relationship between two fundamentally different kinds of mathematical objects.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

On one hand, we have programs.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

A program is a discrete and symbolic object.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Its identity is defined by its compositional structure, a graph of distinct operations.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

A small change to this structure, like flipping a comparison or replacing an addition with a subtraction, can create a completely different program with discontinuous, global changes in behavior.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The space of programs is discrete.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

On the other hand, we have neural networks.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

A neural network is defined by its parameter space, a continuous vector space of real valued weights.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The function a network computes is a smooth, or at least piecewise smooth, function of these parameters.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

This smoothness is the essential property that allows for learning via gradient descent a process of infinitesimal steps along a continuous loss landscape.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

This presents a seeming type mismatch.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

How can a continuous process in a continuous parameter space give rise to a discrete, structured program?

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The problem is deeper than it first appears.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

To see why, we must first be precise about what we mean when we say a network has learned a program.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

It cannot simply be about the input-output function the network computes.