Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Zach Furman

๐Ÿ‘ค Speaker
696 total appearances

Appearances Over Time

Podcast Appearances

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Gradient descent navigates a continuous parameter space using only local information.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

If both processes are somehow arriving at similar destinations, compositional solutions to learning problems, then something interesting is happening in how neural network loss landscapes are structured, something we do not yet understand.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

We will return to this issue at the end of the post.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

So the hypothesis raises as many questions as it answers.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

But it offers something valuable.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

A frame.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

If deep learning is doing a form of program synthesis, that gives us a way to connect disparate observations about generalization, about convergence of representations, about why scaling works, into a coherent picture.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Whether this picture can make sense of more than just these particular examples is what we'll explore next.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

There's a details box here with the title clarifying the hypothesis.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The box contents are omitted from this narration.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Subheading.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Why this isn't enough.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The preceding case studies provide a strong existence proof.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Deep neural networks are capable of learning and implementing non-trivial, compositional algorithms.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

The evidence that Inception V1 solves image classification by composing circuits, or that a transformer solves modular addition by discovering a Fourier-based algorithm, is quite hard to argue with.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

And, of course, there are more examples than these which we have not discussed.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Still, the question remains.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Is this the exception or the rule?

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

It would be completely consistent with the evidence presented so far for this type of behavior to just be a strange-edge case.

LessWrong (Curated & Popular)
"Deep learning as program synthesis" by Zach Furman

Unfortunately, mechanistic interpretability is not yet enough to settle the question.