Damien Tanner
๐ค SpeakerAppearances Over Time
Podcast Appearances
And so you have to then buffer the LLM output into sentences, ship them off sentence by sentence to the voice model.
And then as soon as we get that first chunk of 20 millisecond audio we chunk it up into, we stream that straight back down WebSockets from the Cloudflare Worker straight into the user's browser and can start playing the agent response.
You chose TypeScript to do all this.
We're pretty set on Cloudflare workers from day one.
And it just solves so many infrastructure problems that you're going to run into later on.
Like, I don't think we'll need a DevOps person ever.
It's such a... That's interesting.
It's such a wonderful...
There are constraints you have to build to, right?
You're using V8 JavaScript, browser JavaScript in a Cloudflare worker, right?
Tons of node APIs don't work.
There is a bit of a compatibility layer.
You do have to do things a bit differently.
But what do you get in return?
Your application runs everywhere, 330 locations around the world.
There is essentially zero cold start.
Cloudflare workers start up in the time while the SSL negotiations happening.
The worker has already started.
And you have very few limitations to your scaling, extremely high concurrency.
every instance is very kind of isolated.