Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Damien Tanner

๐Ÿ‘ค Speaker
791 total appearances

Appearances Over Time

Podcast Appearances

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

I mean, you can stream the input, but like you need the complete thing

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

the complete question to send to the LLM to then make a request to the LLM to start generating a response, right?

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

There is no duplex LLM that takes input and generates output at the same time.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

in terms of like, because we have the partial transcripts, yeah.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

So we can stream you the partial transcripts and then say, okay, now it's done.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

Now make the LLM call.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

Then you make the LLM call.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

But interesting, sending text is actually super fast in the context of voice conversation, right?

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

And actually the default example is crazy.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

I didn't think this would work until we tried it, but it just uses a webhook.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

When the user finishes speaking, the basic example sends your Next.js API route a webhook with the user text.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

And turns out the webhook, sending webhook with a few sentences in it, that's like, that's fine.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

It's all the other stuff like then waiting for the LLM to respond.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

Yeah, and we've got a WebSocket endpoint now, so we can kind of shave off that HTTP connection and everything.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

But yeah, then the big heavy latency items come in.

The Changelog: Software Development, Open Source
The era of the Small Giant (Interview)

So generating an LLM response.