Heiki Riesenkampf
👤 PersonAppearances Over Time
Podcast Appearances
And yeah, you know, you asked, have I seen the challenges? Yes, I've seen the challenges, but I feel like two years in the machine learning infraspace is an eternity. And so there will definitely be, you know, new frameworks that we'll have to use or will be new platforms that are probably going to be better pricing than they were a few years ago.
And yeah, you know, you asked, have I seen the challenges? Yes, I've seen the challenges, but I feel like two years in the machine learning infraspace is an eternity. And so there will definitely be, you know, new frameworks that we'll have to use or will be new platforms that are probably going to be better pricing than they were a few years ago.
And yeah, you know, you asked, have I seen the challenges? Yes, I've seen the challenges, but I feel like two years in the machine learning infraspace is an eternity. And so there will definitely be, you know, new frameworks that we'll have to use or will be new platforms that are probably going to be better pricing than they were a few years ago.
Yeah, rely on your past experience, but approach every problem with a fresh mind and willing to be surprised by something new, better, making more sense than what made most sense two years ago.
Yeah, rely on your past experience, but approach every problem with a fresh mind and willing to be surprised by something new, better, making more sense than what made most sense two years ago.
Yeah, rely on your past experience, but approach every problem with a fresh mind and willing to be surprised by something new, better, making more sense than what made most sense two years ago.
the optimizations that have gone to improve the latency of our very first product. Every customer call that we have, latency is always a question that comes up and it's always one of the main parts of the conversation. If you want to have real-time translation, it's always a question of, right, like, what is real-time? Is 2,000 milliseconds latency real-time? Is 1,000 milliseconds real-time?
the optimizations that have gone to improve the latency of our very first product. Every customer call that we have, latency is always a question that comes up and it's always one of the main parts of the conversation. If you want to have real-time translation, it's always a question of, right, like, what is real-time? Is 2,000 milliseconds latency real-time? Is 1,000 milliseconds real-time?
the optimizations that have gone to improve the latency of our very first product. Every customer call that we have, latency is always a question that comes up and it's always one of the main parts of the conversation. If you want to have real-time translation, it's always a question of, right, like, what is real-time? Is 2,000 milliseconds latency real-time? Is 1,000 milliseconds real-time?
Is it 300? When we initially got started, we had a very, let's say, flaky infra in terms of how we scale different streams, how we patch things together. I think I'm proudest of just very sober engineering decisions that led to much more stable infrastructure, that led to quite a lot of improved latency just from
Is it 300? When we initially got started, we had a very, let's say, flaky infra in terms of how we scale different streams, how we patch things together. I think I'm proudest of just very sober engineering decisions that led to much more stable infrastructure, that led to quite a lot of improved latency just from
Is it 300? When we initially got started, we had a very, let's say, flaky infra in terms of how we scale different streams, how we patch things together. I think I'm proudest of just very sober engineering decisions that led to much more stable infrastructure, that led to quite a lot of improved latency just from
optimizing using kind of 80 20 analysis and so trying to reduce networking time trying to reduce the the total number of api hops that we need to do between the api providers and our servers i think i've applied 80 20 very well and gotten a lot of bang for and not a not too much of engineering effort that's probably my
optimizing using kind of 80 20 analysis and so trying to reduce networking time trying to reduce the the total number of api hops that we need to do between the api providers and our servers i think i've applied 80 20 very well and gotten a lot of bang for and not a not too much of engineering effort that's probably my
optimizing using kind of 80 20 analysis and so trying to reduce networking time trying to reduce the the total number of api hops that we need to do between the api providers and our servers i think i've applied 80 20 very well and gotten a lot of bang for and not a not too much of engineering effort that's probably my
That's probably what I'm proudest of, not over-engineering the early version of the product.
That's probably what I'm proudest of, not over-engineering the early version of the product.
That's probably what I'm proudest of, not over-engineering the early version of the product.
We had this bug where the longer the call got, the longer the delay in the original speech and translated speech was. We could not figure out where that came from. It was a very strange bug. I initially thought there was a memory leak, but then we're using TypeScript as our stack, so I was a little bit surprised if it had been something that I would have had control over.
We had this bug where the longer the call got, the longer the delay in the original speech and translated speech was. We could not figure out where that came from. It was a very strange bug. I initially thought there was a memory leak, but then we're using TypeScript as our stack, so I was a little bit surprised if it had been something that I would have had control over.