Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing
Podcast Image

The Neuron: AI Explained

The Humans Behind AI: How Invisible Technologies Trains 80% of the World's Top Models

03 Nov 2025

Transcription

Full Episode

0.031 - 26.722 Corey Knowles

So behind every AI response, there's an invisible army of humans who trained it, labeling images, rating answers, and teaching these models right from wrong. Today, we're going to talk to Casper Elliott from the company that's trained 80% of the world's top AI models. Welcome, humans, to the latest episode of the Neuron Podcast.

0

26.922 - 41.665 Corey Knowles

I'm Corey Knowles, and we're joined, as always, by Grant Harvey, writer of the Neuron Daily AI newsletter. And today, we're diving deep into the human side of AI with Casper Elliott from Invisible Technologies. Casper, thanks so much for joining us.

0

41.645 - 59.166 Caspar Eliot

Pleasure to be here. Thank you for having me. So, Caspar, Invisible Technologies just raised $100 million. Invisible also says it has trained 80% of the world's top AI models. Both of those are incredible stats. Can you just tell us a little bit more about what that process actually looks like?

0

59.387 - 75.689 Grant Harvey

The way I think of it is, so large language models, they're not like traditional machine learning, right? They're non-deterministic. They're based on neural nets. They do some funky things. I think of a large language model like an enthusiastic teenager. Yeah, he really wants to answer questions, and he wants to get smart.

0

75.809 - 89.21 Grant Harvey

But if you want a teenager to learn something, like there are some ways you can teach them, right? You could, for example, you could take them to a library, or you could give them a load of homework, or you could set them a test. We kind of do all those three things.

89.23 - 106.589 Grant Harvey

When you hear supervised fine-tuning or reinforcement learning about human feedback or evaluations, that's actually one of those three things. So supervised fine-tuning is giving a model loads of real high-quality examples of data sets to look like. That's taking your model to the library and saying, here's some textbooks to read. It's going to read the textbooks. They'll tell you what's true.

107.53 - 116.961 Grant Harvey

Reinforcement learning is, okay, you're going to give the model some questions. It'll give some answers, and you're going to say if those answers are good or not. Like you might ask the model to write me a poem about...

116.941 - 146.155 Grant Harvey

Russia and then you'll you'll check that poem and you'll have something about what if that poem is good or bad and you'll give it so great and it'll learn from that and so that's reinforcement learning or you can call it reward modeling you're basically allowing you to change the way you reward your model for different types of answers and then evaluation is building the like the test the model has to take to understand if it's good because companies will release loads of different versions of models and they've got to understand if it's better or worse like I mean you've seen the news about track dbt5 dbt5 they released it

146.135 - 164.013 Grant Harvey

It was obviously better in some metrics, but the audience wasn't happy. And that's why human evaluation is so necessary because people are not deterministic too. People like to have opinions on things and you can't just be like, well, this was better than all our benchmarks. If it feels different to someone and the user doesn't like it, it doesn't matter if it's better. And that's evaluation.

Comments

There are no comments yet.

Please log in to write the first comment.