Mark Blyth
๐ค SpeakerAppearances Over Time
Podcast Appearances
So we have smaller models, and they're effective, as we said, because of cost, and they're popular because of cost. What does that do to the requirements in terms of compute?
So we have smaller models, and they're effective, as we said, because of cost, and they're popular because of cost. What does that do to the requirements in terms of compute?
So we have smaller models, and they're effective, as we said, because of cost, and they're popular because of cost. What does that do to the requirements in terms of compute?
When we think about the alignment in compute and models, we had David Kahn from Sequoia on the show and he said that you would never train a frontier model on the same data center twice. Meaning that essentially there is now a misalignment in the development speed of models and that is much faster than the development speed of new hardware and compute. How do you think about that?
When we think about the alignment in compute and models, we had David Kahn from Sequoia on the show and he said that you would never train a frontier model on the same data center twice. Meaning that essentially there is now a misalignment in the development speed of models and that is much faster than the development speed of new hardware and compute. How do you think about that?
When we think about the alignment in compute and models, we had David Kahn from Sequoia on the show and he said that you would never train a frontier model on the same data center twice. Meaning that essentially there is now a misalignment in the development speed of models and that is much faster than the development speed of new hardware and compute. How do you think about that?
So we are releasing new models so fast that computers are unable to keep up with them. And as a result, you won't want to train your new model on old H100 hardware that is 18 months old. You need continuously the newest hardware for every single new frontier model.
So we are releasing new models so fast that computers are unable to keep up with them. And as a result, you won't want to train your new model on old H100 hardware that is 18 months old. You need continuously the newest hardware for every single new frontier model.
So we are releasing new models so fast that computers are unable to keep up with them. And as a result, you won't want to train your new model on old H100 hardware that is 18 months old. You need continuously the newest hardware for every single new frontier model.
Speaking of that commoditization, the thing that I'm interested by there is kind of the benchmarking or the determination that they are suddenly commoditized or kind of equal performance. You said before LLM evaluation is a minefield. Help me understand why is LLM evaluation a minefield?
Speaking of that commoditization, the thing that I'm interested by there is kind of the benchmarking or the determination that they are suddenly commoditized or kind of equal performance. You said before LLM evaluation is a minefield. Help me understand why is LLM evaluation a minefield?
Speaking of that commoditization, the thing that I'm interested by there is kind of the benchmarking or the determination that they are suddenly commoditized or kind of equal performance. You said before LLM evaluation is a minefield. Help me understand why is LLM evaluation a minefield?
We mentioned that, you know, some of the early use cases in terms of passing the bar, some real kind of wild applications in terms of how models are applied. I do just want to kind of move a layer deeper to the companies building the products and the leaders leading those companies. You've got Zach and Demis who are saying that AGI is further out than we think.
We mentioned that, you know, some of the early use cases in terms of passing the bar, some real kind of wild applications in terms of how models are applied. I do just want to kind of move a layer deeper to the companies building the products and the leaders leading those companies. You've got Zach and Demis who are saying that AGI is further out than we think.
We mentioned that, you know, some of the early use cases in terms of passing the bar, some real kind of wild applications in terms of how models are applied. I do just want to kind of move a layer deeper to the companies building the products and the leaders leading those companies. You've got Zach and Demis who are saying that AGI is further out than we think.
And then you have Sam Altman and you have Dario and Elon in some cases saying it's sooner than we think. What are your reflections and analysis on company leader predictions on AGI?
And then you have Sam Altman and you have Dario and Elon in some cases saying it's sooner than we think. What are your reflections and analysis on company leader predictions on AGI?
And then you have Sam Altman and you have Dario and Elon in some cases saying it's sooner than we think. What are your reflections and analysis on company leader predictions on AGI?
Is it possible to have a dual strategy of chasing AGI and superintelligence, as OpenAI very clearly are, and creating valuable products at the same time that can be used in everyday use? Or is that balance actually mutually exclusive?
Is it possible to have a dual strategy of chasing AGI and superintelligence, as OpenAI very clearly are, and creating valuable products at the same time that can be used in everyday use? Or is that balance actually mutually exclusive?