Dylan Patel
๐ค SpeakerAppearances Over Time
Podcast Appearances
Research gets you O1. Research gets you breakthroughs, and you need to bet on it.
Research gets you O1. Research gets you breakthroughs, and you need to bet on it.
Research and ablations. For ballpark, how much would OpenAI or Anthropic have? I think the clearest example we have, because Meta is also open, they talk about like order of 60k to 100k H100 equivalent GPUs in their training clusters.
Research and ablations. For ballpark, how much would OpenAI or Anthropic have? I think the clearest example we have, because Meta is also open, they talk about like order of 60k to 100k H100 equivalent GPUs in their training clusters.
Research and ablations. For ballpark, how much would OpenAI or Anthropic have? I think the clearest example we have, because Meta is also open, they talk about like order of 60k to 100k H100 equivalent GPUs in their training clusters.
Or whatever, right? I mean, we could get into a cost of like, what is the cost of ownership for a 2000 GPU cluster, 10,000? There's just different sizes of companies that can afford these things. And DeepSeek is...
Or whatever, right? I mean, we could get into a cost of like, what is the cost of ownership for a 2000 GPU cluster, 10,000? There's just different sizes of companies that can afford these things. And DeepSeek is...
Or whatever, right? I mean, we could get into a cost of like, what is the cost of ownership for a 2000 GPU cluster, 10,000? There's just different sizes of companies that can afford these things. And DeepSeek is...
Can you in general actually just zoom out and also talk about the Hopper architecture, the NVIDIA Hopper GPU architecture, and the difference between H100 and H800, like you mentioned, the interconnects?
Can you in general actually just zoom out and also talk about the Hopper architecture, the NVIDIA Hopper GPU architecture, and the difference between H100 and H800, like you mentioned, the interconnects?
Can you in general actually just zoom out and also talk about the Hopper architecture, the NVIDIA Hopper GPU architecture, and the difference between H100 and H800, like you mentioned, the interconnects?
This is very abstract. I think this can be the goal of how some people describe export controls, is this super powerful AI. You touched on the training run idea. There's not many worlds where China cannot train AI models. I think export controls are kneecapping the amount of compute or the density of compute that China can have.
This is very abstract. I think this can be the goal of how some people describe export controls, is this super powerful AI. You touched on the training run idea. There's not many worlds where China cannot train AI models. I think export controls are kneecapping the amount of compute or the density of compute that China can have.
This is very abstract. I think this can be the goal of how some people describe export controls, is this super powerful AI. You touched on the training run idea. There's not many worlds where China cannot train AI models. I think export controls are kneecapping the amount of compute or the density of compute that China can have.
And if you think about the AI ecosystem right now, as all of these AI companies, revenue numbers are up and to the right. Their AI usage is just continuing to grow. More GPUs are going to inference. A large part of export controls, if they work, is just that the amount of AI that can be run in China is going to be much lower.
And if you think about the AI ecosystem right now, as all of these AI companies, revenue numbers are up and to the right. Their AI usage is just continuing to grow. More GPUs are going to inference. A large part of export controls, if they work, is just that the amount of AI that can be run in China is going to be much lower.
And if you think about the AI ecosystem right now, as all of these AI companies, revenue numbers are up and to the right. Their AI usage is just continuing to grow. More GPUs are going to inference. A large part of export controls, if they work, is just that the amount of AI that can be run in China is going to be much lower.
So on the training side, DeepSeek v3 is a great example, which you have a very focused team that can still get to the frontier of AI. This 2,000 GPUs is not that hard to get, all considering in the world. They're still going to have those GPUs. They're still going to be able to train models. But if there's going to be a huge market for AI, if you have strong export controls and you
So on the training side, DeepSeek v3 is a great example, which you have a very focused team that can still get to the frontier of AI. This 2,000 GPUs is not that hard to get, all considering in the world. They're still going to have those GPUs. They're still going to be able to train models. But if there's going to be a huge market for AI, if you have strong export controls and you
So on the training side, DeepSeek v3 is a great example, which you have a very focused team that can still get to the frontier of AI. This 2,000 GPUs is not that hard to get, all considering in the world. They're still going to have those GPUs. They're still going to be able to train models. But if there's going to be a huge market for AI, if you have strong export controls and you