George Hotz
👤 PersonAppearances Over Time
Podcast Appearances
I'm still exploring. I want to deliver a really good experience to people. And yeah, what GPUs I end up going with, again, I'm leaning toward AMD. We'll see. You know, in my email, what I said to AMD is like, just dumping the code on GitHub is not open source. Open source is a culture. Open source means that your issues are not all one-year-old stale issues. Open source means developing in public.
I'm still exploring. I want to deliver a really good experience to people. And yeah, what GPUs I end up going with, again, I'm leaning toward AMD. We'll see. You know, in my email, what I said to AMD is like, just dumping the code on GitHub is not open source. Open source is a culture. Open source means that your issues are not all one-year-old stale issues. Open source means developing in public.
And if you guys can commit to that, I see a real future for AMD as a competitor to NVIDIA.
And if you guys can commit to that, I see a real future for AMD as a competitor to NVIDIA.
And if you guys can commit to that, I see a real future for AMD as a competitor to NVIDIA.
We're taking pre-orders. I took this from Elon. I'm like $100 fully refundable pre-orders.
We're taking pre-orders. I took this from Elon. I'm like $100 fully refundable pre-orders.
We're taking pre-orders. I took this from Elon. I'm like $100 fully refundable pre-orders.
No, I'll try to do it faster. It's a lot simpler. It's a lot simpler than a truck.
No, I'll try to do it faster. It's a lot simpler. It's a lot simpler than a truck.
No, I'll try to do it faster. It's a lot simpler. It's a lot simpler than a truck.
The thing that I want to deliver to people out of the box is being able to run 65 billion parameter Lama in FP16 in real time. In like a good, like 10 tokens per second or five tokens per second or something.
The thing that I want to deliver to people out of the box is being able to run 65 billion parameter Lama in FP16 in real time. In like a good, like 10 tokens per second or five tokens per second or something.
The thing that I want to deliver to people out of the box is being able to run 65 billion parameter Lama in FP16 in real time. In like a good, like 10 tokens per second or five tokens per second or something.
Yeah, or I think Falcon is the new one. Experience a chat with the largest language model that you can have in your house.
Yeah, or I think Falcon is the new one. Experience a chat with the largest language model that you can have in your house.
Yeah, or I think Falcon is the new one. Experience a chat with the largest language model that you can have in your house.
From a wall plug, yeah. Actually, for inference, it's not like even more power would help you get more. Even more power wouldn't get you more. Well, no, the biggest model released is 65 billion parameter Lama, as far as I know.
From a wall plug, yeah. Actually, for inference, it's not like even more power would help you get more. Even more power wouldn't get you more. Well, no, the biggest model released is 65 billion parameter Lama, as far as I know.
From a wall plug, yeah. Actually, for inference, it's not like even more power would help you get more. Even more power wouldn't get you more. Well, no, the biggest model released is 65 billion parameter Lama, as far as I know.