Ray Fernando
👤 PersonAppearances Over Time
Podcast Appearances
So these downloads are pretty big, like four gigabytes, and some of them are several gigabytes. So just depending on the space on your phone. So you can actually run the distilled Lama 8-bit MLX version. And I have the distilled Quen version at 7B. So it just depends on your... Oh, that one's actually not compatible. Which one do I have downloaded? So I think on mine, let's see.
The one I have available is the DeepSeek R1 from Apollo. I think I have it from OpenRouter that's running. So let's take a look here. AI providers, OpenRouter. Yeah. So the one that I have set up right now is from OpenRouter. So OpenRouter will show you all the models. You can select DeepSeek R1 from there, which is awesome. So you can have a conversation.
The one I have available is the DeepSeek R1 from Apollo. I think I have it from OpenRouter that's running. So let's take a look here. AI providers, OpenRouter. Yeah. So the one that I have set up right now is from OpenRouter. So OpenRouter will show you all the models. You can select DeepSeek R1 from there, which is awesome. So you can have a conversation.
The one I have available is the DeepSeek R1 from Apollo. I think I have it from OpenRouter that's running. So let's take a look here. AI providers, OpenRouter. Yeah. So the one that I have set up right now is from OpenRouter. So OpenRouter will show you all the models. You can select DeepSeek R1 from there, which is awesome. So you can have a conversation.
So this just requires me being connected to the internet. We start a new chat. You're like, tell me more about options trading. And so here you're still talking to the model, but you're actually just using open router. And so that's a little bit different than, you know, sending your stuff directly to deep seek. And they should be able to do that.
So this just requires me being connected to the internet. We start a new chat. You're like, tell me more about options trading. And so here you're still talking to the model, but you're actually just using open router. And so that's a little bit different than, you know, sending your stuff directly to deep seek. And they should be able to do that.
So this just requires me being connected to the internet. We start a new chat. You're like, tell me more about options trading. And so here you're still talking to the model, but you're actually just using open router. And so that's a little bit different than, you know, sending your stuff directly to deep seek. And they should be able to do that.
It's possible that this model is busy or it's currently down. That can happen. So, yeah, that happens. Yeah. While that's going, I think we could even start another new chat. Let's see this model. You can select a different model. So let's see.
It's possible that this model is busy or it's currently down. That can happen. So, yeah, that happens. Yeah. While that's going, I think we could even start another new chat. Let's see this model. You can select a different model. So let's see.
It's possible that this model is busy or it's currently down. That can happen. So, yeah, that happens. Yeah. While that's going, I think we could even start another new chat. Let's see this model. You can select a different model. So let's see.
There's so many. Yeah. It's like, how do you know which one does it? I feel like you just go off vibes. Like what's, what's my friend telling me? It's yeah. Like what's the real vibes right now? So the vibes right now, obviously R1 is like the real hotness. People are like totally into that right now. Um, and it makes sense cause you know, reasoning, uh, at a much lower costs. So, um, let's see.
There's so many. Yeah. It's like, how do you know which one does it? I feel like you just go off vibes. Like what's, what's my friend telling me? It's yeah. Like what's the real vibes right now? So the vibes right now, obviously R1 is like the real hotness. People are like totally into that right now. Um, and it makes sense cause you know, reasoning, uh, at a much lower costs. So, um, let's see.
There's so many. Yeah. It's like, how do you know which one does it? I feel like you just go off vibes. Like what's, what's my friend telling me? It's yeah. Like what's the real vibes right now? So the vibes right now, obviously R1 is like the real hotness. People are like totally into that right now. Um, and it makes sense cause you know, reasoning, uh, at a much lower costs. So, um, let's see.
Um, there's probably something going wrong with my API key or something. So AI providers, I can select local model to run. You know, I want to see if there's something small here that we can download. So we could do, yeah, this distilled quen, just for speed purposes, we'll just download the gigabyte one. So this is going to download, wow, that's really fast.
Um, there's probably something going wrong with my API key or something. So AI providers, I can select local model to run. You know, I want to see if there's something small here that we can download. So we could do, yeah, this distilled quen, just for speed purposes, we'll just download the gigabyte one. So this is going to download, wow, that's really fast.
Um, there's probably something going wrong with my API key or something. So AI providers, I can select local model to run. You know, I want to see if there's something small here that we can download. So we could do, yeah, this distilled quen, just for speed purposes, we'll just download the gigabyte one. So this is going to download, wow, that's really fast.
the quen model 1.5 b and so that'll run deep seek locally and so basically it's just downloading it directly from i think hugging face and then the model is being loaded on my phone and um this this is actually optimized to run on apple hardware or apple silicon so that's um you know one way that you can kind of take a look at it uh to run this thing and so what's nice yeah if this phone runs out of internet or i need to ask some questions or do some stuff
the quen model 1.5 b and so that'll run deep seek locally and so basically it's just downloading it directly from i think hugging face and then the model is being loaded on my phone and um this this is actually optimized to run on apple hardware or apple silicon so that's um you know one way that you can kind of take a look at it uh to run this thing and so what's nice yeah if this phone runs out of internet or i need to ask some questions or do some stuff
the quen model 1.5 b and so that'll run deep seek locally and so basically it's just downloading it directly from i think hugging face and then the model is being loaded on my phone and um this this is actually optimized to run on apple hardware or apple silicon so that's um you know one way that you can kind of take a look at it uh to run this thing and so what's nice yeah if this phone runs out of internet or i need to ask some questions or do some stuff
I will have this R1 reasoning model that's a much smaller version to run on device. And I think that's another good point about AI that's running. And you don't always need the most powerful thing running for every single type of thing. I think it's really important to... understand different use cases, you know, because maybe you don't need that depth of reasoning.