42 points by asasidh 15 hours ago | 50 comments
Update: A lot of great feedback for me to improve the app. Thank you all.
abujazar 14 hours ago
It'd also need to be much more precise in hardware specs and cover a lot more models and their variants to be actually useful.
Grading the compatibilty is also an absolute requirement – it's rarely an absolute yes or no, but often a question of available GPU memory. There's a lot of other factors too which don't seem to be considered.
rkagerer 14 hours ago
Are you sure it's not powered by an LLM inside?
abujazar 13 hours ago
ggerules 11 hours ago
codingdave 15 hours ago
I can absolutely run models that this site says cannot be run. Shared RAM is a thing - even with limited VRAM, shared RAM can compensate to run larger models. (Slowly, admittedly, but they work.)
lucb1e 14 hours ago
> coined the term in February 2025
> Vibe coding is a new coding style [...] A programmer can describe a program in words and get an AI tool to generate working code, without requiring an understanding of the code. [...] [The programmer] surrenders to the "vibes" of the AI [without reading the resulting code.] When errors arise, he simply copies them into the system without further explanation.
thaumasiotes 13 hours ago
avereveard 15 hours ago
asasidh 15 hours ago
asasidh 15 hours ago
do_not_redeem 14 hours ago
> Yes, you can run this model! Your system has sufficient resources (16GB RAM, 12GB VRAM) to run the smaller distilled version (likely 7B parameters or less) of this model.
Last I checked DeepSeek R1 was a 671B model, not a 7B model. Was this site made with AI?
jsheard 14 hours ago
OP said they "vibe coded" it, so yes.
kennysoona 11 hours ago
reaperman 14 hours ago
Here[0] are some 1.5B and 8B distilled+quantized derivatives of DeepSeek. However, I don’t find a 7B model, that seems totally made-up from whole cloth. Also, I personally wouldn’t call this 8B model “DeepSeek”.
0: https://www.reddit.com/r/LocalLLaMA/comments/1iskrsp/quantiz...
sudohackthenews 14 hours ago
Not technically the full R1 model, it’s talking about the distillations where Deepseek trained Qwen and Llama models based on R1 output
do_not_redeem 14 hours ago
> Yes, you can run this model! Your system has sufficient resources (16GB RAM, 12GB VRAM) to run this model.
No mention of distillations. This was definitely either made by AI, or someone picking numbers for the models totally at random.
sudohackthenews 13 hours ago
monocasa 14 hours ago
That's not ideal from a token throughput perspective, but I can see min working set of weight memory gains if you can load pieces into vram for each token.
throwaway314155 14 hours ago
drodgers 14 hours ago