Welcome Guest ( Log In | Register )

6 Pages < 1 2 3 4 > » Bottom

Outline · [ Standard ] · Linear+

 Full DeepSeek R1 At Home 🥳🥳🥳

views
     
kurtkob78
post Jan 28 2025, 02:31 PM

Do your best
*******
Senior Member
3,833 posts

Joined: Oct 2006
From: Shah Alam


QUOTE(kingkingyyk @ Jan 28 2025, 02:23 PM)
If you already have some big VRAM GPUs, open source LLM lets you utilize the computing power at no cost (well, you could argue there is TNB bill). smile.gif
*
16GB VRAM sufficient to run o1 level intelligence ?
JonSpark
post Jan 28 2025, 02:31 PM

ai shiteru
*******
Senior Member
4,894 posts

Joined: May 2008
ok.....so world peace achieved?
hellothere131495
post Jan 28 2025, 02:32 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(jmas @ Jan 28 2025, 02:30 PM)
running the 1.5bn q4_k_m locally on my NAS
slow af, but it worked so I dont have to work with o1 free-limit
*
1.5b 4bit is slow? what spec are you using? btw, you probably still need the o1. 1.5b 4 bit is a shit model that hallucinates a lot.
GagalLand
post Jan 28 2025, 02:32 PM

Getting Started
**
Junior Member
239 posts

Joined: May 2022

The speed of China Tech is really terrifying

Luckily such power is in the hand of highly civilised nation

China is the future of humanity
kingkingyyk
post Jan 28 2025, 02:32 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(emefbiemef @ Jan 28 2025, 02:23 PM)
what's the difference, preciousss?
*
https://github.com/deepseek-ai/DeepSeek-R1/...s/benchmark.jpg
Here is the comparison. The full 671b models need over 400GB of memory to run, which is out of reach for most people.

Distillation transfers the knowledge to smaller model (i.e. feeding the QA chain), but smaller model has way fewer parameters so they won't be generating result so well.

The list of distilled models:
- DeepSeek-R1-Distill-Qwen-1.5B
- DeepSeek-R1-Distill-Qwen-7B
- DeepSeek-R1-Distill-Llama-8B
- DeepSeek-R1-Distill-Qwen-14B
- DeepSeek-R1-Distill-Qwen-32B
- DeepSeek-R1-Distill-Llama-70B

32B can let you run on RTX5090 nicely with context (the QA chain is involved to generate further response), but how many of us can justify buying that 5 digits GPU + heavy TNB bill just to run this?

This post has been edited by kingkingyyk: Jan 28 2025, 02:33 PM
GagalLand
post Jan 28 2025, 02:33 PM

Getting Started
**
Junior Member
239 posts

Joined: May 2022

At this rate

Full fledged offline AI robot is possible
hellothere131495
post Jan 28 2025, 02:33 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(kurtkob78 @ Jan 28 2025, 02:31 PM)
16GB VRAM sufficient to run o1 level intelligence ?
*
ah lol. 16GB vram is a kid toy that can only run 4-bit quantized small models.

o1 is a big model. you probably need around 400GB vram to run it (in 4 bit probably). To run the full 32 bit idk need how much. lazy to calculate.
hellothere131495
post Jan 28 2025, 02:35 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(JonSpark @ Jan 28 2025, 02:31 PM)
ok.....so world peace achieved?
*
No. a war has just started? see the news? china deepseek is getting attacked, probably by US
kingkingyyk
post Jan 28 2025, 02:36 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(kurtkob78 @ Jan 28 2025, 02:31 PM)
16GB VRAM sufficient to run o1 level intelligence ?
*
https://ollama.com/
Try the 12b/14b models yourself. cool2.gif
The general performance is quite acceptable, but due to the amount of parameters, won't work well in depth.
7900XTX is the best price-performance card for such purpose. Gives you 24GB VRAM and just cost you 5K.
jmas
post Jan 28 2025, 02:37 PM

I can edit title???
*****
Junior Member
830 posts

Joined: Mar 2010
QUOTE(hellothere131495 @ Jan 28 2025, 02:32 PM)
1.5b 4bit is slow? what spec are you using? btw, you probably still need the o1. 1.5b 4 bit is a shit model that hallucinates a lot.
*
N100 processor je, no GPU, its truenas machine
originally installed ollama just for fun
now considering putting it on my desktop
JonSpark
post Jan 28 2025, 02:50 PM

ai shiteru
*******
Senior Member
4,894 posts

Joined: May 2008
QUOTE(hellothere131495 @ Jan 28 2025, 02:35 PM)
No. a war has just started? see the news? china deepseek is getting attacked, probably by US
*
8 M4 units and no world piss? Best it can do is run Crysis on max?
KitZhai
post Jan 28 2025, 03:01 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



So it's better than Nvidia?
vhs
post Jan 28 2025, 03:01 PM

Getting Started
**
Junior Member
90 posts

Joined: May 2022
QUOTE(GagalLand @ Jan 28 2025, 02:33 PM)
At this rate

Full fledged offline AI robot is possible
*
I think it is still a long way to go before we can get AI to the same level as shown in A.I. movie by Steven Speilberg (2001). What we have now shows the promise of what could be achieved. But the power drain is way too much to put it into a machine that not just needs to run the AI, but still needs portable power to move around as well. There are way more things that need to be improved before we can even get a robot similar to the one that is shown in the Interstellar (2014) movie.

Then we have other benchmark to reach, which seem still very far out of reach.

Robot in Lost in Space (1960s TV series version, not the new movie or Netflix version)
R2D2
C3PO

etc etc

vhs
post Jan 28 2025, 03:03 PM

Getting Started
**
Junior Member
90 posts

Joined: May 2022
QUOTE(hellothere131495 @ Jan 28 2025, 02:35 PM)
No. a war has just started? see the news? china deepseek is getting attacked, probably by US
*
Hackers are doing everyone a favor by doing free blackhat test on new AI system. So that it can be improved better in the future. tongue.gif

kaizoku30
post Jan 28 2025, 03:08 PM

Getting Started
**
Junior Member
86 posts

Joined: Jan 2012
Aiya at home you need robot vacuum, automated home system, not this bull crap to ask hallo how is the weather tomorrow
JonSpark
post Jan 28 2025, 03:09 PM

ai shiteru
*******
Senior Member
4,894 posts

Joined: May 2008
QUOTE(kaizoku30 @ Jan 28 2025, 03:08 PM)
Aiya at home you need robot vacuum, automated home system, not this bull crap to ask hallo how is the weather tomorrow
*
Nah man he's building Skynet at home

Trust me bro
KitZhai
post Jan 28 2025, 03:10 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



Can someone explain this in a easier way?
vhs
post Jan 28 2025, 03:13 PM

Getting Started
**
Junior Member
90 posts

Joined: May 2022
QUOTE(KitZhai @ Jan 28 2025, 03:10 PM)
Can someone explain this in a easier way?
*
https://medium.com/@artturi-jalli/how-to-in...ep-65245873fea8

milolauda
post Jan 28 2025, 03:15 PM

New Member
*
Junior Member
41 posts

Joined: Oct 2010
can calculate toto supreme winning combo
tatmeng
post Jan 28 2025, 03:31 PM

Getting Started
**
Junior Member
191 posts

Joined: Mar 2007


Can run on ipong?

6 Pages < 1 2 3 4 > » Top
 

Change to:
| Lo-Fi Version
0.0164sec    0.71    5 queries    GZIP Disabled
Time is now: 18th December 2025 - 10:48 AM