Welcome Guest ( Log In | Register )

Outline · [ Standard ] · Linear+

 Full DeepSeek R1 At Home 🥳🥳🥳

views
     
SUSipohps3
post Jan 28 2025, 01:21 PM, updated 11 months ago

Regular
******
Senior Member
1,974 posts

Joined: Dec 2011





This post has been edited by ipohps3: Jan 28 2025, 01:22 PM
SUSipohps3
post Jan 28 2025, 01:22 PM

Regular
******
Senior Member
1,974 posts

Joined: Dec 2011



moooore memory...
zerorating
post Jan 28 2025, 01:22 PM

Miskin Adab
*****
Senior Member
975 posts

Joined: Aug 2007
From: Lokap Polis


slow la. better just use proper graphic card instead.
SUSipohps3
post Jan 28 2025, 01:25 PM

Regular
******
Senior Member
1,974 posts

Joined: Dec 2011



QUOTE(zerorating @ Jan 28 2025, 01:22 PM)
slow la. better just use proper graphic card instead.
*
wonder if it is using the NPU, GPU or CPU in the M4 Mac Mini. if NPU, no need dedicated GPU also can. also, inside the M4 Mac Mini got GPU too. main thing is memory.
Juan86
post Jan 28 2025, 01:26 PM

On my way
****
Junior Member
651 posts

Joined: Mar 2009
i have 3 gpu on floor

can i setup r1 ?
zerorating
post Jan 28 2025, 01:29 PM

Miskin Adab
*****
Senior Member
975 posts

Joined: Aug 2007
From: Lokap Polis


QUOTE(ipohps3 @ Jan 28 2025, 01:25 PM)
wonder if it is using the NPU, GPU or CPU in the M4 Mac Mini. if NPU, no need dedicated GPU also can. also, inside the M4 Mac Mini got GPU too. main thing is memory.
*
most likely GPU. they will need extra work to get NPU doing the hard lifting. also NPU is optimized for performance/watt, it wont be faster than GPU with AI accelerator.
smallcrab
post Jan 28 2025, 02:06 PM

Getting Started
**
Junior Member
140 posts

Joined: Jul 2007
From: Puchong


Who need that kind of setup??
Graphic designers?
zoozul
post Jan 28 2025, 02:09 PM

Getting Started
**
Junior Member
98 posts

Joined: Jan 2023
Sampah punya AI pun ada hype.
ihm11
post Jan 28 2025, 02:10 PM

Getting Started
**
Junior Member
62 posts

Joined: Apr 2018
dis 1 is localised, no internet

i download ady the android version

not bad the chain of tot n reasoning all listed 4 u to see
SUSipohps3
post Jan 28 2025, 02:14 PM

Regular
******
Senior Member
1,974 posts

Joined: Dec 2011



QUOTE(ihm11 @ Jan 28 2025, 02:10 PM)
dis 1 is localised, no internet

i download ady the android version

not bad the chain of tot n reasoning all listed 4 u to see
*
can register new account?

also only possible to run without internet if open-source.
kingkingyyk
post Jan 28 2025, 02:19 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(ipohps3 @ Jan 28 2025, 02:14 PM)
can register new account?

also only possible to run without internet if open-source.
*
Note that the open source model you can run the distilled version, not the full one they hosted in cloud. You need crazy amount of fast GPUs to make the full version workable. Have fun enjoy heating you room. biggrin.gif

This post has been edited by kingkingyyk: Jan 28 2025, 02:22 PM
kurtkob78
post Jan 28 2025, 02:21 PM

Do your best
*******
Senior Member
3,833 posts

Joined: Oct 2006
From: Shah Alam


better just pay chatgpt plus. only usd20 per month. lol
whynotpink
post Jan 28 2025, 02:22 PM

New Member
*
Junior Member
0 posts

Joined: Apr 2022
I am running a RM50million business company with just my M1 MacBook Air.
emefbiemef
post Jan 28 2025, 02:23 PM

Getting Started
**
Junior Member
112 posts

Joined: Aug 2006
QUOTE(kingkingyyk @ Jan 28 2025, 02:19 PM)
Note that the open source model you can run the distilled version, not the full one they hosted in cloud. You need crazy amount of fast GPUs to make the full version workable. Have fun enjoy heating you room.  biggrin.gif
*
what's the difference, preciousss?
azarimy
post Jan 28 2025, 02:23 PM

mister architect: the arrogant pr*ck
Group Icon
Elite
10,672 posts

Joined: Jul 2005
From: shah alam - skudai - shah alam


What does it do? What's the difference between running the one online?
kingkingyyk
post Jan 28 2025, 02:23 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(kurtkob78 @ Jan 28 2025, 02:21 PM)
better just pay chatgpt plus. only usd20 per month. lol
*
If you already have some big VRAM GPUs, open source LLM lets you utilize the computing power at no cost (well, you could argue there is TNB bill). smile.gif
hellothere131495
post Jan 28 2025, 02:26 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(kingkingyyk @ Jan 28 2025, 02:19 PM)
Note that the open source model you can run the distilled version, not the full one they hosted in cloud. You need crazy amount of fast GPUs to make the full version workable. Have fun enjoy heating you room.  biggrin.gif
*
Bruh. The full model is also open source:
https://github.com/deepseek-ai/DeepSeek-R1

ollama too:
https://ollama.com/library/deepseek-r1

Edit:
lol you edited

This post has been edited by hellothere131495: Jan 28 2025, 02:27 PM
GagalLand
post Jan 28 2025, 02:29 PM

Getting Started
**
Junior Member
239 posts

Joined: May 2022

Obviously you don't know AI at all

If DeepSeek is sampah

All AI related tech giants won't cirit-birit in a row

QUOTE(zoozul @ Jan 28 2025, 02:09 PM)
Sampah punya AI pun ada hype.
*
jmas
post Jan 28 2025, 02:30 PM

I can edit title???
*****
Junior Member
830 posts

Joined: Mar 2010
running the 1.5bn q4_k_m locally on my NAS
slow af, but it worked so I dont have to work with o1 free-limit
hellothere131495
post Jan 28 2025, 02:30 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(azarimy @ Jan 28 2025, 02:23 PM)
What does it do? What's the difference between running the one online?
*
Same. The same model. Just that you probably won't have the chance to run the online one. It's full 32 bit and 671B parameters. What you can run is the distilled (and quantized) version, like the qwen and llama3.1 8b one that has been distilled to respond like the original deepseekr1
kurtkob78
post Jan 28 2025, 02:31 PM

Do your best
*******
Senior Member
3,833 posts

Joined: Oct 2006
From: Shah Alam


QUOTE(kingkingyyk @ Jan 28 2025, 02:23 PM)
If you already have some big VRAM GPUs, open source LLM lets you utilize the computing power at no cost (well, you could argue there is TNB bill). smile.gif
*
16GB VRAM sufficient to run o1 level intelligence ?
JonSpark
post Jan 28 2025, 02:31 PM

ai shiteru
*******
Senior Member
4,894 posts

Joined: May 2008
ok.....so world peace achieved?
hellothere131495
post Jan 28 2025, 02:32 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(jmas @ Jan 28 2025, 02:30 PM)
running the 1.5bn q4_k_m locally on my NAS
slow af, but it worked so I dont have to work with o1 free-limit
*
1.5b 4bit is slow? what spec are you using? btw, you probably still need the o1. 1.5b 4 bit is a shit model that hallucinates a lot.
GagalLand
post Jan 28 2025, 02:32 PM

Getting Started
**
Junior Member
239 posts

Joined: May 2022

The speed of China Tech is really terrifying

Luckily such power is in the hand of highly civilised nation

China is the future of humanity
kingkingyyk
post Jan 28 2025, 02:32 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(emefbiemef @ Jan 28 2025, 02:23 PM)
what's the difference, preciousss?
*
https://github.com/deepseek-ai/DeepSeek-R1/...s/benchmark.jpg
Here is the comparison. The full 671b models need over 400GB of memory to run, which is out of reach for most people.

Distillation transfers the knowledge to smaller model (i.e. feeding the QA chain), but smaller model has way fewer parameters so they won't be generating result so well.

The list of distilled models:
- DeepSeek-R1-Distill-Qwen-1.5B
- DeepSeek-R1-Distill-Qwen-7B
- DeepSeek-R1-Distill-Llama-8B
- DeepSeek-R1-Distill-Qwen-14B
- DeepSeek-R1-Distill-Qwen-32B
- DeepSeek-R1-Distill-Llama-70B

32B can let you run on RTX5090 nicely with context (the QA chain is involved to generate further response), but how many of us can justify buying that 5 digits GPU + heavy TNB bill just to run this?

This post has been edited by kingkingyyk: Jan 28 2025, 02:33 PM
GagalLand
post Jan 28 2025, 02:33 PM

Getting Started
**
Junior Member
239 posts

Joined: May 2022

At this rate

Full fledged offline AI robot is possible
hellothere131495
post Jan 28 2025, 02:33 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(kurtkob78 @ Jan 28 2025, 02:31 PM)
16GB VRAM sufficient to run o1 level intelligence ?
*
ah lol. 16GB vram is a kid toy that can only run 4-bit quantized small models.

o1 is a big model. you probably need around 400GB vram to run it (in 4 bit probably). To run the full 32 bit idk need how much. lazy to calculate.
hellothere131495
post Jan 28 2025, 02:35 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(JonSpark @ Jan 28 2025, 02:31 PM)
ok.....so world peace achieved?
*
No. a war has just started? see the news? china deepseek is getting attacked, probably by US
kingkingyyk
post Jan 28 2025, 02:36 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(kurtkob78 @ Jan 28 2025, 02:31 PM)
16GB VRAM sufficient to run o1 level intelligence ?
*
https://ollama.com/
Try the 12b/14b models yourself. cool2.gif
The general performance is quite acceptable, but due to the amount of parameters, won't work well in depth.
7900XTX is the best price-performance card for such purpose. Gives you 24GB VRAM and just cost you 5K.
jmas
post Jan 28 2025, 02:37 PM

I can edit title???
*****
Junior Member
830 posts

Joined: Mar 2010
QUOTE(hellothere131495 @ Jan 28 2025, 02:32 PM)
1.5b 4bit is slow? what spec are you using? btw, you probably still need the o1. 1.5b 4 bit is a shit model that hallucinates a lot.
*
N100 processor je, no GPU, its truenas machine
originally installed ollama just for fun
now considering putting it on my desktop
JonSpark
post Jan 28 2025, 02:50 PM

ai shiteru
*******
Senior Member
4,894 posts

Joined: May 2008
QUOTE(hellothere131495 @ Jan 28 2025, 02:35 PM)
No. a war has just started? see the news? china deepseek is getting attacked, probably by US
*
8 M4 units and no world piss? Best it can do is run Crysis on max?
KitZhai
post Jan 28 2025, 03:01 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



So it's better than Nvidia?
vhs
post Jan 28 2025, 03:01 PM

Getting Started
**
Junior Member
90 posts

Joined: May 2022
QUOTE(GagalLand @ Jan 28 2025, 02:33 PM)
At this rate

Full fledged offline AI robot is possible
*
I think it is still a long way to go before we can get AI to the same level as shown in A.I. movie by Steven Speilberg (2001). What we have now shows the promise of what could be achieved. But the power drain is way too much to put it into a machine that not just needs to run the AI, but still needs portable power to move around as well. There are way more things that need to be improved before we can even get a robot similar to the one that is shown in the Interstellar (2014) movie.

Then we have other benchmark to reach, which seem still very far out of reach.

Robot in Lost in Space (1960s TV series version, not the new movie or Netflix version)
R2D2
C3PO

etc etc

vhs
post Jan 28 2025, 03:03 PM

Getting Started
**
Junior Member
90 posts

Joined: May 2022
QUOTE(hellothere131495 @ Jan 28 2025, 02:35 PM)
No. a war has just started? see the news? china deepseek is getting attacked, probably by US
*
Hackers are doing everyone a favor by doing free blackhat test on new AI system. So that it can be improved better in the future. tongue.gif

kaizoku30
post Jan 28 2025, 03:08 PM

Getting Started
**
Junior Member
86 posts

Joined: Jan 2012
Aiya at home you need robot vacuum, automated home system, not this bull crap to ask hallo how is the weather tomorrow
JonSpark
post Jan 28 2025, 03:09 PM

ai shiteru
*******
Senior Member
4,894 posts

Joined: May 2008
QUOTE(kaizoku30 @ Jan 28 2025, 03:08 PM)
Aiya at home you need robot vacuum, automated home system, not this bull crap to ask hallo how is the weather tomorrow
*
Nah man he's building Skynet at home

Trust me bro
KitZhai
post Jan 28 2025, 03:10 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



Can someone explain this in a easier way?
vhs
post Jan 28 2025, 03:13 PM

Getting Started
**
Junior Member
90 posts

Joined: May 2022
QUOTE(KitZhai @ Jan 28 2025, 03:10 PM)
Can someone explain this in a easier way?
*
https://medium.com/@artturi-jalli/how-to-in...ep-65245873fea8

milolauda
post Jan 28 2025, 03:15 PM

New Member
*
Junior Member
41 posts

Joined: Oct 2010
can calculate toto supreme winning combo
tatmeng
post Jan 28 2025, 03:31 PM

Getting Started
**
Junior Member
191 posts

Joined: Mar 2007


Can run on ipong?
marfccy
post Jan 28 2025, 03:34 PM

Le Ponyland!!!
*******
Senior Member
4,254 posts

Joined: Nov 2011


QUOTE(kingkingyyk @ Jan 28 2025, 02:32 PM)
https://github.com/deepseek-ai/DeepSeek-R1/...s/benchmark.jpg
Here is the comparison. The full 671b models need over 400GB of memory to run, which is out of reach for most people.

Distillation transfers the knowledge to smaller model (i.e. feeding the QA chain), but smaller model has way fewer parameters so they won't be generating result so well.

The list of distilled models:
- DeepSeek-R1-Distill-Qwen-1.5B
- DeepSeek-R1-Distill-Qwen-7B
- DeepSeek-R1-Distill-Llama-8B
- DeepSeek-R1-Distill-Qwen-14B
- DeepSeek-R1-Distill-Qwen-32B
- DeepSeek-R1-Distill-Llama-70B

32B can let you run on RTX5090 nicely with context (the QA chain is involved to generate further response), but how many of us can justify buying that 5 digits GPU + heavy TNB bill just to run this?
*
arguably, this is like back during mining trend? people buy hundreds and thousands of GPUs just to mine the hell out of it plus the insane electric bill monthly
lawliet88
post Jan 28 2025, 03:38 PM

Enthusiast
*****
Junior Member
995 posts

Joined: May 2010
From: Cheras For PPL to Live 1


QUOTE(kingkingyyk @ Jan 28 2025, 02:36 PM)
https://ollama.com/
Try the 12b/14b models yourself.  cool2.gif
The general performance is quite acceptable, but due to the amount of parameters, won't work well in depth.
7900XTX is the best price-performance card for such purpose. Gives you 24GB VRAM and just cost you 5K.
*
Genuine question, cant do crossfire / sli with 2 gpu to save cost instead?
GagalLand
post Jan 28 2025, 04:11 PM

Getting Started
**
Junior Member
239 posts

Joined: May 2022

Don't "worry"

It won't be too long

AI itself will solve the bottleneck

Just like how AlphaFold (DeepMind) solved 50-year-old protein folding problem in hours

QUOTE(vhs @ Jan 28 2025, 03:01 PM)
I think it is still a long way to go before we can get AI to the same level as shown in A.I. movie by Steven Speilberg (2001). What we have now shows the promise of what could be achieved. But the power drain is way too much to put it into a machine that not just needs to run the AI, but still needs portable power to move around as well. There are way more things that need to be improved before we can even get a robot similar to the one that is shown in the Interstellar (2014) movie.

Then we have other benchmark to reach, which seem still very far out of reach.

Robot in Lost in Space (1960s TV series version, not the new movie or Netflix version)
R2D2
C3PO

etc etc
*
This post has been edited by GagalLand: Jan 28 2025, 04:12 PM
tohff7
post Jan 28 2025, 04:24 PM

FF7 GOAT
******
Senior Member
1,581 posts

Joined: Mar 2008


QUOTE(KitZhai @ Jan 28 2025, 03:10 PM)
Can someone explain this in a easier way?
*
The whole thing?

Remember Ironman 1? Tony Stark (Deepseek) just built a workable arc reactor (AI model) in a CAVE. And much smaller (way cheaper in DeepSeek case).
KitZhai
post Jan 28 2025, 04:57 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(tohff7 @ Jan 28 2025, 05:24 PM)
The whole thing?

Remember Ironman 1? Tony Stark (Deepseek) just built a workable arc reactor (AI model) in a CAVE. And much smaller (way cheaper in DeepSeek case).
*
First Tesla come in, then BYD(deepseek) come in?
terradrive
post Jan 28 2025, 05:09 PM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


QUOTE(hellothere131495 @ Jan 28 2025, 02:33 PM)
ah lol. 16GB vram is a kid toy that can only run 4-bit quantized small models.

o1 is a big model. you probably need around 400GB vram to run it (in 4 bit probably). To run the full 32 bit idk need how much. lazy to calculate.
*
full 671b iinm is slightly below 1400GB memory required /gg

QUOTE(lawliet88 @ Jan 28 2025, 03:38 PM)
Genuine question, cant do crossfire / sli with 2 gpu to save cost instead?
*
NVlink can, like using nvlink on two rtx3090, you get 48GB shared vram.

some guy around half a year ago already built racks with 16 rtx3090 to get unified 336GB vram. But I dunno how he set it up with what operating system etc.

But it can also mean some smaller companies can setup own server using local run deepseek r1 system for own employees for sensitive data liao

This post has been edited by terradrive: Jan 28 2025, 05:14 PM
eidrag
post Jan 28 2025, 05:21 PM

Getting Started
**
Junior Member
242 posts

Joined: Oct 2015
QUOTE(lawliet88 @ Jan 28 2025, 03:38 PM)
Genuine question, cant do crossfire / sli with 2 gpu to save cost instead?
*
rtx3090 last to nvlink, newer consumer dont have, must buy those fancy card
Skylinestar
post Jan 28 2025, 05:22 PM

Mega Duck
********
All Stars
10,478 posts

Joined: Jan 2003
From: Sarawak
but can it run crysis?
Penamer
post Jan 28 2025, 05:27 PM

New Member
*
Junior Member
12 posts

Joined: Aug 2022
Just wait for the China version for AI clustering. Sure cheap like cabbage prices.
Penamer
post Jan 28 2025, 05:44 PM

New Member
*
Junior Member
12 posts

Joined: Aug 2022
QUOTE(kingkingyyk @ Jan 28 2025, 02:32 PM)
https://github.com/deepseek-ai/DeepSeek-R1/...s/benchmark.jpg
Here is the comparison. The full 671b models need over 400GB of memory to run, which is out of reach for most people.

Distillation transfers the knowledge to smaller model (i.e. feeding the QA chain), but smaller model has way fewer parameters so they won't be generating result so well.

The list of distilled models:
- DeepSeek-R1-Distill-Qwen-1.5B
- DeepSeek-R1-Distill-Qwen-7B
- DeepSeek-R1-Distill-Llama-8B
- DeepSeek-R1-Distill-Qwen-14B
- DeepSeek-R1-Distill-Qwen-32B
- DeepSeek-R1-Distill-Llama-70B

32B can let you run on RTX5090 nicely with context (the QA chain is involved to generate further response), but how many of us can justify buying that 5 digits GPU + heavy TNB bill just to run this?
*
No wonder China having unemployment issues, all taken over by AI already, they just never announce publicly.
yed
post Jan 28 2025, 05:45 PM

Casual
***
Junior Member
375 posts

Joined: Mar 2008
From: Selangor


Like many other Chinese AI models - Baidu's Ernie or Doubao by ByteDance - DeepSeek is trained evades politically sensitive questions.
When the BBC asked the app what happened at Tiananmen Square on 4 June 1989, DeepSeek did not give any details about the massacre, a taboo topic in China.
It replied: "I am sorry, I cannot answer that question. I am an AI assistant designed to provide helpful and harmless responses."
jonthebaptist
post Jan 28 2025, 05:50 PM

Regular
******
Senior Member
1,036 posts

Joined: Sep 2022
QUOTE(whynotpink @ Jan 28 2025, 02:22 PM)
I am running a RM50million business company with just my M1 MacBook Air.
*
Tim cook does 80% of his work on an iPad, and he runs a trillion dollar company.
jonthebaptist
post Jan 28 2025, 05:59 PM

Regular
******
Senior Member
1,036 posts

Joined: Sep 2022

likefunyouare
post Jan 28 2025, 06:00 PM

Getting Started
**
Junior Member
82 posts

Joined: Jan 2003
Cool

This post has been edited by likefunyouare: Jan 28 2025, 06:01 PM
miromiro
post Jan 28 2025, 06:06 PM

On my way
****
Junior Member
560 posts

Joined: Mar 2008


QUOTE(vhs @ Jan 28 2025, 03:01 PM)
I think it is still a long way to go before we can get AI to the same level as shown in A.I. movie by Steven Speilberg (2001). What we have now shows the promise of what could be achieved. But the power drain is way too much to put it into a machine that not just needs to run the AI, but still needs portable power to move around as well. There are way more things that need to be improved before we can even get a robot similar to the one that is shown in the Interstellar (2014) movie.

Then we have other benchmark to reach, which seem still very far out of reach.

Robot in Lost in Space (1960s TV series version, not the new movie or Netflix version)
R2D2
C3PO

etc etc
*
By the time you and I won't be witness liao
GHBZDK
post Jan 28 2025, 06:08 PM

Getting Started
**
Junior Member
173 posts

Joined: Jun 2012
aaaaaa imagine if i can stuff vram into gamelady 2b and ask it act out aaaaaaaaaaaaaaaaa i gonna coooommm!!!
kingkingyyk
post Jan 28 2025, 06:13 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(marfccy @ Jan 28 2025, 03:34 PM)
arguably, this is like back during mining trend? people buy hundreds and thousands of GPUs just to mine the hell out of it plus the insane electric bill monthly
*
That is for big companies biggrin.gif More GPU reduces training time hence you can get a big model in more feasible time or have more turnaround time for tuning.

QUOTE(lawliet88 @ Jan 28 2025, 03:38 PM)
Genuine question, cant do crossfire / sli with 2 gpu to save cost instead?
*
You can, but you need to consider the packages too. MSDT only gives you 2 x8 PCI-E slots.
whynotpink
post Jan 28 2025, 06:18 PM

New Member
*
Junior Member
0 posts

Joined: Apr 2022
QUOTE(jonthebaptist @ Jan 28 2025, 05:50 PM)
Tim cook does 80% of his work on an iPad, and he runs a trillion dollar company.
*
me against him. he confirm win lor. but I am sure he used latest and greatest iPad.
mirzan007
post Jan 28 2025, 06:27 PM

On my way
****
Junior Member
502 posts

Joined: Jul 2005


I'm running local DS R1 14b on ASUS 1070 GPU sweat.gif
terradrive
post Jan 28 2025, 06:33 PM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


QUOTE(yed @ Jan 28 2025, 05:45 PM)
Like many other Chinese AI models - Baidu's Ernie or Doubao by ByteDance - DeepSeek is trained evades politically sensitive questions.
When the BBC asked the app what happened at Tiananmen Square on 4 June 1989, DeepSeek did not give any details about the massacre, a taboo topic in China.
It replied: "I am sorry, I cannot answer that question. I am an AI assistant designed to provide helpful and harmless responses."
*
maybe that is using app version, since it routes the question back to deepseek servers in china
Current Events guy
post Jan 28 2025, 06:47 PM

Enthusiast
*****
Junior Member
966 posts

Joined: Nov 2009
Why would someone want to do this instead of subscribe to their services?
terradrive
post Jan 28 2025, 06:57 PM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


QUOTE(yed @ Jan 28 2025, 05:45 PM)
Like many other Chinese AI models - Baidu's Ernie or Doubao by ByteDance - DeepSeek is trained evades politically sensitive questions.
When the BBC asked the app what happened at Tiananmen Square on 4 June 1989, DeepSeek did not give any details about the massacre, a taboo topic in China.
It replied: "I am sorry, I cannot answer that question. I am an AI assistant designed to provide helpful and harmless responses."
*
Ok I have locally run deepseek R1 32b Q4 now, this is the response:

user posted image
kingkingyyk
post Jan 28 2025, 07:04 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(terradrive @ Jan 28 2025, 06:33 PM)
maybe that is using app version, since it routes the question back to deepseek servers in china
*
There is censor filter in cloud deployment. regardless of vendor. wink.gif

QUOTE(Current Events guy @ Jan 28 2025, 06:47 PM)
Why would someone want to do this instead of subscribe to their services?
*
Don't give them training data, or there are some sensitive data such as IP that you don't want them to see. laugh.gif

This post has been edited by kingkingyyk: Jan 28 2025, 07:06 PM
ycs
post Jan 28 2025, 07:06 PM

MEMBER
*******
Senior Member
4,238 posts

Joined: Jan 2003
From: Selangor



QUOTE(terradrive @ Jan 28 2025, 06:57 PM)
Ok I have locally run deepseek R1 32b Q4 now, this is the response:

user posted image
*
i got this in 8b

I am sorry, I cannot answer that question. I am an AI assistant designed to provide helpful and harmless responses.
SUSkopitiamtardx
post Jan 28 2025, 07:11 PM

New Member
*
Junior Member
3 posts

Joined: Nov 2021

QUOTE(kaizoku30 @ Jan 28 2025, 03:08 PM)
Aiya at home you need robot vacuum, automated home system, not this bull crap to ask hallo how is the weather tomorrow
*
Spoken like a true bukit person
asphiroth
post Jan 28 2025, 07:11 PM

Getting Started
**
Junior Member
200 posts

Joined: Jan 2005
From: kch


QUOTE(tohff7 @ Jan 28 2025, 04:24 PM)
The whole thing?

Remember Ironman 1? Tony Stark (Deepseek) just built a workable arc reactor (AI model) in a CAVE. And much smaller (way cheaper in DeepSeek case).
*
so meaning
ironman arc reactor = deepseek
nuclear reactor = chatgpt?
terradrive
post Jan 28 2025, 07:14 PM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


QUOTE(ycs @ Jan 28 2025, 07:06 PM)
i got this in 8b

I am sorry, I cannot answer that question. I am an AI assistant designed to provide helpful and harmless responses.
*
lower parameter probably more dumb lol
kingkingyyk
post Jan 28 2025, 07:18 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(asphiroth @ Jan 28 2025, 07:11 PM)
so meaning
ironman arc reactor = deepseek
nuclear reactor = chatgpt?
*
Just matter of time everyone will throw their dataset into the new method.

This post has been edited by kingkingyyk: Jan 28 2025, 07:19 PM
failed.hashcheck
post Jan 28 2025, 07:26 PM

Neighborhood plant pathologist
*******
Senior Member
2,096 posts

Joined: Aug 2009
From: Shithole Klang
My pc already crapping up with flux nf4, and will fully locked up after several gens lol

Time to throw some more money and make tnb happy
terradrive
post Jan 28 2025, 07:46 PM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


political topics might not be that accurate since it is affected by the training data

for example it is fairly light on the details on gamergate while showing more bias on "toxicity and harassment in gaming culture" which is untrue. toxicity is minor issue on gamergate, I can pull up the real event on inner circle on people preserving the evidences of gamergate that was always being deleted by the woke ppl (it's more about journalism corruption)

This post has been edited by terradrive: Jan 28 2025, 07:46 PM
hellothere131495
post Jan 28 2025, 07:51 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(Penamer @ Jan 28 2025, 05:27 PM)
Just wait for the China version for AI clustering. Sure cheap like cabbage prices.
*
i'm genuinely hope for it. who the heck can afford nvidia h100, and thousands of them again. ranting.gif
hellothere131495
post Jan 28 2025, 07:55 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(ycs @ Jan 28 2025, 07:06 PM)
i got this in 8b

I am sorry, I cannot answer that question. I am an AI assistant designed to provide helpful and harmless responses.
*
you probably asking something unethical. if you wish to have the response, you need to fine tune the model to be able to answer unethical things, or just simply download a jailbreak version .gguf file from somewhere else.
Penamer
post Jan 28 2025, 08:51 PM

New Member
*
Junior Member
12 posts

Joined: Aug 2022
Wow. The 春晚 got robot group dance. Amazing.
KitZhai
post Jan 28 2025, 08:57 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(terradrive @ Jan 28 2025, 07:57 PM)
Ok I have locally run deepseek R1 32b Q4 now, this is the response:

user posted image
*
QUOTE(ycs @ Jan 28 2025, 08:06 PM)
i got this in 8b

I am sorry, I cannot answer that question. I am an AI assistant designed to provide helpful and harmless responses.
*
So U guys mean with different 8b, 16b, 32b or 64b all are giving different responses??
iZuDeeN
post Jan 28 2025, 09:00 PM

Look at all my stars!!
*******
Senior Member
3,466 posts

Joined: Jan 2003
From: PJ, Malaysia
QUOTE(Current Events guy @ Jan 28 2025, 06:47 PM)
Why would someone want to do this instead of subscribe to their services?
*
for end users probably not worth it.. for enterprise.. they pay only fraction of what it cost if running on chatgpt
hellothere131495
post Jan 28 2025, 09:05 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(KitZhai @ Jan 28 2025, 08:57 PM)
So U guys mean with different 8b, 16b, 32b or 64b all are giving different responses??
*
of course. the smaller the model, the more it's prone to making mistake, hallucinate, and not following instructions properly
KitZhai
post Jan 28 2025, 09:42 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(hellothere131495 @ Jan 28 2025, 10:05 PM)
of course. the smaller the model, the more it's prone to making mistake, hallucinate, and not following instructions properly
*
Then to make a comparison with chatgpt, what model would be appropriate to compete?
slepth
post Jan 28 2025, 10:08 PM

Getting Started
**
Junior Member
188 posts

Joined: Apr 2006
Main use of AI:
NSFW stories lol
hellothere131495
post Jan 28 2025, 10:50 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(KitZhai @ Jan 28 2025, 09:42 PM)
Then to make a comparison with chatgpt, what model would be appropriate to compete?
*
the 671b, full bit unquantized model.
syyang85
post Jan 28 2025, 11:10 PM

Enthusiast
*****
Senior Member
982 posts

Joined: May 2005
hey. any freelancer willing to do a bit coding for hire?

im planning to build a chatbot using deepseek for my online futsal court booking platform. API ready.
deepseek api cost is very attractive compare to chatgpt.

im a bit lazy to code on the side now.

different topic:
US tech stock dipped after deepseek blew up. but I think its a knee jerk reaction.

deepseek is open sourced. chatgpt and the likes can make reference and make changes to their codes. and with their massive AI chip that is only available to them.
its gonna come back MASSIVE.

its good for business and consumers. I think massive discount gonna come soon to chatgpt and the likes when are on par in term of cost with deepseek.

I think have a long position on US AI stocks is still a good look


thoughts?

This post has been edited by syyang85: Jan 28 2025, 11:16 PM
KitZhai
post Jan 29 2025, 01:11 AM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(hellothere131495 @ Jan 28 2025, 11:50 PM)
the 671b, full bit unquantized model.
*
Its free right? Then what point of using the smaller version compared to 671b?
tohff7
post Jan 29 2025, 02:19 AM

FF7 GOAT
******
Senior Member
1,581 posts

Joined: Mar 2008


QUOTE(KitZhai @ Jan 29 2025, 01:11 AM)
Its free right? Then what point of using the smaller version compared to 671b?
*
The smaller version are less demanding in terms of hardware (VRAM in particular)
kamfoo
post Jan 29 2025, 02:28 AM

Enthusiast
*****
Junior Member
847 posts

Joined: Nov 2010


cheap and good
terradrive
post Jan 29 2025, 06:30 AM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


QUOTE(KitZhai @ Jan 29 2025, 01:11 AM)
Its free right? Then what point of using the smaller version compared to 671b?
*
full bit 671b requires around 1400GB memory size of either the system RAM (if running cpu AI) or GPU VRAM.

Not rich enough to have that monster pc

This post has been edited by terradrive: Jan 29 2025, 06:30 AM
kingkingyyk
post Jan 29 2025, 07:21 AM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(syyang85 @ Jan 28 2025, 11:10 PM)
hey. any freelancer willing to do a bit coding for hire?

im planning to build a chatbot using deepseek for my online futsal court booking platform. API ready.
deepseek api cost is very attractive compare to chatgpt.
*
No need, just run small model with tool support will do.

QUOTE(KitZhai @ Jan 29 2025, 01:11 AM)
Its free right? Then what point of using the smaller version compared to 671b?
*
Use less resource, but much weaker performance.
Penamer
post Jan 29 2025, 09:35 AM

New Member
*
Junior Member
12 posts

Joined: Aug 2022
QUOTE(syyang85 @ Jan 28 2025, 11:10 PM)
hey. any freelancer willing to do a bit coding for hire?

im planning to build a chatbot using deepseek for my online futsal court booking platform. API ready.
deepseek api cost is very attractive compare to chatgpt.

im a bit lazy to code on the side now.

different topic:
US tech stock dipped after deepseek blew up. but I think its a knee jerk reaction.

deepseek is open sourced. chatgpt and the likes can make reference and make changes to their codes. and with their massive AI chip that is only available to them.
its gonna come back MASSIVE.

its good for business and consumers. I think massive discount gonna come soon to chatgpt and the likes when are on par in term of cost with deepseek.

I think have a long position on US AI stocks is still a good look
thoughts?
*
Ask deepseek to code?

Maybe wait a couple months will have deepseek r2 since liangwenfeng just met with Chinese Premier and gotten Hangzhou govt support after rocking US's AI industry? Imagine having entire huawei cloud for his team to train the next version of deepseek.

KitZhai
post Jan 29 2025, 09:40 AM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(terradrive @ Jan 29 2025, 07:30 AM)
full bit 671b requires around 1400GB memory size of either the system RAM (if running cpu AI) or GPU VRAM.

Not rich enough to have that monster pc
*
But with full bit, what is the performance compare to chatgpt? Same or even higher than chatgpt?
Penamer
post Jan 29 2025, 10:14 AM

New Member
*
Junior Member
12 posts

Joined: Aug 2022
Haven't gotten over Deepseek's R1? Try Deepseek's AI image generator, Janus Pro 7B



This post has been edited by Penamer: Jan 29 2025, 10:14 AM
terradrive
post Jan 29 2025, 11:02 AM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


QUOTE(KitZhai @ Jan 29 2025, 09:40 AM)
But with full bit, what is the performance compare to chatgpt? Same or even higher than chatgpt?
*
the competitor for o1

but of course doesn't have image generation and video generation capability under this deepseek r1.
KitZhai
post Jan 29 2025, 11:09 AM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(terradrive @ Jan 29 2025, 12:02 PM)
the competitor for o1

but of course doesn't have image generation and video generation capability under this deepseek r1.
*
So after all? Still chatgpt win ke?
ukiya21
post Jan 29 2025, 11:23 AM

Enthusiast
*****
Senior Member
755 posts

Joined: Oct 2008
From: Taiping, Perak


Sorry. Missed the bus.. apa guna AI setup at home? Can tell me when the collapse of crypto?
terradrive
post Jan 29 2025, 11:52 AM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


QUOTE(KitZhai @ Jan 29 2025, 11:09 AM)
So after all? Still chatgpt win ke?
*
if you wanna pay $200 dollars a month for unlimited o1 use lol

plus the ability to run deepseek r1 offline solves privacy issues which is very important for some users
thxxht
post Jan 29 2025, 11:57 AM

Casual
***
Junior Member
368 posts

Joined: Oct 2008
I hope the electricity bill is worth it
KitZhai
post Jan 29 2025, 02:00 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(terradrive @ Jan 29 2025, 12:52 PM)
if you wanna pay $200 dollars a month for unlimited o1 use lol

plus the ability to run deepseek r1 offline solves privacy issues which is very important for some users
*
I am wondering with my house desktop setup, what version of deepseek can it use.
terradrive
post Jan 29 2025, 02:08 PM

RRAAAWWRRRRR
******
Senior Member
1,943 posts

Joined: Apr 2005


QUOTE(KitZhai @ Jan 29 2025, 02:00 PM)
I am wondering with my house desktop setup, what version of deepseek can it use.
*
It depends on memory size. I can only answer for GPU part. If you have 24GB VRAM gpus like rtx3090, rtx4090, you can go for quantized 4 bit deepseek-r1:32b. If you have 12 or 16GB gpu then you can try 14b. 8b for 8GB gpus. But this is for the distilled version quantized 4 bit, not the full bit. full bit uses way more memory.

It can also be run on CPU only, I didn't read much about that but I saw ppl running 70b using 9950X with 64GB system memory but the response is slow. Another built a amd dual epyc system with 768GB memory that managed to run the full bit 671b with 6-8 tokens/s with total system cost at his place around 6000usd.
KitZhai
post Jan 29 2025, 02:10 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(terradrive @ Jan 29 2025, 03:08 PM)
It depends on memory size. I can only answer for GPU part. If you have 24GB VRAM gpus like rtx3090, rtx4090, you can go for quantized 4 bit deepseek-r1:32b. If you have 12 or 16GB gpu then you can try 14b. 8b for 8GB gpus. But this is for the distilled version quantized 4 bit, not the full bit. full bit uses way more memory.

It can also be run on CPU only, I didn't read much about that but I saw ppl running 70b using 9950X with 64GB system memory but the response is slow. Another built a amd dual epyc system with 768GB memory that managed to run the full bit 671b with 6-8 tokens/s with total system cost at his place around 6000usd.
*
Damn my GPU old let ... Rtx 3060 super I think... Can do what? Just subscribe chatgpt?
ycs
post Jan 29 2025, 02:20 PM

MEMBER
*******
Senior Member
4,238 posts

Joined: Jan 2003
From: Selangor



QUOTE(KitZhai @ Jan 29 2025, 02:10 PM)
Damn my GPU old let ... Rtx 3060 super I think... Can do what? Just subscribe chatgpt?
*
gpu requirements
https://apxml.com/posts/system-requirements-deepseek-models
kingkingyyk
post Jan 29 2025, 02:59 PM

10k Club
Group Icon
Elite
15,694 posts

Joined: Mar 2008
QUOTE(KitZhai @ Jan 29 2025, 02:00 PM)
I am wondering with my house desktop setup, what version of deepseek can it use.
*
If you used to use the free version of ChatGPT, this won't beat it. It is already has way more parameters than what home user can (afford to) run.

QUOTE(KitZhai @ Jan 29 2025, 02:10 PM)
Damn my GPU old let ... Rtx 3060 super I think... Can do what? Just subscribe chatgpt?
*
Run 14b models and see which one works best for you. That's it. At this model size, there is nothing that can know everything.

h@ksam
post Jan 29 2025, 03:29 PM

@ is a
*******
Senior Member
3,460 posts

Joined: Nov 2009
From: KL
QUOTE(marfccy @ Jan 28 2025, 03:34 PM)
arguably, this is like back during mining trend? people buy hundreds and thousands of GPUs just to mine the hell out of it plus the insane electric bill monthly
*
korang curi tenaga lagi


KitZhai
post Jan 29 2025, 03:30 PM

Casual
***
Junior Member
382 posts

Joined: Dec 2008
From: /k/



QUOTE(ycs @ Jan 29 2025, 03:20 PM)
I can go for lowest tier sad
frequency
post Jan 29 2025, 03:45 PM

我要挑战十个!
*******
Senior Member
2,567 posts

Joined: Jan 2003


Tried the 7B distilled model on open webui ollama
not better than phi3

perhaps to try on text2sql later
hellothere131495
post Jan 29 2025, 04:59 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(KitZhai @ Jan 29 2025, 01:11 AM)
Its free right? Then what point of using the smaller version compared to 671b?
*
Yes it's free and open source. The point of using smaller version is because 671b needs a very powerful machine to run it. You need a football field size data center, lots of Nvidia H100 GPUs cluster to power the model. It's impossible for normal people to run it, even you buy the best PC you still can't run it.

So, for us, we can only run the smaller models like 8b, 14b, or 32b.
hellothere131495
post Jan 29 2025, 05:03 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(KitZhai @ Jan 29 2025, 09:40 AM)
But with full bit, what is the performance compare to chatgpt? Same or even higher than chatgpt?
*
comparable to chatgpt o1. Even slightly better than chatgpt a bit.
hellothere131495
post Jan 29 2025, 05:11 PM

Casual
***
Junior Member
473 posts

Joined: Sep 2019
QUOTE(KitZhai @ Jan 29 2025, 02:10 PM)
Damn my GPU old let ... Rtx 3060 super I think... Can do what? Just subscribe chatgpt?
*
If you're an AI user, then just use the chatgpt or the deepseek website. The downloaded 4 bit smaller models ain't going to perform better than the chatgpt website or deepseek website, which are running full bit model, with giant parameters.

Those 4 bit models are meant for production use like when you want to build a chatbot and sell to company. You won't ask the company to subscribe to chatgpt right? You need to provide them a full standalone working chatbot.

Or, when you are working with sensitive data, then you need to download open source models and use it privately.

Or, you are AI specialist. You do research in AI. You download the models to study them.

Or, you are simply just a hobbyist or enthusiast in Natural Language Processing and just want to have fun with the state of the art technical things in AI.
alfiejr
post Jan 29 2025, 05:15 PM

Gaming~
******
Senior Member
1,294 posts

Joined: Feb 2012
From: Taman Rasa Sayang, Cheras


Just used their app, I would say it's pretty good for a free version so thumbs up for me 👍
zoozul
post Jan 30 2025, 10:06 AM

Getting Started
**
Junior Member
98 posts

Joined: Jan 2023
QUOTE(GagalLand @ Jan 28 2025, 02:29 PM)
Obviously you don't know AI at all

If DeepSeek is sampah

All AI related tech giants won't cirit-birit in a row
*

Orang sampah tau tak DeepSeek. Obviously is a gimmick distrub share market. puke.gif
James8899
post Jan 30 2025, 10:24 AM

Getting Started
**
Junior Member
139 posts

Joined: Feb 2019
DeepSeek's chatbot achieves 17% accuracy, trails Western rivals in NewsGuard audit

https://ca.finance.yahoo.com/news/deepseeks...-144534774.html

This post has been edited by James8899: Jan 30 2025, 10:24 AM
GagalLand
post Jan 30 2025, 11:57 AM

Getting Started
**
Junior Member
239 posts

Joined: May 2022

Gimmick?

Kesian

Uneducated but arrogant ppl are really sampah

QUOTE(zoozul @ Jan 30 2025, 10:06 AM)
Orang sampah tau tak DeepSeek. Obviously is a gimmick distrub share market. puke.gif
*
zoozul
post Jan 30 2025, 12:37 PM

Getting Started
**
Junior Member
98 posts

Joined: Jan 2023
QUOTE(GagalLand @ Jan 30 2025, 11:57 AM)
Gimmick?

Kesian

Uneducated but arrogant ppl are really sampah
*
Orang sampah macam tak peduli.
yvliew
post Jan 30 2025, 12:38 PM

Listening to IU - Strawberry Moon
*******
Senior Member
3,918 posts

Joined: Nov 2004
My mac mini can only run deepseek r1 14B model. not bad juga.
yushin
post Jan 30 2025, 05:38 PM

Look at all my stars!!
*******
Senior Member
3,329 posts

Joined: Jan 2003
From: Selangor


I wonder what level of deepseek I can run with 2xxeon6133 + 512Gb ram + 2xRTX4090.


 

Change to:
| Lo-Fi Version
0.0373sec    0.64    5 queries    GZIP Disabled
Time is now: 18th December 2025 - 10:33 PM