Open-source Deepseek R1 dethrones commercial AI, now allegedly being hit by cyberattack
DeepSeek hit with large-scale cyberattack, says it's limiting registrations
DeepSeek on Monday said it would temporarily limit user registrations "due to large-scale malicious attacks" on its services.Hayden Field (CNBC)
like this
βπ-ππ
in reply to JOMusic • • •xmunk
in reply to βπ-ππ • • •Korhaka
in reply to xmunk • • •T156
in reply to Korhaka • • •deepseek-ai/DeepSeek-R1 Β· Hugging Face
huggingface.coHappyTimeHarry
in reply to JOMusic • • •Do they not know it works offline too?
I noticed chatgpt today being pretty slow compared to the local deepseek I have running which is pretty sad since my computer is about a bajillion times less powerful
Rogue
in reply to HappyTimeHarry • • •JustARaccoon
in reply to Rogue • • •Rogue
in reply to JustARaccoon • • •Thanks
Any recommendations for communities to learn more?
Frustratingly Their setup guide is terrible. Eventually managed to get it running. Downloaded a model and only after it download did it inform me I didn't have enough RAM to run it. Something it could have known before the slow download process. Then discovered my GPU isn't supported. And running it on a CPU is painfully slow. I'm using an AMD 6700 XT and the minimum listed is 6800 github.com/ollama/ollama/blob/β¦
ollama/docs/gpu.md at main Β· ollama/ollama
GitHubJustARaccoon
in reply to Rogue • • •π‘ Home | Open WebUI
docs.openwebui.comRogue
in reply to JustARaccoon • • •Thanks, I did get both setup with Docker, my frustration was neither ollama or open-webui included instructions on how to setup both together.
In my opinion setup instructions should guide you to a usable setup. It's a missed opportunity not to include a
docker-compose.ymlconnecting the two. Is anyone really using ollama without a UI?JustARaccoon
in reply to Rogue • • •blakenong
in reply to JOMusic • • •Corkyskog
in reply to blakenong • • •blakenong
in reply to Corkyskog • • •No. It literally cannot count the number of R letters in strawberry. It says 2, there are 3. ChatGPT had this problem, but it seems it is fixed. However if you say βare you sure?β It says 2 again.
Ask ChatGPT to make an image of a cat without a tail. Impossible. Odd, I know, but one of those weird AI issues
SoftestSapphic
in reply to blakenong • • •Because there aren't enough pictures of tail-less cats out there to train on.
It's literally impossible for it to give you a cat with no tail because it can't find enough to copy and ends up regurgitating cats with tails.
Same for a glass of water spilling over, it can't show you an overfilled glass of water because there aren't enough pictures available for it to copy.
This is why telling a chatbot to generate a picture for you will never be a real replacement for an artist who can draw what you ask them to.
blakenong
in reply to SoftestSapphic • • •Oh, thatβs another good test. It definitely failed.
There are lots of Manx photos though.
Manx images: duckduckgo.com/?q=manx&iax=imaβ¦
JustARaccoon
in reply to SoftestSapphic • • •SoftestSapphic
in reply to JustARaccoon • • •It searches the internet for cats without tails and then generates an image from a summary of what it finds, which contains more cats with tails than without.
That's how this Machine Learning progam works
FatCrab
in reply to SoftestSapphic • • •SoftestSapphic
in reply to FatCrab • • •So what training data does it use?
They found data to train it that isn't just the open internet?
FatCrab
in reply to SoftestSapphic • • •SoftestSapphic
in reply to FatCrab • • •Oh ok so training data doesn't matter?
It can generate any requested image without ever being trained?
Or does data not matter when it makes your agument invalid?
Tell me how you moving the bar proves that AI is more intelligent than the sum of its parts?
FatCrab
in reply to SoftestSapphic • • •Kogasa
in reply to SoftestSapphic • • •SoftestSapphic
in reply to Kogasa • • •Kogasa
in reply to SoftestSapphic • • •SoftestSapphic
in reply to Kogasa • • •And I care about "how" it works and "what" data it uses because I don't have to walk on eggshells to preserve the sanctity of an autocomplete software
You need to curb your pathetic ego and really think hard about how feeding the open internet to an ML program with a LLM slapped onto it is actually any more useful than the sum of its parts.
Kogasa
in reply to SoftestSapphic • • •ikt
in reply to Kogasa • • •LLM Visualization
bbycroft.netvrighter
in reply to SoftestSapphic • • •SoftestSapphic
in reply to vrighter • • •vrighter
in reply to SoftestSapphic • • •Kuvwert
in reply to vrighter • • •Corkyskog
in reply to blakenong • • •blakenong
in reply to Corkyskog • • •Not trolling you at all:
lemmy.world/comment/14735060
Kuvwert
in reply to Corkyskog • • •Non thinking prediction models can't count the r's in strawberry due to the nature of tokenization.
However openai o1 and deep seek r1 can both reliably do it correctly
Kuvwert
in reply to blakenong • • •blakenong
in reply to Kuvwert • • •Screenshots please
lemmy.world/comment/14735060
rdri
in reply to blakenong • • •lemmy.world/comment/14754440
rdri
2025-01-29 03:40:50
blakenong
in reply to rdri • • •rdri
in reply to blakenong • • •It didn't, I just wanted a short reply. Though it failed when I asked again in the same chat. But when asked to split the word to 2 parts it became sure that the correct answer is 3.
Here is the fun part
Certainly! Let's split "strawberry" into two equal parts and count the 'r's in each:
Letters: S, T, R, A, W, B
'r's: 1 (at position 3)
Letters: E, R, R, Y
'r's: 2 (at positions 2 and 3)
Now, summing them: 1 (first half) + 2 (second half) = 3 'r's.
Wait, this contradicts my earlier correction. It seems I made another mistake. Let's carefully recount:
'r's: 1 (at position 3)
'r's: 1 (only one 'r' at position 2; the third letter is 'r' but already counted in the first half).
Wait, noβthis is getting confusing. Let's re-examine the entire word:
Full word: S, T, R, A, W, B, E, R, R, Y
'r's: Positions 3, 8, and 9 β 3 'r's.
I apologize for the back-and-forth. The correct total is 3 'r's in "strawberry." Thank you for your patience as I worked through this! π
blakenong
in reply to rdri • • •Pieisawesome
in reply to blakenong • • •Itβs because LLMs donβt work with letters. They work with tokens that are converted to vectors.
They literally donβt see the word βstrawberryβ in order to count the letters.
Splitting the letter probably separates them into individual tokens
ikt
in reply to blakenong • • •feel free to ask Google/Bing/Your favourite search engine to do the same π
blakenong
in reply to ikt • • •Kuvwert
in reply to blakenong • • •ibb.co/wVNsn5H
ibb.co/HpK5G5Pp
ibb.co/sp1wGMFb
ibb.co/4wyKhkRH
ibb.co/WpBTZPRm
ibb.co/0yP73j6G
Note that my tests were via groq and the r1 70B distilled llama variant (the 2nd smartest version afaik)
Edit 1:
Incidentally... I propositioned a coworker to answer the same question. This is the summarized conversation I had:
Me: "Hey Billy, can you answer a question? in under 3 seconds answer my following question"
Billy: "sure"
Me: "How many As are in abracadabra 3.2.1"
Billy: "4" (answered in less than 3 seconds)
Me: "nope"
I'm gonna poll the office and see how many people get it right with the same opportunity the ai had.
Edit 2:
The second coworker said "6" in about 5 seconds
Edit 3:
Third coworker said 4, in 3 seconds
Edit 4:
I asked two more people and one of them got it right... But I'm 60% sure she heard me asking the previous employee, but if she didnt we're at 1/5
In probably done with this game for the day.
I'm pretty flabbergasted with the results of my very unscientific experiment, but now I can say (with a mountain of anecdotal juice) that with letter counting, R1 70b is wildly faster and more accurate than humans .
Screenshot-2025-01-29-122217 hosted at ImgBB
ImgBBrdri
in reply to blakenong • • •blakenong
in reply to rdri • • •CouncilOfFriends
in reply to JOMusic • • •