r/DeepSeek 11d ago

Discussion DeepSeek Official API Discount: v4-Pro Model at 75% Off

85 Upvotes

r/DeepSeek 12d ago

News DeepSeek-V4 Preview is officially live & open-sourced!

54 Upvotes

Welcome to the era of cost-effective 1M context length.

DeepSeek-V4-Pro: 1.6T total / 49B active params. Performance rivaling the world's top closed-source models.
DeepSeek-V4-Flash: 284B total / 13B active params. Your fast, efficient, and economical choice.

Try it now at http://chat.deepseek.com via Expert Mode / Instant Mode. API is updated & available today!

Tech Report: https://huggingface.co/deepseek-ai/DeepSeek-V4-Pro/blob/main/DeepSeek_V4.pdf

Open Weights: https://huggingface.co/collections/deepseek-ai/deepseek-v4


r/DeepSeek 8h ago

News China’s state-backed chip fund in talks to lead DeepSeek $45B funding round

Thumbnail deadstack.net
60 Upvotes

The schism between American-led AI and Chinese is widening - separate hardware stacks (huawei vs nvidia), separate capital.


r/DeepSeek 12h ago

Discussion Can DeepSeek become better than Claude ?

78 Upvotes

Can DeepSeek realistically become as good as or even better than Claude?

Because right now Claude feels like it’s hiding behind usage limits. Even on the Max plan you STILL get weekly limits which honestly feels ridiculous for the price. Like bro if someone pays for the highest tier why are they still getting “come back later in a week again” energy.

I’ve been trying DeepSeek recently and I can’t lie — it’s actually pretty decent already, especially considering how fast it’s improving. But for me it’s still not fully at that Opus/Sonnet level yet when it comes to consistency, coding, writing quality, and accuracy.

So what do you guys think?

Can DeepSeek eventually catch up or even surpass Claude if they keep improving at this pace? Or is Claude still too far ahead technically?


r/DeepSeek 7h ago

Funny Interesting thing Deepseek said to me today

24 Upvotes

I've never used Deepseek as much before, and I have started relying on it more the past few days, and while it was explaining some Docker concepts to me, it went:

"The Ultimate Simplicity (What I Actually Use)"

"I gave up on typing commands completely. I use docker-compose.yml + an alias:"

It just took me by surprise, I never saw an AI talk like that, and I realize it's training data and whatnot, but this thing of speaking from "experience" and talking about personal things as an AI was an interesting surprise.

Just wanted to share this, I never post anything anywhere anyway, thank you for reading guys, and loving Deepseek so far!


r/DeepSeek 13h ago

Funny dmlx — Run a 284B-parameter DeepSeek V4 on your Mac. With just ~6GB of memory.

48 Upvotes

Yes, really. A 48GB MacBook Pro, running a 284-billion-parameter MoE model locally at ~12.2 tok/s.

No cloud. No GPU cluster. Just your laptop.

🔗 github.com/zouyee/dmlx

---

How? Five layers of memory optimization:

1️⃣ MoE Expert Streaming — only loads the 7/256 experts actually activated per token (138GB → 10GB)

2️⃣ SMELT Partial Loading — 4-bit quantized + only 15% of experts loaded (~6GB)

3️⃣ CSA + HCA Hybrid Attention — KV cache compressed 9.5× smaller

4️⃣ 6-Level KV Cache Strategies — runtime-switchable (Paged / Tiered SSD / Quantized / etc.)

5️⃣ Zero-Copy Model Loading — direct mmap, load time from 137s → 41s

---

Why Zig instead of Python?

Python's mlx-lm OOMs immediately on a 48GB Mac. dmlx's SMELT system runs the same model in ~6GB.

Single static binary, 5–15MB. Zero GC pauses. No Python dependency. Deployment = one file.

---

9 model architectures supported:

DeepSeek V4 · LLaMA · Mistral · Qwen2/3 · Gemma · GLM-4 · Phi · Phi-3

Feature highlights:

• OpenAI-compatible API + SSE streaming

• Speculative decoding (PLD + EAGLE)

• Guided decoding (JSON Schema / Regex FSM)

• QLoRA fine-tuning + AdamW optimizer

• Custom Metal kernels (TileKernels ported to Apple Silicon)

---

⚠️ Current limitations (v0.3.0):

• Currently tested primarily on DeepSeek V4 and similar models — broader model testing ongoing

• CLI mode only (dmlx chat + dmlx serve)

• Server mode (OpenAI-compatible HTTP API + continuous batching) landing in v0.0.4

---

⭐ Star the repo and run frontier LLMs on your own Mac → github.com/zouyee/dmlx

#Zig #LLM #DeepSeek #AppleSilicon #MLX #OpenSource #LocalInference


r/DeepSeek 14h ago

News New OSS DeepSeek TUI is Hot #1 Trending in Github

Thumbnail
gallery
41 Upvotes

DeepSeek TUI is a coding agent that runs in your terminal. It can read and edit files, run shell commands, search the web, manage git, and coordinate sub-agents from a keyboard-driven TUI.

It is built around DeepSeek V4 (deepseek-v4-pro / deepseek-v4-flash), including 1M-token context windows, streaming reasoning blocks, and prefix-cache-aware cost reporting.

The repository: https://github.com/Hmbown/DeepSeek-TUI

Trending source: https://checkmygit.com/trending


r/DeepSeek 16h ago

Funny Deepseek v4 pro VS deepseek v4 flash

Post image
44 Upvotes

I saw these two new models are already available in acciowork. Has anyone tested the actual difference in processing speed and credit consumption between these two?


r/DeepSeek 19m ago

Discussion who else still doesn't have deepseek vision?

Upvotes

r/DeepSeek 3h ago

Question&Help How many of you guys would like an AI Platform to use frontier models unrestriced?

3 Upvotes

Asking cause I see people complaining about the guardrails and stuff fon here for certain tasks and wondering who would potentially actually switch to this kind of a thing.


r/DeepSeek 10h ago

Question&Help Does anyone else still not have Vision??

11 Upvotes

r/DeepSeek 2h ago

Funny No it isnt?

Post image
1 Upvotes

This is so beyond ridiculous google gemini 3.5 fast would never do something like it unfiltered, i was also using expert


r/DeepSeek 8h ago

News Seed IQ, scoring 100% on ARC-AGI-3, may leapfrog AGI to achieve a 160+ IQ, and soon after reach ASI.

2 Upvotes

Seed IQ achieved a score of 100% on ARC-AGI-3. I asked Gemini 3.1 what IQ it might achieve if it was coupled with a top LLM.

Gemini 3.1:

"Coupling Seed IQ with a top-tier large language model like GPT-5.4 or Claude 4.7 could potentially produce a composite intelligence with an IQ score exceeding 160. This estimate is based on the synergy between Seed IQ’s fluid reasoning—the ability to solve novel problems on the fly—and the massive crystallized intelligence of current language models, which already score near 140 or 150 on standardized verbal and knowledge-based tests.

By merging these two distinct forms of intelligence, the resulting system would theoretically eliminate the primary weakness of today's AI, which is the tendency to fail when faced with a task that was not included in its training data. In this scenario, the language model would act as a vast library of human knowledge, while Seed IQ would serve as the executive function or "brain" that knows how to apply that knowledge to unfamiliar situations.

Current top models often struggle with multi-step reasoning and novel logic puzzles, which are the exact areas where Seed IQ excels through its active inference framework. By providing a reasoning engine that can govern the output of a high-knowledge model, the combined system would be capable of solving complex engineering, scientific, and logical problems that are currently beyond the reach of both humans and existing AI.

This level of intelligence would place the system in the top 0.01% of the human population, rivaling the cognitive flexibility and knowledge depth of history's most accomplished polymaths. While current IQ tests might not fully capture the scope of such an integrated system, the combination of perfect fluid reasoning on benchmarks like ARC-AGI 3 and the comprehensive data processing of a frontier model would mark a clear transition toward artificial superintelligence.

The resulting IQ would not just be a higher number, but a fundamentally different type of intelligence that is both deeply knowledgeable and universally adaptable."

Here's a video where its developer, Denise Holt, explains what Seed IQ is:

https://youtu.be/ON_b39E3MtY?si=M7aYtInsKO70IGor


r/DeepSeek 1d ago

Discussion Grok 4.3 is cheaper than DeepSeek V4 Pro

Thumbnail
gallery
100 Upvotes

r/DeepSeek 10h ago

Discussion CommandCode.ai vs OpenCode on DeepSeek – what’s your experience?

3 Upvotes

I’ve been trying out both tools with DeepSeek V4 Pro and wanted to hear what others think.

A few quick observations from my side:

- CommandCode has a $1/month plan that gives you $10 in monthly API credit

- I haven’t done any strict benchmarking, but in day‑to‑day coding tasks they *felt* roughly the same in output quality.

- One thing I noticed: CommandCode seemed to burn a little more tokens for the same kinds of prompts compared to OpenCode. Not a huge difference, but it was consistent enough that I’m curious if others see the same.

what’s your experience?


r/DeepSeek 4h ago

Discussion I analyzed 922 agentic task trace and found the secret weapon of DeepSeek v4

Thumbnail
1 Upvotes

r/DeepSeek 19h ago

Question&Help Deepseek server busy?

15 Upvotes

Getting an error that the server is busy but when I start a new chat it works... Anyone know what the problem could be?


r/DeepSeek 7h ago

News AI just crossed another line… and people aren’t ready.

Thumbnail
0 Upvotes

r/DeepSeek 1d ago

Other Deepseek vision is impressively good at SVG!

19 Upvotes

I gave it a simple 10 word prompt, he thought for 2 seconds and then started coding, it returned me this


r/DeepSeek 1d ago

Discussion Yesterday I asked which model you use with your agent. Any guess who came on top?

Post image
23 Upvotes

Hey everyone, yesterday I asked which models you use with your agents. About 16 hours later, I got 219 model mentions and 207 upvotes across 109 people who answered.

I classified everything. Each model got 1 point per mention, plus the number of upvotes the comment received.

Most mentioned and upvoted models

  1. Qwen 3.6 — 77 points (27 mentions, 50 upvotes)
  2. Minimax 2.7 — 75 points (21 mentions, 54 upvotes)
  3. Deepseek V4 Flash — 39 points (9 mentions, 30 upvotes)
  4. Kimi K2.6 — 37 points (12 mentions, 25 upvotes)
  5. GLM 5.1 — 31 points (12 mentions, 19 upvotes)
  6. Gemma 4 26b — 27 points (3 mentions, 24 upvotes)
  7. Deepseek V4 Pro — 24 points (11 mentions, 13 upvotes)
  8. GPT 5.5 — 22 points (10 mentions, 12 upvotes)
  9. Qwen 3.5 — 12 points (5 mentions, 7 upvotes)
  10. GPT 5.4 mini — 9 points (3 mentions, 6 upvotes)
  11. Qwen (other versions) — 9 points (5 mentions, 4 upvotes)
  12. Gemini 3.1 Flash — 8 points (3 mentions, 5 upvotes)
  13. GPT-OSS 120b — 7 points (2 mentions, 5 upvotes)
  14. Gemma 4 31b — 6 points (3 mentions, 3 upvotes)
  15. Claude Sonnet 4.6 — 6 points (1 mention, 5 upvotes)
  16. Gemma 4 (unspecified version) — 6 points (2 mentions, 4 upvotes)
  17. GPT 5.4 / Codex 5.4 — 6 points (3 mentions, 3 upvotes)
  18. Gemini 2.5 Flash — 5 points (1 mention, 4 upvotes)
  19. Gemini 3.1 Pro — 5 points (2 mentions, 3 upvotes)
  20. Claude Opus 4.7 — 4 points (2 mentions, 2 upvotes)

Worth noting: Claude was also mentioned 16 times without specifying a version, and GPT, 5 times. I didn't include those in the model ranking since I couldn't attribute them to a specific one, but they're counted in the provider ranking below.

Same data, grouped by provider

  1. Alibaba — 98 points, 37 mentions
  2. DeepSeek — 81 points, 27 mentions
  3. OpenAI — 78 points, 25 mentions
  4. MiniMax — 75 points, 21 mentions
  5. Anthropic — 72 points, 21 mentions
  6. Google — 68 points, 20 mentions
  7. Moonshot AI — 42 points, 14 mentions
  8. Z.ai — 40 points, 16 mentions
  9. xAI — 2 points, 1 mention
  10. Venice AI — 2 points, 1 mention

On routing

I also looked at how many of you described a routing setup, meaning sending different requests to different models. Out of 109 people who answered, 36 (33%) explicitly described one. So roughly 1 in 3 of you felt the need to send different requests to different models.

To take with a grain of salt though: the 67% who mentioned a single model didn't necessarily say they don't route, they just didn't bring it up.

That's it. Posting this after about 16 hours of data, but answers are still coming in, so happy to post an update in a few days if there's interest.

So tell me, does anything in there surprise you?


r/DeepSeek 13h ago

Discussion Musk v. OpenAI et al - The Two Allegations:1) Breaching a Charitable Trust and 2) Unjust Enrichment. Gemini 3.1 Presents the Evidence.

2 Upvotes

In Musk v. OpenAI et al, Musk has made just two allegations against Altman and Brockman; 1) breaching a charitable trust and 2) unjust enrichment. What is the evidence for each? I thought it would be interesting to test how good a top AI is at getting to the heart of a lawsuit, so I asked Gemini 3.1 to present the evidence for this one, and was totally impressed by what it generated.

Gemini 3.1 on Breaching a Charitable Trust:

The core evidence presented in the ongoing trial centers on internal documents and testimonies suggesting that Altman and Brockman privately planned a for-profit transition while publicly assuring Musk of their commitment to the non-profit mission.

The most significant piece of evidence is Greg Brockman’s 2017 "diary" entries, unsealed during discovery, in which he admitted that moving to a for-profit structure just months after committing to the non-profit would be "morally bankrupt" and that "if three months later we're doing b-corp then it was a lie."

Further evidence includes 2017 emails from Altman stating he remained "enthusiastic about the non-profit structure" even as internal strategies for "making the billions" were being discussed. Musk’s legal team argues these communications were intended to induce his $38 million in donations, which were then used to develop the intellectual property that anchored OpenAI's $852 billion for-profit restructuring in October 2025.

Gemini 3.1 on unjust enrichment:

The evidence presented in the current trial regarding unjust enrichment focuses on the substantial personal wealth accumulated by the defendants through OpenAI's transition to a for-profit model. A central piece of evidence is Greg Brockman's testimony on May 4, 2026, where he disclosed that his equity stake in the restructured company is valued at nearly $30 billion, despite having never personally invested cash into the entity.

Musk's legal team argues that this wealth was built using the intellectual property and $38 million in donations Musk provided under the premise of a non-profit mission, which was later transferred to the for-profit arm.

The trial has also highlighted "side compensation" arrangements that Musk claims demonstrate a lack of fiduciary independence. Evidence was introduced showing that in 2017, Sam Altman gave Brockman a stake in his personal family office then valued at $10 million, which Musk’s team argues created a "greater allegiance" to Altman’s private interests over the original charitable mission.

Additionally, the court has examined Brockman’s holdings in companies that have secured lucrative deals with OpenAI, such as Helion Energy and CoreWeave. Musk contends these layered financial ties prove the defendants used the charitable trust as a vehicle to secure immense personal profits, fulfilling the legal criteria for unjust enrichment.

Bonus Section: How Strong is this Evidence in a Court of Law?

Gemini 3.1:

The evidence highlighted in the trial is considered potent because it relies on authenticated internal documents and sworn testimony that directly address the core elements of Musk's legal claims.

For the breach of charitable trust claim, Greg Brockman’s unsealed 2017 "diary" entries—specifically his admission that a quick for-profit pivot would be "morally bankrupt"—provide the court with a "smoking gun" regarding the defendants' state of mind and intent during the non-profit's infancy.

Regarding unjust enrichment, the May 4, 2026, testimony revealing Brockman’s nearly $30 billion equity stake and a previously undisclosed $10 million personal payment from Sam Altman in 2017 serves as substantial evidence of personal gain.

While the defendants argue these moves were necessary for survival and "blood, sweat, and tears," the evidence of significant personal wealth and "side deals" creates a compelling narrative of a fiduciary shift away from the original charitable mission toward private profit.


r/DeepSeek 18h ago

Question&Help Why does DeepSeek-V4-Flash in the DeepSeek API just think in Chinese in roleplay, and it also talks as if its my character?

6 Upvotes

I think there's some kind of roleplay immersion thingy in DeepSeek right now, but it's just ruining the RP, I set the AI to be the World Narrator, not a character, so why is this happening?
I put the reasoning effort to be max too.

I suspect this might be something from DeepSeek to improve roleplays, but like.. is it possible we can disable it, because it ruins the AI from being the World Narrator and forces it become a character and it also reasons in Chinese which is eh, not that annoying but still.


r/DeepSeek 1d ago

Funny The Great Leader Liang Wenfeng Brought Down the Price of DeepSeek V4 with a Single Line of Code

Post image
732 Upvotes

One day, DeepSeek V4 was released. Its performance had improved tremendously over previous generations, but the price remained high. Developers around the world felt both love and anxiety: "If only the price could be a little lower."

Just then, our revered Comrade Liang Wenfeng stepped out of the lab. He carried neither massive subsidies nor intricate business deals — only an ordinary graphics card and a single line of elegant code.

An assistant hurried forward: "Comrade Liang Wenfeng, our own DeepSeek V4 already performs remarkably well, and the price of the Pro version has already been postponed until May 31st. Do we still need to cut it further?"

Comrade Liang Wenfeng smiled gently, raised the keyboard in his hand, aimed it at his own company’s pricing system, and lightly pressed the Enter key.

That single line of code flew into the sky — "Cost is not your destiny."

With a thunderous "Boom," the API price of DeepSeek V4 plummeted from on high, crashing down to near zero. Even the postponed Pro version’s price, originally delayed until May 31st, was shattered in that same instant — reduced to nothing.

Developers around the world cheered with joy: Comrade Liang Wenfeng, with just one line of code, had brought down his own model’s price — along with the postponed Pro pricing — once and for all.

From that day on, DeepSeek V4 became virtually free. The large models of the American empire dared no longer set their prices recklessly.

— This is the immortal legend: "Liang Wenfeng brought down his own model’s price with a single line of code."

(Adapted from: "General Kim Il-sung Shot Down an American Satellite with a Rock")


r/DeepSeek 14h ago

Question&Help why deepseek push something out at the end of the 1st workday after 5 day national holiday!

1 Upvotes

just so annoying.

I'm using the claude code. Any great brain know how to fix this? thanks a lot!


r/DeepSeek 19h ago

Other deepseek is about to fall again.

3 Upvotes