DeepSeek Before V4: Culture, Organization, and Liang Wenfeng's Unique Goals

Original: V4 发布前的 DeepSeek:特质、组织和梁文锋的独特目标 (WeChat/晚点LatePost, ~March 2026)

Summary by Lab Index. This is a condensed English summary of the major claims in the original Chinese article.


Key Departures

Several core DeepSeek members left between late 2025 and early 2026:

Despite these departures, the article emphasizes that more people chose to stay. There has been no group-level attrition.

V4 Status

As of writing (~March 2026), DeepSeek V4 has not yet been officially released. A small-parameter version was given to open-source framework communities around January 2026 for adaptation work. The optimistic timeline was a mid-February release (around Chinese New Year), but the article reports V4 may launch in April 2026.

Compensation and Valuation Challenges

DeepSeek has never raised external funding and has no established company valuation. As competitors like MiniMax and Zhipu (Z.ai) go public with rising stock prices, and Moonshot AI (Kimi) and StepFun prepare IPOs, DeepSeek employees are questioning the value of their equity/option agreements. Liang Wenfeng has recently begun working to establish a company valuation and provide more certainty to team members.

When Liang did briefly meet investors in 2023, he proposed an unusual term: a return cap for investors, similar to OpenAI's arrangement with Microsoft. No institution invested.

The "No Overtime" Culture

DeepSeek is described as the only core AI lab globally that does not overwork. While engineers at Google, OpenAI, xAI, and ByteDance work 70–80 hours per week:

The company provides free after-work benefits like sports courses and gym/venue reimbursements.

Organization: Flat, Cross-functional, ~200 People

Hiring Profile

Before 2025, DeepSeek almost never hired experienced professionals, preferring new graduates and converting interns. Analysis of 172 researchers who contributed to three generations of models (LLM, V2, V3/R1) showed: over 70% held only bachelor's or master's degrees, and over 70% were under 30 years old.

Research Focus Since R1

Rather than capitalizing on V3/R1's viral success with flashy releases, DeepSeek continued along three lines:

  1. Efficiency optimization — squeezing maximum intelligence per unit of GPU compute. This includes the open-source week infrastructure releases (inference kernels, communication libraries, matrix multiplication libraries, data processing frameworks), NSA (Native Sparse Attention), DSA (Dynamic Sparse Attention), and even replacing CUDA/Triton with the Peking University-developed TileLang at the operator level.
  2. Architecture innovations — mHC (Manifold-Constrained Hyper-Connections) for training stability, and Engram for building long-term memory outside the model. mHC is widely expected to be used in V4.
  3. "Non-mainstream" explorations — DeepSeek-OCR (converting text to images before feeding to the model), continuous learning, autonomous learning, and consultations with neuroscience/brain science advisors to explore mechanisms closer to the human brain.

What DeepSeek Is NOT Doing

Liang Wenfeng's Unique AGI Goals

Beyond pursuing intelligence ceiling, Liang prioritizes two things most labs do not:

  1. Building on domestic (Chinese) chip ecosystems — adapting models for domestic GPUs, using Chinese-originated open-source tools (TileLang over Triton), and designing data formats for next-generation domestic chips.
  2. "Original innovation" — pursuing directions that big companies or other startups won't try: the Janus unified multimodal series, the Prover formal verification series, OCR research, continuous learning, and brain-inspired approaches.

Outlook

The article concludes that V4, when released, will likely be the strongest open-source model but will not be overwhelmingly dominant, as "strong" has become increasingly context-dependent across different use cases. DeepSeek faces the tension between Liang's emphasis on original exploration and the industry's pressure to simply "stay the strongest."

A person close to DeepSeek says: "Those who stay still have some idealism." Another observer notes: "Only when more companies like DeepSeek appear will Chinese technology have a chance to go from 'replication' to leading."