Wed Aug 27 2025

DeepSeek Releases V3.1: Hybrid LLM with Think/Non-Think Modes and 128K Token Context Window

DeepSeek launches V3.1, a hybrid AI model with think/non-think modes, 128K tokens, and open-source under MIT license.
Generative AI

Chinese AI startup DeepSeek has announced the release of DeepSeek-V3.1, a next-generation hybrid large language model that blends advanced reasoning with fast response capabilities. Featuring think/non-think modes, an extended 128K token context window, and agent-ready tool integrations, V3.1 is set to challenge industry leaders by combining efficiency, scalability, and open-source accessibility under the MIT license.

Date & Location:
Beijing, August 21, 2025 — Chinese AI startup DeepSeek has officially launched DeepSeek-V3.1, a groundbreaking hybrid large language model designed to usher in the “agent era.”

Key Features:

Hybrid Inference Architecture — The model supports both Think and Non‑Think modes, allowing developers to toggle between deep reasoning and fast responses within a single model framework.

Improved Reasoning Efficiency — “Think” mode delivers faster answer generation compared to DeepSeek‑R1‑0528, thanks to optimized reasoning performance.

Extended 128K Token Context Window — Enables processing of long documents, extended dialogues, and large-scale data in one shot.

Tool & Agent Enhancements — Strengthened multi-step tool usage and agentic workflows with post-training fine-tuning.

Precision & Hardware Optimization — Supports BF16, FP8, and F32 formats for efficient inference on modern hardware, including upcoming Chinese AI chips.

Open‑Source Release Under MIT License — Both base and full model weights are freely available via Hugging Face repositories.

Immediate Availability & API Pricing Transition — Accessible via web chat, mobile app, and API; API pricing update effective September 6, 2025.

Why It Matters:
DeepSeek-V3.1’s hybrid design and vast context capabilities position it as a serious competitor to closed-source giants like GPT-4. Its open licensing, agent-enhanced performance, and compatibility with new hardware make it especially appealing to developers, researchers, and enterprises seeking cost-effective, flexible AI solutions.

References:

Share:
Comments:
No comments yet!