DeepSeek-V3.1 adds hybrid-inference “Think” and “Non-Think” modes for faster reasoning
Expands long-context window to 128,000 tokens, enabling much longer dialogues
Mixture-of-Experts (MoE) design: ~671B parameters, 37B activated per token
Agent upgrades: lower Think-mode latency, improved tool-calling, API and template support