StepFun Step 3.5 Flash
B Tier · 7.8/10
StepFun's (China) agent-focused open-weight model -- Step 3.5 Flash launched 2026-02-01. 196B sparse MoE, ~11B active. Benchmarks slightly ahead of DeepSeek V3.2 at over 3x smaller total size. Step 3 (321B / 38B active, Apache 2.0) and Step3-VL-10B multimodal also in the family
Score Breakdown
The Good and the Bad
What we like
- +Step 3.5 Flash at 196B total / 11B active beats DeepSeek V3.2 on several mainstream benchmarks at 3x smaller total parameter count -- an efficient architecture result that suggests StepFun's MoE routing is unusually well-tuned for agentic workloads
- +Agent-focused tuning explicitly -- tool use, function calling, and multi-step planning are prioritized in training objectives, not just general chat. For teams building agent systems with Chinese open-weight models, Step 3.5 Flash is a stronger starting point than general-purpose DeepSeek or Qwen base models
- +Fills an important gap in the site's Chinese open-weight roster (already have DeepSeek, Qwen, GLM, Kimi, MiniMax -- StepFun was the notable absence). StepFun has real traction in the China market and growing Western mindshare through OpenRouter
- +Step3-VL-10B multimodal variant provides a small-footprint vision option in the same family -- 10B size is realistic for consumer GPUs and the vision tower is competitive with Qwen3-VL-10B on standard benchmarks
What could be better
- −Smaller Western community than DeepSeek or Qwen -- fewer tutorials, quants, and third-party fine-tunes. Production adoption outside China is still early, which means less Stack Overflow coverage when you hit edge cases
- −PRC content filters apply -- same category of refusals on politically sensitive topics as DeepSeek, Qwen, GLM. Not a dealbreaker for most commercial work but worth flagging
- −Step 3.5 Flash's benchmark wins are strongest on Chinese-heavy evaluations and lighter on purely-English benchmarks versus comparably-sized Western models. English writing quality trails Claude / Mistral / Llama meaningfully
- −Release cadence is fast (Step 3 → Step 3.5 in months) but model naming / versioning is confusing -- expect churn in which specific variant is the 'current best' through Q2 2026
Pricing
Self-hosted (Apache 2.0)
- ✓Apache 2.0 license on Step 3 + Step3-VL-10B
- ✓Weights on Hugging Face (stepfun-ai/step3)
- ✓Agent-focused tuning -- tool use, planning, multi-step reasoning
API (StepFun / OpenRouter)
- ✓Available on OpenRouter and select hosted providers
- ✓StepFun direct API for enterprise customers
- ✓Pricing in line with other mid-tier Chinese open-weight models
System Requirements
Hardware needed to self-host. Min = smallest viable setup (usually heavy quantization). Max = full-precision / production-grade.
| Model variant | Min | Max |
|---|---|---|
| Step 3.5 Flash (196B total / 11B active MoE)Agent-focused tuning; MoE routing activates ~11B params per token | 64 GB RAM + 16 GB GPU (Q3 offload) | 4× H100 FP8 |
| Step 3 (321B total / 38B active, Apache 2.0) | 128 GB RAM + 24 GB GPU (Q3) | 4× H100 FP8 |
| Step3-VL-10B (multimodal) | 12 GB VRAM Q4 (RTX 3060 / 4070) | 1× A100 40 GB FP16 |
Known Issues
- Step 3.5 Flash was released 2026-02-01 with initial quantization issues for Q3 and below -- community Q5 quants are the practical deployment target as of April 2026. Full FP16 weights require significant multi-GPU setup (196B total params even with 11B active)Source: Hugging Face discussions, Reddit r/LocalLLaMA · 2026-02
- Refuses discussion of Tiananmen, Taiwan sovereignty, Xi Jinping, and other politically sensitive topics per PRC regulations -- same content filters as other Chinese open-weight modelsSource: Hugging Face discussions, user reports · 2026-02
Best for
Teams building agent systems on Chinese open-weight foundations who want something other than DeepSeek or Qwen, especially if agentic tool-use is the primary workload. Also good for Chinese-market products where StepFun's domestic tuning advantages matter. And for anyone looking to add diversity to their open-weight evaluation matrix beyond the top-3 Chinese labs.
Not for
English-first creative writing, non-technical users, or workflows that touch politically sensitive topics (PRC content filters apply). Also not the right pick if you need the absolute best Western-community ecosystem -- Llama, Qwen, or DeepSeek will have more third-party tooling.
Our Verdict
StepFun Step 3.5 Flash (February 2026) is the notable Chinese open-weight release that doesn't come from the top-3 labs (DeepSeek, Alibaba/Qwen, Zhipu/GLM). The agent-focused tuning and MoE efficiency (196B total / 11B active beats DeepSeek V3.2 at 3x larger) give it a real niche for teams building tool-use agents on open weights. Western community adoption is still early but StepFun is growing OpenRouter presence and the Step 3.5 Flash release is a credible first step into broader ecosystem relevance. Worth adding to any Chinese-open-weight shortlist alongside DeepSeek and Qwen.
Sources
- StepFun research: Step 3 (accessed 2026-04-17)
- Hugging Face: stepfun-ai/step3 (accessed 2026-04-17)
- aibase news: Step 3.5 Flash (accessed 2026-04-17)
Alternatives to StepFun Step 3.5 Flash
Llama 4 (Meta)
Meta's open-weights flagship family -- Scout (10M context), Maverick (multimodal 400B MoE), Behemoth in preview
Mistral AI
European AI lab with open and commercial models -- Mistral Small 4 (Mar 2026, 119B MoE Apache 2.0 unified model), Medium 3 (Apr 9 2026), and Voxtral TTS (open-source speech, Mar 2026)
DeepSeek
Near-frontier reasoning for pennies on the dollar -- the open-source LLM that made Silicon Valley nervous
Gemma 4 (Google)
Google DeepMind's open-weights model family -- multimodal, 256K context, runs on edge devices
Qwen (Alibaba)
Alibaba's open-weights + API family -- Qwen 3.6-Plus (Mar 30 2026, 1M context + always-on CoT + agentic tool-use), Qwen3.5 Small (2B runs on iPhone, 9B matches 120B-class models), plus Qwen3.5-Omni native multimodal. Apache 2.0 on the open sizes
GLM / Z.ai (Zhipu AI)
Zhipu AI's open-weights family -- GLM-5.1 (launched 2026-04-07) is 744B MoE / 40B active, topped SWE-Bench Pro at 58.4 (beating GPT-5.4 and Claude Opus 4.6), MIT licensed, 200K context. Trained entirely on 100K Huawei Ascend 910B chips -- first frontier model with zero Nvidia in the training stack
Kimi K2.5 (Moonshot)
Moonshot's 1T-parameter MoE open-weights flagship -- best open-source agentic coder, rivals Claude Opus 4.5
Nemotron (Nvidia)
Nvidia's open-weights family -- hybrid Mamba-Transformer MoE architecture, optimized for efficient reasoning on Nvidia hardware
MiniMax M2 / M2.5
MiniMax's open-weights frontier -- first open model to match Claude Opus 4.6 on SWE-Bench at 10-20× lower cost
Falcon (TII)
UAE's Technology Innovation Institute open-weights family -- Falcon 3 optimized for efficient sub-10B deployment on consumer hardware
gpt-oss (OpenAI)
OpenAI's FIRST open-weight models -- gpt-oss-120b (single 80GB GPU, near parity with o4-mini on reasoning) and gpt-oss-20b (runs on 16GB edge devices). Apache 2.0. Launched 2025-08-05. gpt-oss-safeguard ships in 2026 as the safety-tuned variant
IBM Granite 4.0
IBM's enterprise-focused open-weight family -- Granite 4.0 hybrid Mamba-2 + transformer architecture (70-80% memory reduction vs pure transformer), 3B to 32B sizes, Apache 2.0. First open model family to secure ISO 42001 certification. Nano 350M runs on CPU with 8-16GB RAM. 3B Vision variant landed 2026-04-01
Arcee Trinity-Large-Thinking
Arcee AI's US-made open-weight frontier reasoning model -- launched 2026-04-01. 398B total params, ~13B active. Sparse MoE (256 experts, 4 active = 1.56% routing). Apache 2.0, trained from scratch. #2 on PinchBench trailing only Claude 3.5 Opus. ~96% cheaper than Opus-4.6 on agentic tasks
Olmo 3 (AI2)
Allen Institute for AI's fully-open frontier reasoning models -- Olmo 3 family (2025-11-20) includes 7B and 32B sizes, four variants (Base, Think, Instruct, RLZero). Apache 2.0 with fully open data + checkpoints + training logs. Olmo 3-Think 32B matches Qwen3-32B-Thinking at 6x fewer training tokens
AI21 Jamba2
AI21 Labs' hybrid SSM-Transformer (Mamba-style) open-weight family -- Jamba2 launched 2026-01-08. Two sizes: 3B dense (runs on phones / laptops) and Jamba2 Mini MoE (12B active / 52B total). Apache 2.0, 256K context, mid-trained on 500B tokens
Cohere Command A
Cohere's enterprise-multilingual flagship -- 111B params, 256K context, runs on 2x H100. 23 languages. CC-BY-NC 4.0 on weights (research / non-commercial), commercial requires Cohere enterprise contract. Follow-ups: Command A Reasoning + Command A Vision