Step 3.5 Flash
step-3.5-flash is our flagship reasoning model, designed for high-complexity tasks requiring deep logic and fast execution. It features:
- Mixture of Experts Architecture (MoE): Combines a 196B parameter knowledge base with sparse activation (activating around 11B parameters per token) to deliver the logical depth of ultra-large models while keeping inference fast.
- 256K Long Context: Maintains logical consistency when processing massive datasets or long documents, making it ideal for multi-stage reasoning and research workflows.
- Native Agent Capabilities: Excels at tool call orchestration, multi-step problem decomposition, and long-context agent development, making it the preferred foundation for engineering and automation workloads.
- Extreme Efficiency: Optimized for production throughput and cost-effective deployment without compromising on cutting-edge reasoning performance.
Chat Completion Example
The following code demonstrates how to use thestep-3.5-flash model for logical reasoning.
copy
Obtaining Reasoning Content
When StepFun’s reasoning models handle complex problems, they include areasoning field in the output to display the model’s thinking process. Developers can check for the existence of this field to obtain the model’s thinking information.
reasoning field to get the model’s thinking process.
Notes
- JSON Mode Limitation: The current version does not temporarily support JSON mode.
- Error Handling and Logging: A Trace ID is added to model outputs. Please include this ID when reporting any issues with reasoning behavior.