1. Model Introduction
Qwen3.6-35B-A3B is the first open-weight variant of the Qwen3.6 series developed by Alibaba. Built on direct feedback from the community, Qwen3.6 prioritizes stability and real-world utility, delivering substantial upgrades in agentic coding and thinking preservation. Qwen3.6 features a Gated Delta Networks combined with sparse Mixture-of-Experts architecture (35B total parameters, 3B activated), supporting multimodal inputs (text, image, video) and natively handles context lengths of up to 262,144 tokens, extensible to over 1M tokens. Key Features:- Agentic Coding: Handles frontend workflows and repository-level reasoning with greater fluency and precision
- Thinking Preservation: New option to retain reasoning context from historical messages, streamlining iterative development
- Efficient Hybrid Architecture: Gated Delta Networks + sparse MoE (35B total / 3B active) for high-throughput inference
- Hybrid Reasoning: Thinking mode enabled by default with step-by-step reasoning, can be disabled for direct responses
- Tool Calling: Built-in tool calling support with
qwen3_coderparser - Multi-Token Prediction (MTP): Speculative decoding support for lower latency
- Multimodal: Unified vision-language model supporting text, image, and video inputs
| Model | Weights |
|---|---|
| Qwen3.6-35B-A3B (BF16) | Qwen/Qwen3.6-35B-A3B |
| Qwen3.6-35B-A3B (FP8) | Qwen/Qwen3.6-35B-A3B-FP8 |
2. SGLang Installation
SGLang>=0.5.10 is required for Qwen3.6. You can install from source or use a Docker image:
Command
3. Model Deployment
This section provides deployment configurations optimized for different hardware platforms and use cases.3.1 Basic Configuration
Interactive Command Generator: Use the configuration selector below to automatically generate the appropriate deployment command for your hardware platform and capabilities.3.2 Configuration Tips
- Speculative decoding (MTP) can significantly reduce latency for interactive use cases.
- Mamba Radix Cache: Qwen3.6’s hybrid Gated Delta Networks architecture supports two mamba scheduling strategies via
--mamba-scheduler-strategy:- V1 (
no_buffer): Default. No overlap scheduler, lower memory usage. - V2 (
extra_buffer): Enables overlap scheduling and branching point caching with--mamba-scheduler-strategy extra_buffer --page-size 64. Requires FLA kernel backend (NVIDIA GPUs only). Trades higher mamba state memory for better throughput.
- V1 (
- The
--mem-fraction-staticflag is recommended for optimal memory utilization, adjust it based on your hardware and workload. - Context length defaults to 262,144 tokens. If you encounter OOM errors, consider reducing it, but maintain at least 128K to preserve thinking capabilities.
- CUDA IPC Transport: Add
SGLANG_USE_CUDA_IPC_TRANSPORT=1as an environment variable to use CUDA IPC for transferring multimodal features, significantly improving TTFT (Time To First Token). Note: this consumes additional memory proportional to image size, so you may need to lower--mem-fraction-staticor--max-running-requests. - Multimodal Attention Backend: Use
--mm-attention-backend fa3on H100/H200 for better vision performance, or--mm-attention-backend fa4on B200. - For processing large images or videos, you may need to lower
--mem-fraction-staticto leave room for image feature tensors. - Hardware requirements:
- BF16: ~35B parameters require ~70GB of GPU memory for weights. TP=1 fits on all supported hardware.
- FP8: The FP8 quantized model requires ~35GB for weights. TP=1 fits on all supported hardware.
| Hardware | Memory | BF16 TP | FP8 TP |
|---|---|---|---|
| H100 | 80GB | 1 | 1 |
| H200 | 141GB | 1 | 1 |
| B200 | 183GB | 1 | 1 |
4. Model Invocation
Deploy Qwen3.6-35B-A3B with the following command (H200, all features enabled):Command
4.1 Basic Usage
For basic API usage and request examples, please refer to:4.2 Vision Input
Qwen3.6 supports image and video inputs as a unified vision-language model. Image Input Example:Example
Example
4.3 Advanced Usage
4.3.1 Reasoning Parser
Qwen3.6 supports Thinking mode by default. Enable the reasoning parser during deployment to separate the thinking and content sections. The thinking process is returned viareasoning_content in the streaming response.
To disable thinking and use Instruct mode, pass chat_template_kwargs at request time:
- Thinking mode (default): The model performs step-by-step reasoning before answering. No extra parameters needed.
- Instruct mode (
{"enable_thinking": false}): The model responds directly without a thinking process.
Example
{"enable_thinking": false} via chat_template_kwargs:
Example
4.3.2 Thinking Preservation
Qwen3.6 has been trained to preserve and leverage thinking traces from historical messages. Enable this for agent scenarios where maintaining full reasoning context improves decision consistency:Example
4.3.3 Tool Calling
Qwen3.6 supports tool calling capabilities. Enable the tool call parser during deployment.Example
