Skip to content

Qwen 3.6 27B

Qwen 3.6 27B is a Qwen 3.6 native vision-language model from Alibaba built on a hybrid linear-attention plus sparse mixture-of-experts architecture, with a context window of 256K tokens and improvements in agentic coding, math and code reasoning, spatial intelligence, and object detection.

ReasoningTool UseImplicit CachingFile InputVision (Image)
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'alibaba/qwen3.6-27b',
prompt: 'Why is the sky blue?'
})

About Qwen 3.6 27B

Qwen 3.6 27B, released on April 22, 2026, is a Qwen 3.6 native vision-language model in Alibaba's Qwen 3 family. It is built on a hybrid architecture that combines linear attention mechanisms with a sparse mixture-of-experts (MoE) framework, a design intended to keep inference efficient at long context while preserving the capability profile of a larger network.

Compared with the prior 3.5-35B-A3B generation, Qwen 3.6 27B brings improvements across several axes. Agentic coding ability is stronger, which matters for pipelines that chain tool calls and multi-step plans. Mathematical and code reasoning have been upgraded for benchmark-style problem solving and real-world programming tasks. Spatial intelligence, object localization, and object detection are sharper, which improves the model's grounding when it must reason about positions of elements within an image.

Native vision-language support means images are treated as first-class inputs alongside text rather than processed through a bolt-on encoder. Qwen 3.6 27B accepts file input and is tagged for reasoning, tool use, and implicit caching, so it can ingest documents and images, decide when to invoke registered tools, and reuse cached prefixes when serving repeated long prompts. The context window of 256K tokens accommodates extended multimodal sessions, document plus image inputs, and long agent traces.

You can integrate Qwen 3.6 27B through AI SDK, Chat Completions API, Responses API, Messages API, or other API formats, from TypeScript or Python. Maximum output is 256K tokens tokens per request.