# Qwen3.5-Flash The Qwen3.5 native vision-language Flash models are built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. Compared to the 3 series, these models deliver a leap forward in performance for both pure text and multimodal tasks, offering fast response times while balancing inference speed and overall performance. ## Model Information - **Organization**: [qwen](/llm.txt) - **Slug**: qwen3-5-flash - **Available at Providers**: 13 - **Release Date**: February 25, 2026 ### Benchmark Scores - Weekly: 0.18 ## Providers | Provider | Name | $ Input (per 1M) | $ Output (per 1M) | Free | Link | |----------|------|-----------------|------------------|------|------| | [AIHubMix](/llm/aihubmix.txt) | qwen3.5-flash | 0.03 | 0.28 | | [View](https://aihubmix.com/model/qwen3.5-flash) | | [ApiYI](/llm/apiyi.txt) | qwen3.5-flash | | | | | | [ApiYI](/llm/apiyi.txt) | qwen3.5-flash-2026-02-23 | | | | | | [Kilo Code](/llm/kilocode.txt) | Qwen: Qwen3.5-Flash | 0.10 | 0.40 | | [View](https://kilo.ai/models/qwen/qwen3.5-flash-02-23) | | [Nano-GPT](/llm/nanogpt.txt) | Qwen3.5 Flash | | | | | | [Nano-GPT](/llm/nanogpt.txt) | Qwen3.5 Flash Thinking | | | | | | [OpenRouter](/llm/openrouter.txt) | Qwen3.5-Flash | 0.10 | 0.40 | | [View](https://openrouter.ai/qwen/qwen3.5-flash-20260224) | | [Poe](/llm/poe.txt) | Qwen3.5-Flash | | | | [View](https://poe.com/qwen3.5-flash/api) | | [Routeway](/llm/routeway.txt) | Qwen: Qwen 3.5 Flash | 0.09 | 0.37 | | [View](https://routeway.ai/models) | | [302.AI](/llm/302ai.txt) | qwen3.5-flash | 0.03 | 0.29 | | [View](https://302ai-en.apifox.cn/207705113e0) | | [ZenMUX](/llm/zenmux.txt) | Qwen: Qwen3.5-flash | 0.10 | 0.40 | | | | [Writingmate](/llm/writingmate.txt) | Qwen: Qwen3.5-Flash | | | | [View](https://writingmate.ai/models/qwen/qwen3.5-flash-02-23) | | [Arena AI](/llm/arenaai.txt) | | | | | | --- [← Back to all providers](/llm.txt)