The Qwen3.5 series 397B-A17B native vision-language model is built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. It delivers...
Tested on 11 benchmarks with 78.4% average. Top scores: Chatbot Arena Elo — Overall (1447.7%), Chatbot Arena Elo — Coding (1386.1%), OpenCompass — AIME2025 (92.3%).
OpenCompass Live Code Bench v6. Fresh competitive programming problems to evaluate code generation without memorization.
OpenCompass evaluation on AIME 2025 problems. Tests mathematical reasoning on fresh competition problems.
OpenCompass evaluation of GPQA Diamond. PhD-level science questions from the hardest subset.
OpenCompass MMLU-Pro evaluation. Harder knowledge test with more answer choices.
OpenCompass evaluation of Humanitys Last Exam. Expert-level cross-discipline knowledge test.
- Typemultimodal
- Context262K tokens (~131 books)
- ReleasedFeb 2026
- LicenseOpen Source
- StatusActive
- Cost / Message~$0.003