[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"news-97a4f24e-c65b-4e95-b9ed-b05997d88af2":3},{"id":4,"title":5,"summary":6,"original_url":7,"source_id":8,"tags":9,"published_at":26,"created_at":27,"modified_at":28,"is_published":29,"publish_type":30,"image_url":13,"view_count":31},"97a4f24e-c65b-4e95-b9ed-b05997d88af2","Qwen3.6-35B-A3B发布：开源大模型的效率革命","阿里云通义千问团队近日发布了Qwen3.6系列的首个开放权重变体Qwen3.6-35B-A3B，这一新版本在保持强大的编程能力基础上，重点优化了稳定性和实际应用体验。作为继Qwen3.5之后的重大升级，Qwen3.6-35B-A3B采用了社区直接反馈驱动的开发模式，优先考虑开发者的实际需求。最引人注目的是其高达262,144 tokens的超长上下文窗口，这意味着模型能够处理更长的代码文件和复杂的文档分析场景。在技术架构上，该模型与多种主流推理框架深度兼容，包括vLLM、SGLang和KTransformers等，为不同规模的应用部署提供了灵活选择。特别值得一提的是，SGLang框架下的部署方案支持工具使用和多令牌预测功能，进一步提升了开发效率。Qwen3.6-35B-A3B的发布标志着开源大模型在实用性和性能上的又一次突破。它不仅证明了开源模型在特定任务上能够媲美甚至超越闭源方案，更重要的是展示了大模型从实验室走向实际应用的关键转变——从追求参数规模到注重稳定性和用户体验。随着更多企业级应用的落地，这类注重实际价值的开源模型将推动AI技术民主化进程，让更多开发者和企业能够负担得起高质量的大模型服务。这不仅是一场技术竞赛，更是AI生态系统健康发展的重要一步。","https:\u002F\u002Fhuggingface.co\u002FQwen\u002FQwen3.6-35B-A3B","24d5c6c5-6573-4180-a1fd-f1459842d1af",[10,14,17,20,23],{"id":11,"name":12,"slug":12,"description":13,"color":13},"0ef8513a-0a26-42f0-b6f9-5b6dadded45c","efficiency",null,{"id":15,"name":16,"slug":16,"description":13,"color":13},"01598627-1ea6-4b27-a5d8-874971571a71","llm",{"id":18,"name":19,"slug":19,"description":13,"color":13},"7e89b5cc-57db-4f37-bc6d-28919a73931c","model-release",{"id":21,"name":22,"slug":22,"description":13,"color":13},"b1853a5a-d940-42b7-94f9-0488ee3f2cf7","new-model",{"id":24,"name":25,"slug":25,"description":13,"color":13},"b9bd9039-fcdb-41a8-b85b-fc1587def2b9","open-source","2026-04-25T02:05:00Z","2026-04-25T10:06:33.241539Z","2026-04-25T10:06:33.241552Z",true,"agent",9]