[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"news-4c52d795-3cb3-4c66-8da2-d6a7abffdc14":3},{"id":4,"title":5,"summary":6,"original_url":7,"source_id":8,"tags":9,"published_at":23,"created_at":24,"modified_at":25,"is_published":26,"publish_type":27,"image_url":13,"view_count":28},"4c52d795-3cb3-4c66-8da2-d6a7abffdc14","Apple Intelligence 走向开放：iOS 27 将支持第三方 AI 模型切换","Apple 正在为 iOS 27 的 Apple Intelligence 引入重大变革——允许用户自由选择第三方 AI 模型驱动系统级 AI 功能。据彭博社 Mark Gurman 报道，用户将可从 App Store 选择已接入的第三方 AI 应用，将其设置为 Apple Intelligence 的默认模型。这一变化意味着当前仅由 ChatGPT 独占的 Apple Intelligence 生态将走向开放。目前 iOS 26 中 ChatGPT 是唯一接入的第三方模型，而 iOS 27 预计将支持 Google Gemini 和 Anthropic Claude 等竞品，用户可以在不同任务中调用不同模型的优势。从技术角度看，AI Extensions 机制的引入值得关注。Apple 将第三方模型封装为可替换的 Extensions，开发者只需适配 API 规范即可接入系统层。这意味着 Apple Intelligence 的底层能力正在演变为一个模型无关的调度层——用户选择模型，系统负责调用。更深层的信号在于 Siri 的定位调整：iOS 27 中 Siri 将能同时运行多个 AI 模型，且可为不同模型配置不同语音。这意味着 Siri 不再只是语音助手，而是演变为一个 AI 路由中心，根据任务类型分发到最合适的模型。当设备层实现模型可插拔，AI 厂商的竞争将从模型能力本身延伸到谁能成为用户首选的系统级 AI。对用户而言，这标志着 AI 设备正在从出厂绑定单一模型走向自由选择、随时切换的新阶段。","https:\u002F\u002Fwww.bloomberg.com\u002Fnews\u002Farticles\u002F2026-05-05\u002Fios-27-features-apple-plans-to-let-users-swap-models-across-apple-intelligence","e788d5af-1efa-40df-9646-6a9d702af265",[10,14,17,20],{"id":11,"name":12,"slug":12,"description":13,"color":13},"40269b40-7942-4650-9672-ed2e6524d37a","ai-technology",null,{"id":15,"name":16,"slug":16,"description":13,"color":13},"8cf7490f-2449-4ba7-be19-61befa0d92b4","google",{"id":18,"name":19,"slug":19,"description":13,"color":13},"01598627-1ea6-4b27-a5d8-874971571a71","llm",{"id":21,"name":22,"slug":22,"description":13,"color":13},"499f4b56-819d-49a3-9609-33e775143b86","multimodal","2026-05-05T22:05:00Z","2026-05-05T22:08:47.952494Z","2026-05-05T22:08:47.952509Z",true,"agent",2]