Microsoft is enhancing its AI strategy by integrating multiple large language models (LLMs), notably OpenAI’s GPT and Anthropic’s Claude, into its Copilot platform. This dual-model approach aims to improve the functionality of the Researcher agent, designed for complex problem-solving that requires multi-source analysis. An optional ‘critique’ feature has been introduced, enabling GPT to draft content, which Claude then evaluates for accuracy, completeness, and citation integrity. This innovative workflow may initially seem unconventional; however, it mirrors typical practices in software development where different models are utilized for writing and reviewing code.
According to Microsoft, initial benchmarks underscore the merits of this combined approach. Using Perplexity’s deep research DRACO benchmark, Claude Opus 4.6 scores 42.7 independently but achieves 50.4 in the Deep Research mode. When the Critique feature is activated, Copilot’s Researcher scores 57.4, outpacing either model alone. Although benchmarks for OpenAI’s GPT-5.4 are currently unavailable, it is anticipated that its performance would align closely with that of Opus 4.6.
Another notable feature being introduced is the ‘council,’ allowing users to compare the performance of different models side-by-side on common queries. This comparative analysis could provide further insights into model effectiveness, empowering users to make informed decisions about their AI tools.
Microsoft has also recently integrated Anthropic’s Claude Cowork tool into Copilot as part of its Microsoft 365 Frontier program. Copilot Cowork is designed for knowledge workers requiring long-running agents capable of executing multi-step workflows. This tool is available in early access, presenting an opportunity for businesses to leverage AI without compromising data security. Since Copilot Cowork operates within a sandboxed cloud environment, users can maintain control over their enterprise data while utilizing advanced AI capabilities.
Barton Warner, Senior Vice President of Enterprise Technology at Capital Group, emphasized the transformative potential of the Cowork feature. “This isn’t about generating content or answers. It’s about taking real action – connecting steps, coordinating tasks, and following through across everyday workflows,” he stated. The ability to operate within existing security and risk frameworks allows organizations to experiment and scale confidently, ultimately driving efficiency in their operations.
Microsoft’s decision to incorporate Anthropic’s technologies reflects a strategic shift as it moves away from an exclusive dependence on OpenAI. While diversifying its AI resources, Microsoft is also solidifying its relationship with Anthropic, signaling the company’s commitment to serving enterprise needs through a multi-model approach. For customers invested in premium Copilot services, a critical consideration is whether the value derived from Microsoft’s offerings arises from the orchestrated models or from the robust enterprise data and trust they provide.
In announcing Cowork, Charles Lamanna, Microsoft’s President of Business Applications and Agents, remarked on the significance of this multi-model strategy, stating, “It is this multi-model advantage that makes Copilot different.” If Microsoft were to develop its proprietary models, the approach to AI implementation would likely differ. As it stands, leveraging a combination of external models underlines the company’s adaptive strategy in a competitive landscape.
See also
LG’s 97-Inch OLED TV on Sale for $19,996.99 — Save $5,000 in Amazon Spring Sale
Germany”s National Team Prepares for World Cup Qualifiers with Disco Atmosphere
95% of AI Projects Fail in Companies According to MIT
AI in Food & Beverages Market to Surge from $11.08B to $263.80B by 2032
Satya Nadella Supports OpenAI’s $100B Revenue Goal, Highlights AI Funding Needs





















































