Microsoft Is Testing Claude Against Its Own Copilot. Here’s Why.

Microsoft Is Testing Claude Against Its Own Copilot. Here’s Why.

More

Descriptions:

Nate B. Jones addresses a growing tension inside large organizations: individual contributors who know their approved AI default — often Microsoft Copilot or Google Gemini — is underperforming, but lack a convincing way to surface that finding without sounding like they are expressing a personal preference. The video reframes the argument from “this tool is bad” to a measurable cost claim: if the default tool requires four extra hours per week on a specific job compared to a specialist alternative, that is a statement an organization can act on.

The framework covers four criteria for picking the right job to test — it runs at least weekly, takes at least 30 minutes, has a real audience for the output, and can be graded instantly by someone who has done it by hand. The same job is then run through both the corporate default and a challenger (Claude, ChatGPT, or Codex) with identical inputs and success criteria, tracking time spent, rework required, and output quality. Jones argues that five to fifteen data points collected over a week constitute more relevant evidence than the vendor demo used in the original procurement decision.

The video walks through how the ask changes as it escalates from a direct manager to a director to an executive, and how to neutralize objections around shadow IT and procurement complexity. Anthropic’s Claude and OpenAI’s tools are identified as the two realistic enterprise-scale alternatives given current capital levels and model shipping velocity.


📺 Source: AI News & Strategy Daily | Nate B Jones · Published April 30, 2026
🏷️ Format: Opinion Editorial

1 Item

Channels

1 Item

Companies