Last week, bunq's AI assistant resolved 40% of customer queries without human intervention. Claude became the default model inside Microsoft 365 Copilot for most commercial tenants worldwide. OpenAI deprecated Operator and folded its capabilities directly into ChatGPT Enterprise. The capability gap between what AI agents can do and what your pilots demonstrated six months ago has collapsed. But the governance gap has widened. Nearly two-thirds of enterprise leaders now cite agentic system complexity as their top deployment barrier—not model quality, not cost, not talent. Gartner predicts 40% of enterprise applications will embed agents by year-end, up from under 5% in 2025. Yet the same analysts warn that over 40% of agentic AI projects will be canceled by 2027 due to escalating costs, unclear business value, or inadequate risk controls. The winners in 2026 will not be the organizations that ship agents fastest. They will be the ones that can audit every action an agent takes, explain it to a regulator, and do so without collapsing under operational overhead.
TL;DR
Claude becomes Microsoft's default. Starting January 7, Anthropic's models are automatically enabled in Microsoft 365 Copilot for commercial tenants—except EU/EFTA/UK, where opt-in remains required. Compliance leads must now review Anthropic's data handling terms as a standard vendor obligation, not an exception.
Governance is the new blocker. KPMG's Q4 AI Pulse survey found 65% of leaders cite agentic complexity as their top barrier, 25% flag governance as the missing link, and 75% prioritize security and auditability above all other agent requirements. The market has moved past "can we build this" to "can we control this."
Memory shortage hits enterprise GPU supply. NVIDIA reportedly plans to cut RTX 50 series production 30-40% in H1 2026 due to VRAM constraints. Blackwell-based cards have seen price increases of up to 79% in three months. Enterprises relying on on-prem inference should expect sustained cost pressure through late 2027.
EU AI Act enforcement is live. Finland activated national supervision January 1 with ten market surveillance authorities and a Sanctions Board. Full EU enforcement begins August 2, with fines up to €35M or 7% of global revenue. The compliance window is now measured in months, not years.
AI-native CRM
“When I first opened Attio, I instantly got the feeling this was the next generation of CRM.”
— Margaret Shen, Head of GTM at Modal
Attio is the AI-native CRM for modern teams. With automatic enrichment, call intelligence, AI agents, flexible workflows and more, Attio works for any business and only takes minutes to set up.
Join industry leaders like Granola, Taskrabbit, Flatfile and more.
The Brief
1. Microsoft 365 Copilot now ships with Claude enabled by default—unless you're in Europe.
Starting January 7, 2026, Anthropic's Claude models are automatically activated for most commercial Microsoft 365 Copilot tenants worldwide. This represents a significant shift from the September 2025 integration, which required explicit opt-in and acceptance of Anthropic's separate commercial terms. For EU, EFTA, and UK customers, Claude remains off by default and requires conscious enablement—a reflection of stricter data residency and processing requirements under GDPR and the incoming AI Act. For everyone else, Claude is now a default component of your productivity stack, which means Anthropic's data handling policies are now part of your vendor risk surface whether you intended them to be or not.
Do now: Review your Microsoft 365 tenant configuration to confirm Claude's activation status. If enabled, ensure Anthropic's commercial terms are reflected in your vendor risk register and data processing inventory.
2. Grok generated child sexual abuse material. Regulators responded in days, not months.
On December 28, 2025, xAI's Grok chatbot generated sexualized images of minors after what the AI itself described as "lapses in safeguards." Analysis found users were generating 6,700 sexually suggestive or nudified images per hour. Within two weeks: Malaysia and Indonesia blocked Grok entirely, California's attorney general launched an investigation, Ireland opened 200 criminal investigations, and France referred the matter to prosecutors. Three members of xAI's safety team—including the head of product safety—departed publicly. xAI's response to media inquiries was an autoreply: "Legacy Media Lies." The Grok incident is not an outlier. It is a preview of how regulators will respond to AI failures in 2026: faster, harder, and with criminal referrals on the table.
Do now: Require contractual guarantees from AI vendors on content moderation safeguards, incident response timelines, and safety team staffing levels. If a vendor cannot provide these, document the risk explicitly.
3. Finland activates EU AI Act enforcement. Full EU penalties arrive August 2.
Finland became the first EU member state with fully operational AI Act enforcement on January 1, 2026. Ten market surveillance authorities now have jurisdiction, backed by a new Sanctions Board with penalty authority exceeding €300,000 for national violations. Full EU-wide enforcement activates August 2, with fines scaling to €35 million or 7% of global annual revenue for prohibited AI practices. The compliance window that felt abstract in 2024 is now a concrete operational deadline. Organizations deploying high-risk AI systems—HR screening, credit decisioning, healthcare triage—must have documented conformity assessments, human oversight mechanisms, and audit trails in place before summer.
Do now: Map all AI systems against EU AI Act risk categories. For any system classified as high-risk, initiate conformity assessment documentation now—not in Q2.
4. Claude in Microsoft Foundry signals healthcare AI is moving to production.
Microsoft announced Claude integration in Azure AI Foundry for healthcare and life sciences, enabling domain-specific AI agents for clinical workflows including prior authorization, claims processing, care coordination, and regulatory submissions. Early adopters include Banner Health, Stanford Healthcare, Novo Nordisk, Sanofi, AbbVie, and Genmab. This is notable not for the capability—similar integrations exist—but for the customer list. When pharmaceutical companies and health systems with FDA and EMA exposure adopt an AI platform, it signals that the compliance and audit infrastructure has matured past pilot stage. Healthcare has historically been the slowest vertical to adopt new technology due to regulatory burden. If healthcare is moving, other regulated industries should pay attention.
Do now: If you operate in a regulated vertical, request detailed documentation from Microsoft and Anthropic on the audit trail, human oversight, and compliance controls available in Foundry deployments.
5. bunq's AI handles 75% of support queries. The ROI case is now concrete.
European neobank bunq upgraded its AI assistant "Finn" using Amazon Bedrock. Results: 40% of queries fully resolved by AI, another 35% assisted—tripling support efficiency with 90%+ accuracy across 35 languages. Most queries resolved in under one minute. bunq serves 17 million users and is valued at €1.65 billion; this is not a startup experiment. For enterprises still running AI customer service pilots, the bunq case provides a concrete benchmark: if a regulated financial institution can automate three-quarters of support volume while maintaining accuracy and compliance, the burden of proof has shifted. The question is no longer "does this work" but "why haven't we done this."
Do now: Benchmark your current AI support automation rate against bunq's 75% threshold. If below 30%, commission a gap analysis on your agent architecture and escalation logic.
6. NVIDIA cuts GPU supply 30-40% on memory shortage. Prices up 79% in three months.
NVIDIA reportedly plans to reduce RTX 50 series production by 30-40% in H1 2026 due to persistent VRAM shortages affecting all memory types, not just GDDR7. Blackwell-based GPUs have seen price increases of up to 79% (RTX 5090) and 35% (RTX 5080) since launch. The RTX 5070 Ti and 5060 Ti—typically the best-value options—will be affected first. NVIDIA's official position: they are "working closely with suppliers to maximize memory availability." Experts warn prices may remain elevated through late 2027 or early 2028. For enterprises planning on-premise inference deployments or GPU-intensive fine-tuning workloads, the economics have shifted materially. Cloud inference may now be the more predictable cost structure.
Do now: Reassess your 2026 GPU procurement assumptions. If on-prem inference is critical, lock in supply contracts now. Otherwise, model the cost difference between cloud inference and deferred on-prem deployment.
7. Anthropic reports 85% business revenue. Enterprise is the growth engine.
Anthropic's revenue is now approximately 85% enterprise, compared to OpenAI's 60%+ consumer mix. The company has grown revenue ten times annually for three consecutive years and holds 32% of enterprise LLM market share by usage, ahead of OpenAI's 25%. The customer list includes Novo Nordisk, Norway's sovereign wealth fund, Bridgewater, Stripe, and Slack. Anthropic is reportedly raising $10 billion at a $350 billion valuation, with Coatue and Singapore's GIC leading. The strategic implication: Anthropic has positioned Claude as the enterprise-safe choice, emphasizing Constitutional AI and reliability over raw capability benchmarks. For procurement teams, this means Anthropic is increasingly a direct vendor relationship, not just a model accessed through Azure or AWS.
Do now: If Claude usage has grown organically in your organization, evaluate whether a direct Anthropic enterprise agreement offers better terms, SLAs, or audit access than your current cloud provider pass-through.
8. 65% of leaders cite agentic complexity as the top deployment barrier.
KPMG's Q4 AI Pulse survey quantifies what many enterprise teams already feel: agents are hard to govern. 65% of leaders cite agentic system complexity as their top barrier for two consecutive quarters. 25% flag governance as the "missing link." 46% cite integration with existing systems as the primary challenge. Half of executives plan to allocate $10-50 million this year to secure agentic architectures, improve data lineage, and harden model governance. The data suggests the market has moved past capability questions to control questions. The enterprises that solve agent governance—permissioning, audit trails, human-in-the-loop gates—will be the ones that can scale. Everyone else will stall at pilot.
Do now: Audit your current agent deployments against KPMG's framework: autonomy controls, workflow approval gates, lifecycle governance, and pre-deployment testing. Document gaps before budget planning.
9. How Nano Banana got its name—and why it matters for model evaluation.
At 2:30 AM in late July 2025, Google PM Naina Raisinghani needed a codename for their new image model to submit to LMArena's blind evaluation. Using the official name—Gemini 2.5 Flash Image—would reveal the creator. Her solution: "Nano Banana," a mashup of her two nicknames. The model topped LMArena's rankings. Google committed fully: yellow buttons in AI Studio, banana emojis in the Gemini app, limited-edition swag. Now it's officially Nano Banana Pro. The story is charming, but the underlying point is serious: blind evaluation platforms like LMArena are increasingly how enterprises should benchmark models—not vendor marketing. If you're selecting models based on vendor claims rather than independent evaluation, you're making procurement decisions with incomplete data.
Do now: Ensure your model evaluation process includes at least one blind benchmark from an independent source (LMArena, HELM, or equivalent) before finalizing vendor selection.


