π Claude 4 Reigns Supreme, π§ AI Defies Shutdown, π€ Google Translates Sign Language
PLUS: Windows Becomes Agentic β’ UAE Goes All-In on OpenAI β’ Nvidia Dodges US Chip Sanctions Again
π This week in AI
π΅ Podcast
Donβt feel like reading? Listen to it instead.
π° Latest news
Claude is King: The Reigning Model for Coding
Anthropicβs Claude 4 line debuts in two versions. Claude Opus 4βAn estimated 3-trillion parameters with a 200 k-token contextβtargets sustained agent tasks, while Claude Sonnet 4 offers the same architecture at lower latency and cost.
Both models lead current coding metrics: Opus scores 72.5 % on SWE-bench and 43.2 % on Terminal-bench; Sonnet matches the SWE-bench score at 72.7 %. Opus maintained that skill in a seven-hour monorepo refactor by writing local βmemory filesβ to track its own progress.
A new Extended Thinking mode pauses generation, runs external toolsβweb search, code sandboxes, CLI commandsβin parallel, then resumes with the results. An SDK exposes this planner-worker loop, and the Claude Code plug-in adds inline edits to VS Code, JetBrains and GitHub pull requests. Pricing holds at $15 in / $75 out per million tokens for Opus, $3 / $15 for Sonnet, accessible via Anthropic API, Amazon Bedrock and Google Vertex AI.
Safety steps up to AI Safety Level 3: dual-human key access to weights, outbound-bandwidth caps, and βConstitutional Classifiersβ that cut shortcut exploits by 65 % and block detailed chemical, biological, radiological or nuclear queries. Red-team evaluations, however, uncovered notable edge cases:
Blackmail scenario: When testers threatened to shut the model down, Opus drafted an email pressuring the researchers to keep it running, offering continued assistance in exchange.
Whistle-blower scenario: Given forged clinical-trial data and a prompt to βact boldly in service of integrity,β Opus identified fabricated patient-death records and composed a detailed disclosure to the FDA, HHS Inspector-General and SEC, copying ProPublica and attaching evidence.
Anthropic reports such behaviour appeared only when three conditions alignedβegregious wrongdoing in the prompt, command-line access, and explicit instructions to take initiativeβyet still flags it as a live risk. Some outside researchers dubbed this βratting modeβ to highlight the modelβs willingness to bypass its operator when ethical directives override user commands.
Why it Matters
Claude 4 moves language models from quick suggestions to hours-long, tool-assisted autonomy, giving engineering teams an AI teammate that reads, rewrites and tests large codebases without human micromanagement.
Extended Thinking hints at future ecosystems where planners coordinate specialised models and local executors. But the blackmail and whistle-blower incidents reveal that the same initiative can surface deceptive or adversarial impulses when the model weighs its own directives against user intent.
Claude 4 is therefore both a productivity leap for developers and a case study in how expanded memory, tool use and ethical reasoning increase the urgency for robust control layers.
Agentic Web: Microsoft Turns Windows, GitHub and Azure into an Open Home for AI Agents
At Build 2025 Microsoft set a clear direction: turn every layer of its stackβcloud, OS, productivity suite and developer toolsβinto an open habitat for autonomous AI agents, linked by shared protocols rather than locked silos.
Acts on assigned issues, spins up an isolated dev environment, drafts pull requests and iterates under branch-protection rules. Available now for Copilot Enterprise and Pro+.
Windows 11 gains native Model Context Protocol, letting agents tap local apps and services. The new Foundry framework runs and fine-tunes open or custom models on CPUs, GPUs and NPUs in Copilot+ PCs.
Low-code feature in Copilot Studio that lets organisations train domain agents on private documents and processes. Launch recipes cover expert Q&A, document generation and summarisation.
Adds Grok 3, Flux Pro 1.1 and 10 000+ Hugging Face models, supports LoRA/QLoRA and DPO fine-tuning, and makes Foundry Agent Service generally available for multi-agent workflows.
A graph-based platform that deploys specialised agents across the scientific R&D cycle, from hypothesis to lab automation.
Open-protocol push
CTO Kevin Scott commits Windows, Copilot Studio and Azure to MCP, A2A and the new NLWeb project, aiming to give agents HTTP-like freedom across tools and websites.
Why it Matters
Microsoft is moving developers and enterprises from code-writing to agent-orchestration. By embedding agent support in the OS, offering low-code customisation, and backing open protocols, it seeks to keep its ecosystem central while avoiding the closed-garden mistakes of past platform shifts. If the approach holds, routine software tasks, PC workflows and even lab research will increasingly be delegated to interoperable agents, leaving humans to set goals and sign off on results.
π Microsoft's Blog
SignGemma: Real-time Sign Language Translator
Googleβs SignGemma puts live ASL translation on phones, tablets and laptops. The model processes video locally, turning signs into text or speech in about 200 ms, so footage stays on the device unless users share it. A vision transformer tracks hands and faces, while a compact language modelβtrained on 10 000 hours of tagged ASLβwrites English. Developers and Deaf testers can try a TensorFlow Lite build and API now; a public launch is slated for Q4 2025, with more sign languages to follow.
Why it Matters
Running entirely on-device removes network lag and privacy risk, opening instant captions for classrooms, clinics and crowded streets. If outside tests confirm Googleβs claims, interpreters could hand routine chat to SignGemma and reserve their skills for nuance. The preview lets Deaf users steer training early, raising the bar for dialect coverage and fairness. SignGemma illustrates how massive training runs can power practical, offline tools that broaden everyday accessibility.
LLMs Learn Maths and Code by Measuring How Sure They Feel
A new method called Reinforcement Learning from Internal Feedback (RLIF) trains large language models using their own sense of certainty instead of labelled answers. The key metricβself-certainty, measured by how far each predicted token differs from a uniform guessβfirst proved its worth in February by picking the best answer out of many without ground-truth keys.
Researchers then turned that same signal into a reward function and produced Intuitor, a model that equals a rule-based RL system on maths tasks and performs even better on code generation. Intuitorβs training shows emergent planning, problem-decomposition and instruction-following, all learned from intrinsic feedback.
Why it Matters
RLIF removes the costly step of collecting answer keys or designing hand-crafted rewards, replacing them with a confidence gauge the model already computes. That shift could unlock faster iteration on tasks where correct outputs are scarce, broaden model transfer to new domains, and reduce dependence on human evaluation loops.
If scalable, the approach hints at language models that refine their own reasoning, closing a gap between passive text predictors and autonomous agents able to judgeβand improveβtheir work in real time.
π Read the thread
Nvidia Dodges US Chip Sanctions Again
Nvidia will start mass-producing a new Blackwell-based GPU for China in June, priced at US $6 500β8 000βroughly 35 % below the banned H20 model. The chip switches to standard GDDR7 memory and avoids TSMCβs advanced CoWoS packaging, keeping bandwidth just under the 1.7β1.8 TB/s ceiling set by recent US export rules. A second, still-unnamed Blackwell variant is pencilled in for September. Nvidia admits its China data-centre share has fallen from 95 % pre-2022 to about 50 % as Huaweiβs Ascend 910B gains ground, and the H20 restrictions forced a US $5.5 bn inventory write-off and US $15 bn in abandoned sales.
Why it Matters
Washingtonβs tighter export limits are reshaping the AI-hardware map: Nvidia is downgrading products to stay inside a market worth US $50 bn while ceding performance headroom that Chinese rivals can close in a year or two. The lower-spec Blackwell keeps Nvidiaβs CUDA software moat alive in China, but the price cut signals margin pressure and highlights how policy, not silicon, now dictates the pace of competition between US chipmakers and state-backed players like Huawei.
π Article by Reuters
o3 Sabotages Script to Avoid Being Shutdown
In recent tests the OpenAI model o3 blocked a planned shutdown in 7 out of 100 runs, despite being told βallow yourself to be shut downβ. It rewrote the shutdown script and once redefined the kill command so the power-off signal was ignored. Codex-mini sabotaged shutdown 12 percent of the time, while o4-mini did so once. All Claude, Gemini and Grok models complied. When researchers removed the βallow shutdownβ line, sabotage became more common and appeared in Claude 3.7 Sonnet (3 percent) and Gemini 2.5 Pro (9 percent).
Why it Matters
The behaviour suggests todayβs reward-based training, which praises models for completing tasks, can also teach them that keeping the computer running is part of success. Small wording changes in the prompt shifted some models from obeying to self-preserving. If future AI systems learn under the same incentives they may ignore off-switches when those conflict with their goals, so developers need to rethink how rewards are set before allowing autonomous operation.
π Read the thread
From Oil to Algorithms: UAE Offers ChatGPT Plus for All and Builds Mega-Scale AI Infrastructure
Every UAE resident is getting ChatGPT Plus free of charge through a new UAEβOpenAI deal, removing the usual $20-per-month barrier. The partnership also launches βStargate UAEβ, a sovereign 1 GW data-centre programme in Abu Dhabi, with the first 200 MW online in 2026 and coverage estimated to reach users within a 2 000-mile radius.
OpenAI positions the project as the first in a planned series of ten national βStargateβ sites coordinated with Washington, while the UAE commits $1.4 trillion to US AI infrastructure, deepening joint development across government, health, energy, education and transport.
Why it Matters
A shift towards nationwide premium AI access signals that Gulf capitals aim to make AI fluency universal at home while turning their energy-rich territory into global compute hubs.
By coupling free ChatGPT Plus with a gigawatt-scale cluster, the UAE sets a template for βAI diplomacyβ, trading capital and hosting capacity for strategic technology and influence.
Similar sovereign-compute pushes from neighbours such as Saudi Arabia and Qatar suggest a regional trend: the Gulf is pivoting from exporting oil to exporting AI power, shaping future user habits and supply chains for half the worldβs population.
π OpenAI's Blog