AI News, Week of April 27: GPT-5.5 Ships, DeepSeek V4 Goes Open-Weight, Anthropic Pulls $65B

GPT-5.5 lands with doubled long-context scores, DeepSeek releases two MIT-licensed open-weight models at a fraction of frontier price, and Anthropic collects $65 billion in investment pledges from Amazon and Google — all in one week.

Try the New Models on Writingmate
200+ models
One subscription
No API keys
Cancel anytime
Abstract collage of AI model logos and benchmark charts representing the week of April 27 2026 in AI news
Artem Vysotsky

Author, Co-Founder & CEO

Artem Vysotsky

Sergey Vysotsky

Reviewer, Co-Founder & CMO

Sergey Vysotsky

8 min read
Updated: 04/27/2026

Four days. Four frontier model launches. One company — Anthropic — absorbed $65 billion in new investment pledges. The week of April 20–27, 2026 was the most compressed burst of AI news since the DeepSeek R1 shock of early 2025. Here is everything that moved the needle, with benchmarks and prices — no vague hype.

Model Launches

The week opened on Monday, April 20 with Moonshot AI shipping Kimi K2.6 — a 1-trillion-parameter open-weight model with 32 billion active parameters, a 256,000-token context window, and an agent-swarm capability that now scales to 300 sub-agents coordinating across 4,000 steps (up from 100 sub-agents and 1,500 steps in K2.5). On Humanity's Last Exam (HLE-Full) with tools, K2.6 scored 54.0 — topping GPT-5.4 (52.1), Claude Opus 4.6 (53.0), and Gemini 3.1 Pro (51.4) at the time of release. Weights ship on Hugging Face under a Modified MIT license. In a live demonstration, K2.6 ran continuously for twelve hours to port and optimize an inference engine in Zig, making over 4,000 tool calls and raising throughput from 15 to 193 tokens per second.

Three days later, Thursday, April 23, OpenAI shipped GPT-5.5 — codename Spud — rolling out to Plus, Pro, Business, and Enterprise users in ChatGPT and Codex. The API opened on April 24 at $5 / $30 per million input/output tokens with a 1M-token context window. The headline benchmark numbers: Terminal-Bench 2.0 at 82.7%, and MRCR v2 at 1M tokens jumping from 36.6% (GPT-5.4) to 74.0% — more than doubling. OpenAI positions GPT-5.5 as a step toward a unified ChatGPT–Codex–browser super app for enterprise, excelling at multi-step computer use, autonomous research, and cross-tool agentic execution. Six weeks elapsed between GPT-5.4 and GPT-5.5 — a cadence that suggests OpenAI is operating on a software-release cycle rather than a model-research cycle.

The next morning, Friday, April 24, DeepSeek previewed V4-Pro and V4-Flash — the first major architecture refresh since V3 and the most consequential open-weight release of the year so far. V4-Pro carries 1.6 trillion parameters (49B active); V4-Flash runs at 284B parameters (13B active). Both models support a 1M-token context window and ship under an MIT license permitting commercial fine-tuning. The new hybrid attention mechanism — alternating Compressed Sparse Attention (CSA) and Heavily Compressed Attention (HCA) across 61 transformer layers — cuts V4-Pro's inference FLOPs to 27% of V3.2's at 1M context and KV-cache to 10%. V4-Pro's Codeforces rating of 3,206 is the highest competitive-programming score any model has posted; its SWE-bench Verified score is 80.6%. Pricing: V4-Flash at $0.14 / $0.28 per million in/out tokens, V4-Pro at $1.74 / $3.48 per million in/out tokens — roughly one-seventh the cost of GPT-5.5 at the Pro tier, opening high-volume agentic use cases that were cost-prohibitive at frontier prices.

Rounding out the week, xAI made Grok 4.3 Beta available to SuperGrok Heavy subscribers ($300/month). At 1 trillion parameters — twice the size of Grok 4.20 and trained significantly longer — 4.3 adds native video input, in-chat generation of PDFs, PowerPoint slides, and spreadsheets, and expanded beta access to Grok Computer, xAI's autonomous desktop agent. The 16-agent Heavy system and 2M-token context window from 4.20 both carry over. Full rollout to standard SuperGrok subscribers is expected mid-to-late May.

Bar chart comparing GPT-5.5, DeepSeek V4-Pro, Kimi K2.6, and Grok 4.3 across SWE-bench Verified, Terminal-Bench 2.0, and HLE-Full benchmark scores for the week of April 27 2026

Product Updates

The biggest product stage of the week was Google Cloud Next '26 in Las Vegas on April 22. Google launched the Gemini Enterprise Agent Platform — a full-stack environment with Agent Designer, Agent-to-Agent Orchestration, Agent Registry, Agent Identity, Agent Gateway, and Agent Observability. CEO Sundar Pichai framed the event as the company's formal entry into the agentic era. Google Cloud simultaneously announced a $750 million partner fund for its 120,000-member ecosystem. On the hardware side, 8th-generation TPUs arrived in two purpose-built chip variants engineered for agent workloads, alongside a new cross-cloud Lakehouse and Knowledge Catalog as part of the Agentic Data Cloud announcement.

OpenAI introduced Workspace Agents in ChatGPT: cloud-hosted AI agents that operate across ChatGPT and Slack with org-level approval flows, memory, and analytics — currently in research preview for Business, Enterprise, Edu, and Teachers plans. The company also launched ChatGPT for Clinicians, a free tool for verified US clinicians covering evidence review, real-time documentation, and medical research at the point of care. Codex for Students became available for verified university students in the US and Canada, with $100 in ChatGPT credits per eligible student to extend Codex usage.

Anthropic released two new products this week. Claude Design — an Anthropic Labs product — lets users collaborate with Claude to create slides, one-pagers, prototypes, and visual layouts inside a canvas interface, powered by Claude Opus 4.7 and available in research preview for Pro, Max, Team, and Enterprise subscribers. Alongside it, Managed Agents moved to public beta on the Claude Platform, offering stable session, harness, and sandbox interfaces for long-running agent tasks with durable state and faster startup times.

Industry & Funding

The defining financial story of the week — and possibly of the quarter — unfolded across two days. On Monday, April 20, Amazon announced an investment of up to $25 billion in Anthropic: $5 billion now at the startup's $380 billion valuation, with $20 billion more tied to commercial milestones. In return, Anthropic committed to spending over $100 billion on AWS over the next decade and bringing nearly 1 gigawatt of Trainium2 and Trainium3 capacity online by year-end. Amazon's move came just two months after it had separately agreed to invest up to $50 billion in OpenAI — the cloud giant is evidently hedging across both sides of the frontier model race.

Then on Thursday, April 24, Google committed up to $40 billion in Anthropic — $10 billion in immediate cash at a $350 billion valuation, with $30 billion more conditional on performance targets. Combined with Amazon's pledge made earlier in the week, Anthropic collected $65 billion in new investment commitments within five days. Context: Anthropic's annual run-rate revenue has already crossed $30 billion, up from roughly $9 billion at the end of 2025, driven heavily by explosive Claude Code adoption in enterprise software teams. The dual mega-investment reflects both genuine confidence in the company's growth and each tech giant's desire to lock in preferred cloud-compute commitments before Anthropic's next valuation step.

Separately, OpenAI confirmed it closed a record $122 billion funding round and is now generating $2 billion per month in revenue, with leadership describing an IPO as a near-term strategic priority.

Research & Papers

Sakana AI published The AI Scientist-v2 (arXiv: 2504.08066) — an end-to-end agentic system that autonomously proposes hypotheses, runs experiments, analyzes results, and writes full manuscripts. Why it matters: one AI-generated paper cleared peer review at an ICLR workshop, marking the first independently verified case of a fully autonomous AI producing an accepted scientific publication. Compared to v1, the new system eliminates human-authored code templates, generalizes across ML domains, and introduces a progressive tree-search methodology managed by a dedicated experiment-manager agent. The code is open-sourced at SakanaAI/AI-Scientist-v2 on GitHub.

The DeepSeek V4 technical report drew substantial academic attention for the hybrid CSA/HCA attention architecture described above and for the decision to release weights that run natively on Huawei Ascend chips — the first major frontier open-weight model to explicitly support non-NVIDIA hardware at launch. Multiple researchers called it one of the most rigorously written model papers of the year, with the KV-cache reduction numbers sparking particular discussion about future long-context inference costs.

Community Buzz

r/LocalLLaMA went into sustained eruption after DeepSeek dropped V4-Flash weights. The combination of an MIT license, 1M-token context, and sub-$0.15-per-million-token API pricing generated multiple Hacker News front-page threads and hundreds of Reddit comments within 24 hours. The community's instinctive question with every DeepSeek release — can I run this locally? — quickly gave way to appreciation for the API economics. Multiple threads noted that V4-Flash is fast enough and cheap enough to replace Claude Haiku or Gemma 4 in high-throughput tool-calling pipelines at a fraction of the cost, with one thread titled "Buried lede: DeepSeek V4 Flash is incredibly inexpensive" capturing the community's read precisely.

"Grok 4.3 (beta) has been released by @xai 🔥🔥 The new version of Grok is currently available to the highest tier plan SuperGrok Heavy. With 1T params Grok 4.3 is twice as big as Grok 4.20, and was also trained much longer. It should provide a decent jump in performance."

Mark Kretschmann (@mark_k) on X

Sam Altman sparked a broader conversation on X this week with a post about the unusual emotional attachment users form to specific model versions: "If you have been following the GPT-5 rollout, one thing you might be noticing is how much of an attachment some people have to specific AI models. It feels different and stronger than the kinds of attachment people have had to previous kinds of technology." The thread prompted widespread debate about whether rapid incremental model updates create friction when users have built workflows around specific model behaviors — a tension that will only grow as OpenAI accelerates its release cadence.

The Writingmate Angle

This was one of the most consequential weeks for the model catalog in months. GPT-5.5 is rolling into the Writingmate model list now that the API is live — subscribers can reach it for complex agentic coding, autonomous research, and tasks that benefit from its doubled long-context performance. DeepSeek V4-Flash and V4-Pro are expected to follow within days as API stability settles; they make a strong case for high-volume use cases where cost-per-task is the deciding factor. Kimi K2.6 should arrive shortly after, with Cloudflare Workers AI already offering access. Check /models for the current lineup — if a model from this week isn't there yet, it's usually a matter of days. One subscription, no separate provider accounts required.

Frequently Asked Questions

Artem Vysotsky

Written by

Artem Vysotsky

Ex-Staff Engineer at Meta. Building the technical foundation to make AI accessible to everyone.

Sergey Vysotsky

Reviewed by

Sergey Vysotsky

Ex-Chief Editor / PM at Mosaic. Passionate about making AI accessible and affordable for everyone.

Ready to experience the power of AI?

Access 200+ AI models, custom agents, and powerful tools - all in one subscription.