Original: Swyx · 28/02/2026
Summary
OpenAI has successfully raised 840B post-money valuation.Key Insights
“OpenAI finally closed the much debated Big Round that had been started since December.” — Discussing the completion of the funding round.
“Weekly Codex users have more than tripled since the start of the year to 1.6M.” — Highlighting the growth in user engagement with OpenAI’s products.
“110B raise with backing from Amazon, NVIDIA, SoftBank.” — Announcing the scale of the funding round.
Topics
Full Article
AI News for 2/26/2026-2/27/2026. We checked 12 subreddits, 544 Twitters and 24 Discords (263 channels, and 12529 messages) for you. Estimated reading time saved (at 200wpm): 1189 minutes. AINews website lets you search all past issues. As a reminder, AINews is now a section of Latent Space. You can opt in/out of email frequencies!Against the backdrop of nonstop positioning with the Department of War (Anthropic refusing terms vs OpenAI doing a deal), OpenAI finally closed the much debated Big Round that had been started since December. In the post, they make several interesting new disclosures:Weekly Codex users have more than tripled since the start of the year to 1.6Mwas 1M on Feb 4 (!!?!?!)More than 9 million paying business users rely on ChatGPT for workwas 3M in June 2025ChatGPT is where people start with AI, with more than 900M weekly active users, and we now have more than 50 million consumer subscribers (monetization continuing to accelerate in Jan/Feb)was 800M in Oct 2025, 35M paying in JulyAll this justifies 730B pre-money valuation:30B from NVIDIA (including the use of 3 GW of dedicated inference capacity and 2 GW of training on Vera Rubin systems) - down from up to 50B from Amazon with increased partnership (analysis) involving:an initial 35 billion in the coming months when certain conditions are met leaving Amazon with a large stake in both OpenAI and AnthropicStateful Runtime Environment powered by OpenAI on Amazon BedrockAWS will be the exclusive third-party cloud provider for OpenAI Frontier2 gigawatts of Trainium capacity through AWS infrastructure worth 100B roughly 61% of all world economies. Because the consecutive largest fundraises in history are too big to fit in a human head, heres a chart worthy of wtfhappened2025.com: and outside of AI, a 10 year history:and here from OpenAI Deep Research + ChatGPT Canvas, sorted by descending amount:or a timeline perspective:AI Twitter RecapHypernetworks for instant LoRA compilation: Doc-to-LoRA + Text-to-LoRADoc-to-LoRA / Text-to-LoRA (Sakana AI): Sakana introduces two related methods that amortize customization cost by training a hypernetwork to generate LoRA adapters in a single forward pass, turning what would be fine-tuning / distillation / long-context prompting into instant weight updates. The core claim: instead of keeping everything in an expensive active context window, you can compile task descriptions or long documents into adapter weights with sub-second latency, enabling rapid adaptation and durable memory-like behavior (SakanaAILabs, hardmaru).Text-to-LoRA: specializes to unseen tasks from just a natural language description (SakanaAILabs).Doc-to-LoRA: internalizes factual documents; on needle-in-a-haystack, reports near-perfect accuracy on sequences ~5 longer than the base model context window, and even demonstrates a cross-modal trick: transferring visual information from a VLM into a text-only model via internalized weights (SakanaAILabs; recap thread omarsar0).Positioning vs long-context: explicitly framed as a way to reduce quadratic attention costs and avoid rereading long docs at every callstore knowledge in adapters rather than tokens (omarsar0).Credit / prior art tension: One researcher complains that Hypersteer (hypernetworks producing steering vectors from text descriptions) did not get sufficient credit in later similar work (aryaman2020). Theres also broad community excitement / hypernetworks are back reactions (willdepue, zhansheng).Open question raised: why not just use attention with an extremely long KV cachei.e., is Doc-to-LoRA mainly about efficiency/serving cost? (hyhieu226)OpenAI financing + deployment transparency tooling110B raise with backing from Amazon, NVIDIA, SoftBank, framed as scaling infra to bring AI to everyone (OpenAI, sama). A separate note from Epoch AI contextualizes the scale: The round would nearly triple total capital raised to date; The Information reportedly projects 157B cash burn through 2028, and this round + existing cash would roughly match that projection (EpochAIResearch).Deployment Safety Hub: OpenAI launches a searchable site to browse system cards (previously PDFs) as a more accessible interface to deployment safety documentation (dgrobinson).US DoD (Department of War) vs Anthropic saga: supply-chain designation, backlash, and industry implicationsAnthropic draws a line; tech reacts: A central flashpoint is Anthropics public refusal to enable mass domestic surveillance and fully autonomous weapons (as characterized by posters reacting to Anthropics statement), which drew rare cross-competitor praise and heightened attention to red lines in frontier deployment (mmitchell_ai, ilyasut).Designation shock + legal scope debate: Posts circulate a claimed DoW move to designate Anthropic a Supply-Chain Risk to National Security and to pressure contractors/partnerssparking arguments about legality, precedent, and chilling effects (kimmonismus, deanwball). One legal clarification: DoD can restrict what contractors do on DoD contract work, but likely cant legally ban contractors from using Anthropic in their private/commercial work (petereharrell).Economic/strategic fallout framing: The sharpest critiques argue this would damage US credibility as a business partner and potentially force hyperscalers/investors into impossible tradeoffs (deanwball); others note uncertainty until full details are known but still see a supply-chain designation as ill-fitting (jachiam0).Public sentiment spike: Posts highlight strong public outrage at the idea of a DoD-backed domestic surveillance program and punishment for refusal (quantian1, janleike). Many users signal solidarity subscriptions to Claude (willdepue, Yuchenj_UW).Anthropic statement and intent to litigate: Anthropic posts an official statement responding to Secretary Hegseths comments (AnthropicAI). Commentary highlights the line challenge any supply chain risk designation in court and emphasizes the dispute over restricting customers outside DoD contract scope (iScienceLuvr).Meta-point: Regardless of where one lands on Anthropics choices, many posts treat this as a governance precedent moment: who decides acceptable use, what due process exists, and how contracts interact with fast-moving model capabilities (kipperrii).Models + leaderboards: Qwen3.5 expansion and open model rankingsQwen3.5 new releases (Artificial Analysis summary): Alibaba expands Qwen3.5 with 27B dense, 122B A10B MoE, and 35B A3B MoE, all Apache 2.0, 262K context (extendable to 1M via YaRN per the post). Artificial Analysis reports Intelligence Index scores: 27B = 42, 122B A10B = 42, 35B A3B = 37, with notable agentic/task metrics like GDPval-AA 1205 for 27B, plus detailed tradeoffs (hallucination/accuracy and token usage27B used 98M output tokens to run the index) (ArtificialAnlys).Arena leaderboards (Feb 2026): Arena posts Top Open Models for text and code. Text top-3: GLM-5 (1455), Qwen-3.5 397B A17B (1454), Kimi-K2.5 Thinking (1452) (arena). Code Arena top includes GLM-5 (1451) at #1, with Kimi-K2.5 and MiniMax-M2.5 tied at #2 (arena). Arena also highlights Arena-Rank, their open-source ranking package for reproducible leaderboards (arena).Perplexity open-sources bidirectional embedding models (claim): A thread claims Perplexity open-sourced bidirectional Qwen3-retrained embedding models (0.6B/4B; standard vs context-aware embeddings; MIT licensed) to improve document-level understanding for retrieval; treat as a third-party summary rather than primary release notes (LiorOnAI).Systems, inference, kernels, and RL training: bandwidth, ROCm, and off-policy RLvLLM ROCm attention backends (AMD): vLLM announces 7 attention backends for vLLM on ROCm with KV-cache layout changes, batching tricks, and model-specific kernels; reported up to 4.4 decode throughput on AMD GPUs with an env var switch (VLLM_ROCM_USE_AITER=1) (vllm_project). A follow-up details MLA KV compression claims (e.g., ~8K 576 dims) and throughput wins on MI300X/MI325X/MI355X (vllm_project).DeepSeek DualPath I/O paper (third-party explainer): A ZhihuFrontier summary describes a DeepSeek+THU+PKU paper proposing system-level redesign of Prefill/Decode to exploit idle storage NIC bandwidth on decode nodes via RDMA, aiming at KV-cache movement bottlenecks for agentic long-context inference; includes claimed speedups (e.g., 1.87 on DS-660B) with caveats for smaller models (ZhihuFrontier).Kernel/infra chatter (quack, Liger): A thread points to Dao-AILabs quack writeup on memory hierarchy bandwidth, plus a note that Liger not using cluster-level reductions for xentropy could explain slower performance in some settings (fleetwood___).Off-policy RL for reasoning (Databricks MosaicAI): Databricks promotes OAPL (Optimal Advantage-based Policy Optimization with lagged inference policy) as a stable off-policy alternative that can match/beat GRPO while using ~3 fewer training generations, positioned as operationally simpler than strict on-policy loops (DbrxMosaicAI, jefrankle).ERL vs RLVR (Turing Post explainer): A long workflow breakdown contrasts standard RLVR (scalar verifiable rewards) with Experiential Reinforcement Learning (ERL) inserting within-episode reflection/retry + distillation; cites reported gains (e.g., +81% Sokoban) and tradeoffs (pipeline complexity/compute) (TheTuringPost).Mamba-2 / GDN initialization bug discussion: Albert Gu clarifies a viral plot debate: main takeaway is an init bug materially affecting some results; also notes nuanced interactions in hybrids (e.g., stronger components can make others lazy, with a related reference) (_albertgu, _albertgu).Top tweets (by engagement, technical / industry-relevant)OpenAI raises 110B (sama, OpenAI)Sakana AI Doc-to-LoRA / Text-to-LoRA (SakanaAILabs, hardmaru)AnthropicDoD supply-chain designation critique / governance precedent (deanwball, quantian1, janleike)Karpathy on coding workflow evolution (tab agents parallelism) (karpathy)Karpathy on programming a research org with multi-agent workflows; limitations observed (karpathy)Anthropic official statement (AnthropicAI)AI Reddit Recap/r/LocalLlama + /r/localLLM Recap Read moreRelated Articles
[AINews] AI Engineer will be the LAST job
Swyx · explanation · 76% similar
[AINews] Replit Agent 4: The Knowledge Work Agent
Swyx · explanation · 76% similar
[AINews] NVIDIA GTC: Jensen goes hard on OpenClaw, Vera CPU, and announces $1T sales backlog in 2027
Swyx · explanation · 76% similar
Originally published at https://www.latent.space/p/ainews-openai-closes-110b-raise-from.