AI Infra Dao

AI Infra Brief|EU AI Grid Expansion and Cost-Cutting Agent Infra (2026.02.15)

February 15, 2026 — I’m tracking fresh momentum across sovereign AI buildouts, GPU capacity, agent-ready web infrastructure, and open source efficiency gains — all within the past 48 hours.

🧭 Key Highlights

🇪🇺 EU AI Grid expands to Latvia, Estonia, Finland, Germany, and Italy

🖥️ HIVE BUZZ signs $30M in customer AI GPU agreements over two years

💰 Rizz Network secures $5M capital commitment for AI-enabled wireless expansion

🚀 MiniMax M2.5 posts SOTA results — BrowseComp 76.3%, SWE‑Bench 80.2%

🧠 Gemini 3 Deep Think debuts extended reasoning mode — ARC‑AGI2 84.6%

📄 Cloudflare “Markdown for Agents” converts HTML to Markdown with ~80% token reduction

💾 Pierre “Code Storage” soft-launches API-first Git layer for agent coding

⚠️ ICML prompt injection controversy: every paper in batch had injected instructions

🔧 AdaLLM open-sources NVFP4 inference for RTX 4090 — 2.4x lower peak VRAM

Sovereign AI & Computing Infrastructure

🇪🇺 EU AI Grid Expands to Five Countries

According to Manilatimes, the EU AI Grid expands to Latvia, Estonia, Finland, Germany, and Italy, extending beyond the initial Vilnius deployment. The federated network treats AI as a metered utility via the TokenVisor commercial platform, aiming to keep value and control within Europe.

The EU AI Grid expansion marks substantive progress in Europe’s sovereign AI infrastructure, with a federated model allowing countries to share computing resources while maintaining data sovereignty. TokenVisor platform provides AI capabilities as a metered utility similar to water and electricity.

🖥️ HIVE BUZZ Signs $30M in Customer AI GPU Agreements

According to Insidehpc, HIVE BUZZ signs $30M in customer AI GPU agreements over two years — 504 liquid-cooled Dell server-based GPUs slated for Manitoba, targeted online by March 31, 2026, with ~$15M in ARR projected.

Liquid-cooled GPU deployment reflects AI data centers’ evolution toward more efficient cooling technologies, with Manitoba’s location likely leveraging its low-temperature climate and clean energy advantages.

💰 Rizz Network Secures $5M Capital Commitment for AI-Enabled Wireless Expansion

According to Streetinsider, Rizz Network secures a $5M capital commitment from Nimbus Capital for $RZTO to power Rizz Wireless, an MVNO blending telecom infrastructure with blockchain and AI using AI-driven analytics for rewards and efficiency.

The fusion of AI with telecom infrastructure represents a new network service paradigm, with MVNOs (Mobile Virtual Network Operators) using AI to optimize user experience and operational efficiency.

Model Performance & Inference

🚀 MiniMax M2.5 Posts SOTA Results

According to Patmcguinness, MiniMax M2.5 posts SOTA results — BrowseComp 76.3%, SWE‑Bench Verified 80.2%, priced at $0.30 per million input tokens.

MiniMax M2.5’s performance improvements and pricing strategy reflect intensifying competition between open-source and frontier vendor models, with $0.30 per million input tokens positioning it at a cost-advantaged tier.

🧠 Gemini 3 Deep Think Debuts Extended Reasoning Mode

According to Patmcguinness, Gemini 3 Deep Think debuts an extended reasoning mode — ARC‑AGI2 84.6%, Codeforces 3455.

Extended reasoning mode represents progress in AI systems’ complex task-solving capabilities, with ARC‑AGI2 and Codeforces serving as important benchmarks for reasoning and programming abilities.

Agent Infrastructure & Web Optimization

📄 Cloudflare “Markdown for Agents” Achieves ~80% Token Reduction

According to Linksurge, Cloudflare launches “Markdown for Agents” converting HTML to Markdown on the edge when requested via Accept: text/markdown — ~80% token reduction (e.g., ~16k → ~3k), lower RAG costs, new machine-readable headers (ai-train, search, ai-input), and a noted “AI cloaking” risk.

Agents need to read large volumes of web content, and HTML-to-Markdown conversion dramatically reduces token consumption, directly lowering operational costs for RAG (Retrieval-Augmented Generation) systems. New machine-readable headers provide finer content control for agents, but “AI cloaking” risks (malicious content optimized specifically for agents) require vigilance.

💾 Pierre “Code Storage” Soft-Launches API-First Git Layer

According to Hacker News, Pierre “Code Storage” soft-launches an API-first Git layer for agent coding at massive scale (single createCommitFromDiff API, native grep/archive/ephemeral branches); community flags potential storage ingress cost concerns.

Agent coding requires specialized code storage infrastructure, and Pierre’s API-first design simplifies agent operations on Git repositories, though storage costs become a concern for scaled deployment.

Open Source Ecosystem & Efficiency

🔧 AdaLLM Open-Sources NVFP4 Inference for RTX 4090

According to Reddit, AdaLLM open-sources NVFP4 inference for RTX 4090 targeting Qwen3 and Gemma3 — 2.4x lower peak VRAM vs. FP16, ~297 tok/s for Qwen3‑8B, eliminating FP16 fallbacks.

NVFP4 precision format achieves high-performance inference on consumer-grade GPUs, lowering hardware barriers for local deployment. ~297 tokens/second throughput makes Qwen3‑8B practical on RTX 4090.

Security & Risk

⚠️ ICML Prompt Injection Controversy

According to Reddit, ICML prompt injection controversy: a reviewer reports every paper in their batch had injected instructions embedded in PDF metadata, triggering debate on detection vs. deterrence.

Prompt injection has permeated from AI application layers to academic publishing workflows, with researchers attempting to influence review processes by embedding instructions in paper PDFs, reflecting new security challenges for AI systems in content moderation and metadata processing.

🔍 Infra Insights

Today’s news collectively points to core AI infrastructure trends: substantive expansion of sovereign AI infrastructure and systematic cost optimization of agent infrastructure.

Sovereign AI construction is moving from concept to deployment, with the EU AI Grid’s five-country expansion and federated operational model providing a practical template for sovereign AI in other regions. HIVE BUZZ’s $30M GPU agreements and Rizz Network’s AI wireless financing confirm computing infrastructure’s evolution toward regionalization and specialization.

Cost optimization in agent infrastructure has become an innovation hotspot. Cloudflare’s Markdown for Agents achieves 80% token reduction through edge conversion, directly lowering RAG costs; Pierre’s API-first code storage layer simplifies agent coding workflows. Both embody the “infrastructure designed for agents” philosophy — evolving from human-readable to machine-readable, from interaction-friendly to programmatic-friendly.

Open-source models continue catching up to frontier vendors. MiniMax M2.5’s SOTA results on BrowseComp and SWE‑Bench, combined with $0.30 per million token pricing, reflect the open-source ecosystem’s dual progress in performance and cost competitiveness. AdaLLM’s NVFP4 inference achieves high performance on consumer GPUs, lowering local deployment barriers.

The ICML prompt injection controversy reveals a new dimension of AI security — attack surfaces expanding from model and application layers to content production and distribution workflows. Challenges faced by academic publishing systems presage the necessity of AI-native content moderation infrastructure.

Overall, developments in the past 48 hours show AI infrastructure evolving in three directions: sovereignization (local control), agent-ization (optimized for AI systems), and democratization (open-source lowering barriers).