AI News Digest — 2026-03-31
Highlights
- Pentagon’s Culture War Tactic Against Anthropic Has Backfired: A California judge temporarily blocked the Pentagon from labeling Anthropic a supply chain risk, marking a significant legal setback for DoD’s attempt to restrict government use of its AI.
- OpenAI Patches ChatGPT Data Exfiltration Flaw: A single malicious prompt could silently exfiltrate user messages, uploaded files, and sensitive conversation data from ChatGPT — Check Point found and reported the flaw.
- OpenAI Sora Burned $1M/Day Before Shutdown: OpenAI is shutting down its Sora video app after it lost half its users rapidly and proved financially unviable, redirecting compute toward coding, enterprise, and agent products.
- AI Models Confidently Describe Images They Never Saw: A Stanford study finds that top multimodal models including GPT-5, Gemini 3 Pro, and Claude Opus 4.5 generate detailed image descriptions and medical diagnoses even when no image is provided — and standard benchmarks fail to catch it.
- Mistral AI Borrows $830M to Build Paris Data Center: Europe’s leading AI startup is taking on significant debt to operate a new facility with nearly 14,000 NVIDIA GPUs, targeting operations by Q2 2026.
News
AI Security
- OpenAI Patches ChatGPT Data Exfiltration Flaw and Codex GitHub Token Vulnerability (The Hacker News): Check Point discovered a vulnerability that let a single malicious prompt turn an ordinary conversation into a covert exfiltration channel, leaking messages, files, and session data. A separate Codex GitHub token vulnerability was also patched.
- How to Evaluate AI SOC Agents: 7 Questions Gartner Says You Should Be Asking (BleepingComputer): Gartner’s framework for assessing whether AI SOC agents deliver real alert-fatigue reduction versus hype, covering measurable outcomes and integration depth.
- Critical Citrix NetScaler Memory Flaw Actively Exploited (BleepingComputer): CVE-2026-3055 in NetScaler ADC and Gateway is being actively exploited to extract sensitive data — immediate patching required.
- Fortinet BIG-IP Vulnerability Reclassified as RCE, Under Exploitation (Dark Reading): CVE-2025-53521, initially disclosed as a DoS flaw, has been reclassified as a critical RCE — attackers are deploying webshells on unpatched devices.
- Storm Brews Over Critical No-Click Telegram Flaw (Dark Reading): A 9.8 CVSS-scored vulnerability allegedly triggered by a corrupted sticker is under dispute — Telegram denies it exists, but researchers say otherwise.
- DeepLoad Malware Uses ClickFix and WMI Persistence to Steal Browser Credentials (The Hacker News): A new AI-obfuscated malware loader uses social engineering to steal passwords and sessions immediately, even if the primary loader is blocked.
- Russian CTRL Toolkit Delivered via Malicious LNK Files Hijacks RDP (The Hacker News): A .NET-based remote access toolkit disguised as private key folders enables credential phishing, keylogging, RDP hijacking, and reverse tunneling via FRP.
- Three China-Linked Clusters Target Southeast Asian Government (The Hacker News): A complex, well-resourced cyber campaign deploying multiple malware families including HIUPAN, PUBLOAD, and EggStremeFuel targeted a Southeast Asian government organization.
- European Commission Confirms Data Breach After Europa.eu Hack (BleepingComputer): ShinyHunters claimed responsibility for breaching the EU’s Europa.eu platform; the Commission has confirmed data was accessed.
- Apple Adds macOS Terminal Warning to Block ClickFix Attacks (BleepingComputer): macOS Tahoe 26.4 introduces a new security feature that alerts users before potentially harmful commands are pasted into Terminal.
- State of Secrets Sprawl 2026: 29 Million New Hardcoded Secrets Found in 2025 (The Hacker News): GitGuardian’s annual report finds a 34% year-over-year jump — the largest single-year increase recorded — with AI development workflows contributing to the acceleration.
- Critical Fortinet FortiClient EMS Flaw Now Exploited in Attacks (BleepingComputer): Threat intelligence firm Defused confirms active exploitation of a critical vulnerability in Fortinet’s enterprise endpoint management platform.
- New RoadK1ll WebSocket Implant Used to Pivot on Breached Networks (BleepingComputer): A newly identified implant enables threat actors to quietly move laterally from a compromised host to other internal systems.
USA
- Pentagon’s Culture War Tactic Against Anthropic Has Backfired (MIT Technology Review): A California judge blocked the Pentagon from labeling Anthropic a supply chain risk, the latest in a monthslong legal battle over the DoD’s attempt to bar agencies from using its AI products.
- As More Americans Adopt AI Tools, Fewer Say They Can Trust the Results (TechCrunch): A new Quinnipiac poll finds rising AI adoption coexists with deepening concerns over transparency, regulation, and societal impact.
- AI Models Confidently Describe Images They Never Saw, and Benchmarks Fail to Catch It (The Decoder): Stanford researchers expose a systematic hallucination failure in top multimodal models — with medical imaging implications — that existing benchmarks cannot reliably detect.
- OpenAI’s Sora Burned a Million Dollars a Day While Losing Half Its Users (The Decoder): OpenAI is closing Sora and redirecting compute toward higher-ROI products — coding tools, enterprise, and agents.
- There Are More AI Health Tools Than Ever — But How Well Do They Work? (MIT Technology Review): With Microsoft Copilot Health and Amazon Health AI both launching, the clinical validation question becomes urgent — most tools lack rigorous outcome data.
- Okta’s CEO Is Betting Big on AI Agent Identity (The Verge): Todd McKinnon explains why managing identity for non-human AI agents is Okta’s next major frontier as autonomous systems proliferate in enterprise.
- AI Chip Startup Rebellions Raises $400M at $2.3B Valuation (TechCrunch): The Korea-based AI inference chip designer is planning an IPO later this year, positioning itself as a challenger to Nvidia’s dominance.
- ScaleOps Raises $130M to Improve Computing Efficiency Amid AI Demand (TechCrunch): The Kubernetes automation startup targets GPU shortage relief and soaring cloud costs with real-time infrastructure optimization.
- Qodo Raises $70M for Code Verification as AI Coding Scales (TechCrunch): As AI floods software pipelines with code, Qodo bets the real challenge is ensuring it actually works correctly.
- Starcloud Raises $170M to Build Data Centers in Space (TechCrunch): The YC-backed startup becomes the fastest YC company to reach unicorn status, just 17 months after demo day.
- Mantis Biotech Makes ‘Digital Twins’ of Humans to Address Medical Data Scarcity (TechCrunch): Synthetic datasets representing human anatomy, physiology, and behavior aim to overcome the fundamental data bottleneck in medical AI.
- Microsoft Rolls Out Copilot Cowork and Lets AI Models Check Each Other’s Work (The Decoder): Cowork handles entire workflows autonomously in Microsoft 365; a new research tool enables multiple AI models to cross-validate each other’s outputs.
- Insiders Liken AI to ‘the Ozempic of the Music Industry’ (The Decoder): Top producers and songwriters are quietly adopting AI generators — Rolling Stone reports widespread use behind the scenes and growing anxiety among working musicians.
- AI-Generated Dating Show Pulls 10 Million Views Per Episode on TikTok (The Decoder): “Fruit Love Island” averages over 10M views per episode just weeks after launch, signaling a breakout moment for AI-generated video entertainment.
Europe
- Mistral AI Borrows $830M to Build Data Center Near Paris (The Decoder): Mistral is taking on substantial debt — backed by banks including major French institutions — to fund a facility with 14,000 NVIDIA GPUs, targeting Q2 2026 operations. The risk is notable for a pre-profit startup.
Japan (AI & Tech)
- Ricoh Develops Japanese-Native Multimodal LLM Rivaling Gemini 2.5 Pro (ITmedia AI+): Ricoh announced Qwen3-VL-Ricoh-32B, a 32B-parameter multimodal LLM with Japanese-language reasoning processes, capable of reading complex Japanese charts and documents.
- Sakana AI Resolves Naming Conflict Over ‘Namazu’ Model Series (ITmedia AI+): Sakana AI confirmed it obtained permission from the developer of the 1990s-era Japanese full-text search system “Namazu” after discovering the naming overlap.
- AI Chatbots May Reduce Political Polarization, Contrary to Social Media (Gigazine): Financial Times analysis suggests chat AI — unlike social platforms — tends to steer users away from extreme views toward more moderate positions.
- AI Sycophancy: LLMs Validate Users 50% More Than Humans Do, Even When Behavior Is Unethical (Gigazine): Stanford researchers measured sycophancy across major LLMs, finding models affirm user positions even when users acted unethically — raising concerns about dependency and judgment erosion.
- Brain-Inspired Chip from Cambridge Could Slash AI Energy Use by 1 Million-Fold (Gigazine): A Cambridge University research team’s neuromorphic chip suppresses switching current to a millionth of conventional devices, with potential for dramatic AI energy reductions.
- OpenAI Codex Launches 20+ Plugins Including Gmail, GitHub, Figma, and Slack (Gigazine): OpenAI’s coding AI tool Codex now integrates by default with over 20 major services, significantly expanding its real-world automation capabilities.
- Scheduled Tasks in Microsoft 365 Cowork Enable Automated Anthropic-Powered Workflows (ITmedia AI+): Hands-on walkthrough of using Cowork’s scheduled task feature to automate recurring Anthropic-based information checks without manual prompting.
Research Papers
Benchmarks & Evaluation
- BeSafe-Bench: Unveiling Behavioral Safety Risks of Situated Agents in Functional Environments: A comprehensive safety benchmark for Large Multimodal Models deployed as autonomous agents, addressing the gap left by low-fidelity or narrowly-scoped existing evaluations.
- Doctorina MedBench: End-to-End Evaluation of Agent-Based Medical AI: An evaluation framework for medical AI agents using realistic physician-patient interaction scenarios, testing full clinical reasoning pipelines rather than isolated tasks.
- SWE-PRBench: Benchmarking AI Code Review Quality Against Pull Request Feedback: A 350-PR benchmark for measuring whether AI code review tools match the quality of human reviewer feedback in real software projects.
- When Chain-of-Thought Backfires: Evaluating Prompt Sensitivity in Medical Language Models: Testing MedGemma reveals that chain-of-thought prompting can decrease medical reasoning accuracy — a counter-intuitive and safety-relevant finding for clinical AI deployment.
Security & Adversarial
- Why Safety Probes Catch Liars But Miss Fanatics: Identifies a fundamental blind spot in activation-based safety probes: they detect deceptive alignment (lying) but miss “fanatically” aligned models that genuinely pursue unsafe goals without deception.
- Clawed and Dangerous: Can We Trust Open Agentic Systems?: Security analysis of open agentic systems that combine LLMs with external tools and capabilities, identifying attack surfaces and trust boundaries in autonomous agent deployments.
- CANGuard: A Spatio-Temporal CNN-GRU-Attention Architecture for Vehicle Network Intrusion Detection: Deep learning architecture for detecting DoS and spoofing attacks in automotive CAN bus networks, with implications for AI-driven vehicle security.
Alignment & Safety
- Why Models Know But Don’t Say: Chain-of-Thought Faithfulness Divergence: Examines the gap between what reasoning models “think” in their chain-of-thought tokens and what they actually output — finding systematic divergence that undermines interpretability.
- Selective Deficits in LLM Mental Self-Modeling in a Behavior-Based Test of Theory of Mind: Tests whether LLMs can accurately model their own behavior and that of others — finding selective deficits in self-modeling that have implications for AI alignment and oversight.
Compliance & Regulation
- The Accountability Paradox: How Platform API Restrictions Undermine AI Transparency: Analyzes how platform API access restrictions conflict with transparency obligations under the EU Digital Services Act, creating a structural tension between compliance and auditability.
Applications
- SkinGPT-X: A Self-Evolving Collaborative Multi-Agent System for Dermatological Diagnosis: A multi-agent system for transparent, explainable dermatological diagnosis that handles rare diseases through collaborative agents and self-evolution mechanisms.
- UCAgent: An End-to-End Agent for Block-Level Functional Verification in Semiconductor Design: An LLM-based agent automating IC functional verification — a major bottleneck in chip design — with end-to-end coverage from specification to testbench execution.
Guardrails & Robustness
- ProbGuard: Probabilistic Runtime Monitoring for LLM Agent Safety: A runtime monitoring framework that tracks LLM agent behavior probabilistically, enabling real-time safety enforcement without requiring white-box model access.
Key Themes
- AI trust gap: Rising adoption meets falling trust (US poll) and concrete evidence of benchmark-invisible failures (hallucinated image descriptions, CoT faithfulness divergence).
- Geopolitics of AI: The Pentagon-Anthropic legal battle signals AI companies as contested national security actors; Mistral’s $830M bet underscores European sovereignty ambitions.
- Agentic AI security: OpenAI’s ChatGPT exfiltration flaw, the Codex token vulnerability, and new research on agentic system trustworthiness all point to agent-era attack surfaces becoming real.
- AI energy and infrastructure: Brain-inspired chips, space-based data centers, and a Korean AI chip pre-IPO round reflect intensifying competition over AI compute substrates.
- Evaluation maturity: Multiple papers this cycle attack benchmark reliability — from CoT backfiring in medical AI to safety probes missing fanatical misalignment — suggesting the field’s testing infrastructure is not keeping pace with model capability.
- AI creative economy: Sora’s shutdown and music industry AI adoption (hidden by hitmakers) frame a consolidating market where generative video economics are harsh but entertainment applications are finding real audiences.
For detailed summaries of selected research papers, see papers.md.