← GenAI Radar Archive

Industry Signals

Model Release · Restricted Access Anthropic · Apr 2026

Anthropic Launches Project Glasswing — Restricted Cybersecurity Model for Vetted Partners

Anthropic quietly launched Project Glasswing (also referred to internally as Claude Mythos), a cybersecurity-specialised model variant available only to a whitelist of more than 45 vetted partner organisations. Unlike standard Claude releases, access requires explicit clearance through Anthropic's safety review process. The model is optimised for offensive security research, vulnerability analysis, and red-teaming workflows where standard models are intentionally limited. The move reflects a broader industry trend toward tiered model access based on use-case legitimacy rather than flat commercial availability.

Why it matters: Restricted-access specialist models signal that labs are beginning to differentiate capability tiers by role and trust level rather than subscription tier alone. Security teams and red-teamers at non-partner organisations need to plan for a capability gap; those inside the partner network gain a significant edge in AI-assisted threat research.

→ Anthropic
Adoption · Developer Tools OpenAI · Sam Altman / X (Twitter) · Apr 2026

OpenAI Codex Hits 3 Million Weekly Active Users — Altman Announces Usage Reset

OpenAI's Codex coding assistant surpassed 3 million weekly active users, according to a post by CEO Sam Altman on X. Altman simultaneously announced that usage limits for Plus and Pro subscribers would be reset to accommodate demand spikes, acknowledging that the rate of adoption had exceeded provisioned capacity. Codex, which powers ChatGPT's code interpreter and API coding features, has become a key battleground for developer mindshare alongside competitors including GitHub Copilot, Google Gemini Code Assist, and Anthropic's Claude for coding.

Why it matters: Developer tool adoption at this scale creates compounding switching costs. Teams and individuals who integrate Codex deeply into their workflows become sticky users regardless of pricing changes. Adoption velocity is now a primary competitive metric alongside benchmark performance.

→ OpenAI / Sam Altman on X
Startup · AI Governance Modus / Lightspeed · Apr 2026

Modus Raises $85 Million for AI Audit Agents

Modus, a startup building autonomous AI audit agents, raised $85 million across a seed round and Series A led by Lightspeed Venture Partners. The company's agents continuously monitor enterprise AI deployments for compliance drift, output quality degradation, and policy violations without requiring human reviewers to manually sample outputs. The fundraise arrives as enterprises face increasing pressure from the European Union (EU) AI Act high-risk system audit requirements and internal governance mandates to demonstrate ongoing oversight of deployed models.

Why it matters: AI audit and governance tooling is becoming a category, not a niche. The combination of regulatory deadlines and the operational complexity of monitoring production AI systems is creating durable demand. Teams responsible for AI risk should begin evaluating third-party audit tooling alongside building internal capability.

→ Lightspeed / Modus
Edge AI · On-Device Google · Apr 2026

Google Ships Offline AI Dictation Using On-Device Gemma Models

Google shipped an offline AI dictation feature for Android powered entirely by on-device Gemma models, requiring no cloud connection for transcription or context-aware editing. The feature integrates with Gboard and allows users to dictate, correct, and reformat text locally, even in airplane mode. It represents the most prominent consumer deployment of the Gemma model family on-device to date, and a direct demonstration of Google's strategy to leverage open-weight models for private, low-latency on-device inference.

Why it matters: On-device AI eliminates a meaningful barrier for privacy-sensitive use cases: no data leaves the device. For developers and product teams, this validates the viability of small language models (SLMs) for production consumer features at scale. Expect the on-device capability bar to rise steadily as Gemma and competing SLM families improve.

→ Google AI / Gemma

Models & Tools

Model Preview · Open Source · MoE DeepSeek · Apr 2026

DeepSeek V4 Preview: ~1 Trillion Parameter MoE at $0.28 per Million Tokens

DeepSeek released a technical preview of V4, a Mixture of Experts (MoE) model with approximately 1 trillion total parameters trained on Huawei Ascend 950PR chips. The model will be released under the Apache 2.0 licence at launch, expected in late April 2026. Pricing is set at $0.28 per million (M) input tokens, a fraction of comparable frontier model costs. The Ascend 950PR hardware detail signals a deliberate strategy to build a training stack independent of NVIDIA, following U.S. export controls limiting access to H100 and H200 graphics processing units (GPUs).

Why it matters: A trillion-parameter open-weight MoE at sub-$0.30 per million input tokens would be a dramatic price-performance dislocation if benchmarks hold. Teams currently paying frontier API rates should model the cost implications of a capable, Apache 2.0 alternative arriving within weeks. The Ascend hardware angle also matters geopolitically: it demonstrates viable large-scale AI training outside the NVIDIA ecosystem.

→ DeepSeek

Safety & Security

Workforce · AI Displacement Challenger, Gray & Christmas / Multiple Sources · Apr 2026

79,000 Tech Jobs Eliminated in Q1 2026 — Half Attributed to AI Automation

Outplacement firm Challenger, Gray and Christmas reported that 79,000 technology sector jobs were eliminated in the first quarter (Q1) of 2026, with approximately half of those layoffs explicitly attributed by employers to AI automation replacing roles. The pace is the fastest since the post-pandemic correction of 2023. Functions most affected include customer support, QA testing, content production, and entry-level software engineering. The data follows a broader pattern of enterprises accelerating headcount reductions as AI-assisted productivity tools reach sufficient reliability for production deployment.

Why it matters: AI displacement is no longer a projection: it is appearing in quarterly filings and outplacement data. Teams and organisations need to be proactive about reskilling strategies, particularly in functions where AI assistance is reaching task-completion quality. The human-in-the-loop question is shifting from "if" to "which tasks and at what ratio."

→ Challenger, Gray & Christmas

Policy & Regulation

Regulation · EU AI Act European Commission · Apr 2026

EU AI Act High-Risk System Compliance Deadline Looms — August 2026

The EU AI Act's compliance obligations for high-risk artificial intelligence (AI) systems enter full force in August 2026, with enterprises now in the final four months of a mandatory transition window. High-risk categories include AI used in recruitment, credit scoring, critical infrastructure, biometric identification, and educational assessment. Non-compliant organisations face fines of up to 3% of global annual turnover. Enterprise legal and AI governance teams are accelerating documentation, conformity assessment, and human oversight mechanism implementations to meet the deadline. The EU's AI Office has begun publishing updated guidance on the audit trail and record-keeping standards expected from deployers.

Why it matters: August 2026 is a hard legal deadline, not a soft guideline. Organisations deploying AI in any of the high-risk categories within the EU need to confirm compliance readiness now. The documentation burden alone is substantial; teams without existing AI governance infrastructure will struggle to self-certify in time without third-party tooling or consultancy support.

→ European Commission — EU AI Act

Term of the Day

Architecture · Scaling

Mixture of Experts (MoE)

A Mixture of Experts (MoE) is a neural network architecture that partitions a model's parameters into multiple distinct subnetworks, called "experts," and uses a learned routing function (the "gating network") to activate only a small subset of those experts for any given input token or sequence. Rather than passing every input through all parameters — as in a dense transformer — an MoE model selects, say, 2 or 8 experts out of potentially hundreds or thousands, performing computation only on those selected paths.

This design achieves a key efficiency win: total parameter count (and thus model capacity and knowledge) can be scaled dramatically without a proportional increase in floating-point operations (FLOPs) per forward pass. A 1 trillion total-parameter MoE model might activate only 20-50 billion parameters per token, giving it the inference cost profile of a far smaller dense model while retaining the expressiveness of a much larger one.

MoE architectures present distinctive engineering challenges: expert load balancing (preventing the router from over-relying on a few experts), communication overhead when experts are distributed across hardware, and training instability that can arise when gating decisions are differentiable but sharp. Modern implementations use auxiliary loss terms to encourage expert utilisation diversity and top-k routing (typically top-2 or top-8) to keep activation patterns tractable.

Why it is relevant today: DeepSeek's V4 preview puts MoE squarely in focus. The architecture is also the foundation of GPT-4 (undisclosed at launch), Mistral's Mixtral series, and Google's Gemini models. As the cost/capacity tradeoff becomes a central competitive dimension in 2026, understanding how MoE models behave differently from dense models at inference time is practical knowledge for anyone designing AI pipelines or evaluating model costs.

Research Papers

Safety · Alignment · Reasoning arXiv:2503.00555 · March 2025

Safety Tax: Safety Alignment Makes Your Large Reasoning Models Less Reasonable

Huang et al. — Georgia Tech / Emory

This paper identifies a fundamental tension in aligning large reasoning models (LRMs): applying conventional safety alignment (via DirectRefusal datasets and RLHF-style fine-tuning) reliably reduces harmful output but measurably degrades reasoning capability across standard benchmarks. The authors term this penalty the "Safety Tax" and demonstrate that the tradeoff worsens as reasoning model size increases, raising structural questions about whether the same alignment techniques that work for general-purpose language models are appropriate for reasoning-optimised architectures.

Practitioner takeaway: Teams deploying LRMs in high-reasoning-demand contexts (code, math, planning) should be aware that safety fine-tuning may degrade the capabilities they are deploying for. The paper motivates research into alignment methods that preserve chain-of-thought integrity while still enforcing behavioural constraints.

→ arXiv:2503.00555
Safety · Reinforcement Learning · Red-Teaming arXiv:2601.18292 · January 2026

TriPlay-RL: Tri-Role Self-Play Reinforcement Learning for LLM Safety Alignment

Tan et al.

TriPlay-RL proposes a closed-loop reinforcement learning (RL) framework in which three model roles — attacker, defender, and evaluator — iteratively co-evolve without manual annotation. The attacker generates adversarial prompts, the defender learns to respond safely, and the evaluator scores outcomes, driving all three roles to improve over successive iterations. The approach significantly outperforms static red-teaming baselines and RLHF-based safety fine-tuning on adversarial safety benchmarks, and can improve alignment without degrading general capability by as much as standard fine-tuning does.

Practitioner takeaway: Self-play alignment is an attractive alternative to expensive human red-teaming at scale. TriPlay-RL's tri-role setup is architecturally transferable and could be implemented on top of open models. Read alongside the Safety Tax paper for context on why annotation-free approaches that preserve reasoning quality are drawing increasing research attention.

→ arXiv:2601.18292
Architecture · Survey · MoE arXiv:2507.11181 · July 2025

Mixture of Experts in Large Language Models: A Comprehensive Review

Zhang et al.

A comprehensive review of the Mixture of Experts (MoE) architecture as applied to large language models, covering routing mechanisms (top-k, expert choice, hash routing), load balancing strategies, hierarchical MoE designs, and integration with multimodal and multi-task learning frameworks. The paper catalogues open research challenges including expert specialisation measurement, sparse-to-dense distillation, and calibration of gating networks at deployment time. A useful reference for practitioners evaluating MoE-based models or building on top of open MoE architectures.

Practitioner takeaway: With DeepSeek V4 and the broader MoE wave accelerating, this survey is timely reading for engineers who need to understand what they are deploying and optimising. Pay particular attention to the load balancing and distillation sections — these directly affect inference reliability and cost when serving MoE models in production.

→ arXiv:2507.11181