← Radar Archive

Industry Signals

Supply Chain Β· Physical AI Β· Digital Simulation Microsoft Industry Blog Β· 24 Mar 2026

Microsoft Supply Chain 2.0: Physical AI, Simulation Agents, and a Three-Layer Operational Architecture

Microsoft published a detailed technical blueprint for what it calls "Supply Chain 2.0" β€” the convergence of physical artificial intelligence (AI) (NVIDIA Omniverse digital twins), simulation agents, and agentic planning on Fabric IQ and Work IQ. The architecture has three layers: (1) a physics-simulation layer (digital twin synchronised with physical operations via Internet of Things (IoT) sensor feeds), (2) a planning agent layer (agents with access to the operational ontology and solver application programming interfaces (APIs)), and (3) an execution layer (agents that dispatch to enterprise resource planning (ERP), warehouse management system (WMS), and transportation management system (TMS) systems within constraint boundaries). Physical AI refers to models trained on sensor and production data rather than text β€” enabling more accurate demand sensing and disruption prediction than statistical forecasting alone.

Why it matters: The three-layer architecture is a practitioner blueprint that makes explicit where optimisation logic sits: the planning agent layer calls solver APIs to generate feasible plans within constraint boundaries. Physical AI improves the quality of input parameters (demand forecasts, yield rates, lead times) fed to the optimisation layer β€” better inputs reduce uncertainty set size, which directly tightens the gap between a robust plan and the deterministic ideal. This is the same Microsoft intelligence stack underpinning the Capgemini CargoPilot Agent (covered 6 April): today's signal provides the architectural blueprint; that issue provided the commercial instantiation.

β†’ Microsoft Industry Blog: Supply Chain 2.0
Logistics Β· Agentic AI Β· Production Execution Logistics Viewpoints Β· 1 Apr 2026

Q1 2026 Supply Chain AI Execution: Constrained Agents Enter Live Operations at UPS, FedEx, and DHL

Logistics Viewpoints' Q1 2026 supply chain AI execution report confirms that agentic AI has moved from planning tools into live operational workflows at the world's three largest carriers. UPS runs a real-time route re-optimisation agent that fires automatically when driver delays threaten downstream cascade failures. FedEx deploys an agent that re-sequences delivery routes using live traffic and weather data without requiring planner approval. DHL's compliance agent resolves cross-border documentation exceptions within four minutes of a shipment scan β€” down from an average of 3.2 hours. All three deployments share an identical architectural pattern: a constrained agent operating within a feasibility envelope set by an offline optimisation model, with automatic human escalation triggered only when proposed actions would breach the pre-computed constraint boundary.

Why it matters: The constrained-agent pattern is the production-safe architecture for real-time logistics AI: the solver runs offline to define the feasibility boundary; the agent operates online within it. This provides a clean answer to the "ML versus solver" question: they are complementary, not competitive. The offline Mixed Integer ProgrammingAn optimisation framework where some decision variables are constrained to integer values while others are continuous. MIP is the workhorse for combinatorial scheduling, routing, and resource allocation. β€” first explained April 8 2026. model sets the constraint envelope; the live ML agent makes fast decisions within it. For practitioners building agentic logistics systems, the key design question is not "which AI model" but "how do we compute the constraint boundary the agent must respect?"

β†’ Logistics Viewpoints: Q1 2026 Supply Chain Trends
Transportation Β· Route Optimisation Β· Logistics AI Trimble Β· Q1 2026

Trimble Transportation Pulse Report 2026: AI Route Optimisation Crosses from Pilot to Production Standard

Trimble's Transportation Pulse Report for 2026 confirms that AI-powered route optimisation has matured past the pilot phase into production-grade deployment across major carriers. The shift documented in the report moves beyond static route planning into dynamic, real-time optimisation: AI agents now autonomously handle routing, carrier vetting, invoice generation, customs documentation, and disruption flagging across the logistics workflow. McKinsey's 2025 data cited in the report puts documented results at 10-15% fuel cost reduction and 15-20% faster average delivery times for enterprises with production AI routing deployments. The key barrier identified is integration: connecting a route optimisation model to a transportation management system (TMS) consumes 40-60% of the project timeline, with data quality issues (named by 57% of respondents) as the primary operational obstacle. The underlying optimisation engine for commercial-grade route optimisation remains the Vehicle Routing Problem (VRP), solved via MIP or metaheuristic approaches depending on fleet scale and constraint density.

Why it matters: The integrality gap in VRP formulations varies significantly by problem structure. For small-to-medium instances with time windows and capacity constraints, tight MIP formulations can achieve gaps close to 1 and be solved exactly. At enterprise fleet scale (thousands of vehicles, millions of stops), gaps become large enough that exact MIP is intractable and metaheuristics or hybrid ML-VRP approaches are required, with optimality guarantees traded for speed. The Trimble report's integration-effort finding (40-60% of project time) is an indirect signal of this formulation complexity: a well-constructed VRP formulation with the right decomposition strategy is not a commodity that ships as a plug-and-play component. The 2026 production deployments at major carriers almost certainly involved significant OR engineering to achieve tractable integrality gaps at scale.

β†’ Trimble: Transportation Pulse Report 2026
Supply Chain Β· Trade Intelligence Β· Scenario Optimisation C3.ai Β· 9 Apr 2026

C3.ai Launches Tariff-Resilient Supply Chain Suite as 90-Day Pause Triggers Mass Replan Wave

Following the US administration's 90-day tariff pause announced April 9, C3.ai released its Tariff-Resilient Supply Chain artificial intelligence (AI) suite β€” a set of pre-built decision models that calculate landed costs across all sourcing, routing, and duty combinations under multiple tariff scenarios simultaneously. The suite runs parallel Mixed Integer ProgrammingAn optimisation framework where some decision variables are constrained to integer values while others are continuous. MIP is the workhorse for combinatorial scheduling, routing, and resource allocation. β€” first explained April 8 2026. optimisations across a tariff scenario set, returning a Pareto frontier of cost, speed, and carbon exposure rather than a single-point recommendation. Early customer data from three manufacturing firms in Q1 2026 shows an 8–14% reduction in tariff-related landed cost variance, primarily through proactive sourcing allocation shifts before tariff announcements rather than reactive rerouting after them.

Why it matters: Scenario optimisation under policy uncertainty is structurally equivalent to two-stage stochastic programming (covered 8 April): the first-stage decision (supplier allocation, safety stock levels) must be made before the tariff policy resolves; the second-stage recourse (rerouting, mode-shifting) executes after the uncertainty clears. The 90-day pause creates a specific decision window β€” practitioners who can run fast Mixed Integer Programming (MIP) scenario replanning before the pause expires will lock in sourcing arrangements under the current duty regime.

β†’ C3.ai: Build a Tariff-Resilient Supply Chain    β†’ ConversightAI: Tariffs and Decision Intelligence

Research Papers

Research Β· Energy Dispatch Β· GPU Optimisation πŸ“‹ Case Study arXiv:2512.06715 Β· Dec 2025

GPU-Accelerated Optimisation Solver for Unit Commitment: First-Order Methods for Energy Dispatch at Scale

The unit commitment problem in energy grid dispatch asks which generators should be online, how much each should produce, and at what times, subject to minimum up- and down-time constraints, ramp rates, reserve requirements, and network constraints. It is a large-scale Mixed Integer ProgrammingAn optimisation framework where some decision variables are constrained to integer values while others are continuous. Mixed Integer Programming (MIP) is the workhorse for combinatorial scheduling, routing, and resource allocation. β€” first explained April 8 2026. problem solved repeatedly in practice under tight time constraints. This paper presents a GPU-accelerated solver using first-order methods (closely related to PDHG) for unit commitment, demonstrating that GPU parallelism enables real-time dispatch decisions on problem scales previously requiring overnight batch runs. The approach achieves solution quality comparable to commercial MIP solvers on benchmark unit commitment instances, with runtimes that fit within the 5-15 minute operational dispatch windows used by system operators. The paper also characterises the integrality gap of standard unit commitment formulations, showing that LP relaxation quality varies significantly with the tightness of the startup cost linearisation and the formulation of ramping constraints.

What problem it solves: Grid operators run unit commitment solvers every 15-60 minutes to determine the generation dispatch schedule for the next several hours. Standard MIP solvers at national grid scale can exceed operational time limits, requiring warm-started heuristics that sacrifice optimality. GPU-accelerated first-order methods provide a complementary approach: faster LP relaxation solving allows better bound computation within the available time, tightening the effective integrality gap achievable in the dispatch window.

Why it matters: The paper characterises the integrality gap of standard unit commitment formulations, showing that LP relaxation tightness varies significantly with formulation choices β€” specifically the startup cost linearisation and the ramp constraint form. GPU-accelerated first-order methods reduce LP solve time per branch-and-bound node, enabling runtimes that fit within the 5–15 minute operational dispatch window for problem scales previously requiring overnight batch runs.

β†’ View Paper on arXiv
📋 Case Study — Friday Deep-Dive Source: paper benchmark — arXiv:2512.06715 Expand for full case ▼
Problem
A grid operator decides every 15–60 minutes which generators to run for the next several hours. At national grid scale, solving this to optimality previously required overnight batch runs. ↓ Expand to see formulation, difficulty, and practitioner takeaway
Research Β· Interior Point Methods Β· LP Theory Zanetti & Gondzio Β· arXiv 2025

A Factorisation-Based Regularised Interior Point Method Using the Augmented System: The Theoretical Basis for HiPO

This paper by Filippo Zanetti and Jacek Gondzio presents the algorithmic foundation of HiPO, the new interior point solver integrated into HiGHS 1.12. The method applies regularisation techniques to the augmented system formulation of the interior point algorithm for linear programs, enabling more robust and efficient factorisations on large, sparse LP instances. The augmented system approach avoids forming the normal equations (which can be numerically ill-conditioned for degenerate LP instances), instead factorising a larger but better-conditioned matrix. Multi-threading across the factorisation step delivers the more predictable running times that distinguish HiPO from existing interior point solvers on instances where degeneracy causes barrier method convergence to slow unpredictably.

What problem it solves: Interior point (barrier) methods for LP can stall near the optimum on degenerate instances, where many constraints are near-active simultaneously. Regularisation stabilises the augmented system factorisation in these regions, making convergence more predictable without sacrificing the polynomial-time theoretical guarantees of the barrier method. For practitioners using HiGHS inside branch-and-bound or Column Generation, HiPO's improved degeneracy handling translates to more consistent LP solve times across the search tree.

Why it matters: The HiPO paper provides the mathematical basis for the LP-solving improvement in HiGHS 1.12, the open-source solver covered alongside Gurobi 13.0 PDHG in the 6 April issue. For practitioners evaluating the Gurobi Primal-Dual Hybrid Gradient (PDHG) versus HiGHS HiPO choice for large-scale LP: PDHG excels on very large, unstructured LP instances where GPU parallelism dominates; HiPO excels on large but structured instances where degeneracy is a practical problem and factorisation quality matters more than raw compute throughput. The integrality gap remains unchanged either way, but both provide faster access to LP bounds that drive branch-and-bound convergence.

β†’ HiGHS: Official Site    β†’ HiGHS on GitHub (includes HiPO release notes)
Research Β· Portfolio Optimisation Β· ML+OR Nature Scientific Reports Β· 2025

A Machine Learning Approach to Risk-Based Asset Allocation in Portfolio Optimisation

This paper in Nature Scientific Reports proposes a dynamic asset allocation framework that integrates machine learning (ML) prediction with classical Markowitz-style portfolio optimisation. ML methods including deep reinforcement learning (RL) and ensemble models provide time-varying estimates of the covariance matrix and expected returns, which are fed into a mean-variance or mean-Conditional Value at Risk (CVaR) portfolio optimisation solve at each rebalancing period. The key contribution is a theoretical analysis showing that when the ML covariance estimates satisfy a bounded estimation error condition, the integrated ML-optimisation system achieves better risk-adjusted returns than either pure ML allocation or classical static optimisation alone. The paper compares performance across multiple benchmark asset universes and demonstrates that the integrality gap concept from discrete portfolio construction (where integer constraints arise from lot sizes, transaction costs, or cardinality limits) strongly predicts out-of-sample portfolio return degradation when LP relaxation solutions are used directly as allocation weights.

What problem it solves: Static Markowitz optimisation assumes that expected returns and covariances are known constants, which fails in practice because both are non-stationary. This paper treats parameter estimation as a machine learning problem and solves the downstream portfolio optimisation exactly at each period, making both the prediction and the decision step rigorous. The explicit treatment of integrality effects in constrained portfolio construction (lot sizes, cardinality) is rare in the ML-finance literature and directly relevant to practitioners building quantitative equity systems.

Why it matters: The finding that integrality gap magnitude predicts out-of-sample return degradation when LP relaxations are used directly as portfolio weights is an empirical result with immediate practitioner implications: it quantifies, in financial return terms, the cost of ignoring integer constraints. For quantitative teams evaluating whether to invest in exact Mixed Integer Programming (MIP) portfolio construction versus LP relaxation approximations, this paper provides the analytical framework to size the decision: the measured return gap from ignoring integer constraints is the cost of the approximation.

β†’ Nature Scientific Reports: ML for Risk-Based Asset Allocation

Term of the Day

Integer Programming Β· New Term First in this issue Β· 10 Apr 2026

Integrality Gap

The integrality gap of a mixed integer programming (MIP) instance is the ratio between the optimal value of the linear programming (LP) relaxation (obtained by removing all integrality constraints and allowing integer variables to take any real value in range) and the optimal value of the MIP itself. For a minimisation problem, the LP relaxation always provides a lower bound on the MIP optimal, so the integrality gap equals MIP_opt divided by LP_opt, a number always greater than or equal to 1. For a maximisation problem, the LP relaxation is always an upper bound, so the gap equals LP_opt divided by MIP_opt, again at least 1. A gap of exactly 1.0 means the LP relaxation is perfectly tight: the optimal LP solution already satisfies all integer constraints (or can be rounded to an optimal integer solution at no cost). A gap of 1.5 means the LP relaxation underestimates (for minimisation) or overestimates (for maximisation) the true optimum by 50%.

The integrality gap is not a solver performance metric. It is a property of the LP relaxation formulation of the specific problem instance. Formulations of the same underlying problem can have radically different integrality gaps depending on how constraints are modelled. Adding valid inequalities (cutting planes) tightens the LP relaxation, bringing LP_opt closer to MIP_opt and thereby reducing the gap. Reformulating the problem to use stronger variable bounds or disaggregated constraints can achieve a gap of 1 on instances where the standard formulation has a gap of 2 or more. This is the central reason why OR formulation expertise remains irreplaceable even as solvers become faster: no amount of hardware acceleration can close a gap that formulation has left open.

The integrality gap connects directly to today's solver releases. Gurobi 13.0's PDHG and HiGHS 1.12's HiPO make LP relaxation solving faster, which accelerates branch-and-boundA general algorithmic framework for solving combinatorial optimisation problems exactly, by recursively partitioning the feasible region and computing bounds to prune branches. Branch and Bound (B&B) is the core engine in all modern MIP solvers. β€” first explained April 8 2026. on well-formulated problems with small gaps. On problems with large gaps, faster LP solving allows more LP relaxations to be computed in the same time, but cannot prevent the exponential branching that a large gap implies. The gap also appears in Column GenerationAn algorithmic technique for solving large-scale linear programs and MIPs by iterating between a Restricted Master Problem and a pricing subproblem that finds the most improving variable to add. Column Generation (CG) is the foundation of Branch-and-Price. β€” first explained April 9 2026.: the quality of the LP relaxation solved at the restricted master problem level determines how quickly the pricing subproblem identifies improving columns, with a tight formulation converging faster than a loose one.

Why practitioners misread this

Confusion 1: The integrality gap versus the MIP gap in solver output. When a solver is running, it displays an optimality gap (also called the MIP gap): the difference between the best feasible integer solution found so far (the incumbent) and the best lower bound (for minimisation), expressed as a percentage of the incumbent. This gap changes throughout the solve and reaches 0 when optimality is proved. The integrality gap is a theoretical, fixed property of a specific LP relaxation versus the true integer optimum. You cannot read the integrality gap from the solver log, because computing it would require already knowing the MIP optimal value. Confusing the two leads practitioners to interpret a large running MIP gap as evidence of a bad formulation (possibly true) or to believe a gap of 0% at termination means the LP relaxation was tight (not implied).

Confusion 2: Hardware closes the integrality gap. Gurobi PDHG and HiGHS HiPO make LP solving faster. A faster LP solve means branch-and-bound can evaluate more nodes per second, potentially proving optimality faster on well-formulated problems. But on problems with large integrality gaps, the branching tree is exponentially larger regardless of LP speed: the solver must enumerate more nodes before finding and proving the optimum. Hardware acceleration reduces wall-clock time per node; it does not reduce the number of nodes required when the formulation is weak. Adding valid inequalities (cutting planes), tightening variable bounds, or using a stronger reformulation are the tools that reduce the integrality gap and therefore reduce the tree size. Practitioners who respond to a slow-solving MIP by upgrading hardware are often addressing the wrong bottleneck.

Confusion 3: Instance-level gap versus problem-class gap. The integrality gap of a specific MIP instance is a number. The integrality gap of a problem class (such as the Travelling Salesman Problem (TSP), vertex cover, or bin packing) refers to the worst-case ratio across all instances, which governs the approximation ratio achievable by LP-rounding algorithms. Some problem classes have constant integrality gaps (vertex cover has a gap of at most 2, which underlies the classic 2-approximation algorithm), while others have unbounded gaps. Practitioners sometimes refer to "the integrality gap of our routing problem" as if it were a constant, when in fact it varies by instance structure, constraint density, and how the formulation is constructed.

Diagnostic test for your formulation: Solve the LP relaxation of your MIP and compare the LP optimal value to the best integer solution you can find (either by running the MIP to proven optimality on a small instance, or by using a known upper bound for the problem class). If the ratio exceeds 1.2 to 1.3 for a minimisation problem, strong cutting planes or reformulation are likely to yield a larger improvement than hardware upgrades. If the gap is below 1.05, the formulation is already tight and solver speed becomes the primary lever.

Related:

Mixed Integer Programming Β· Branch and Bound Β· Column Generation Β· Cutting Planes (not yet in Concepts)
β†’ arXiv:2512.06715 β€” GPU solver paper with unit commitment integrality gap analysis    β†’ Gurobi: GPU LP Solving (PDHG explainer)

Upcoming Conferences

Conference Dates Location Key Track / Deadline
INFORMS Analytics+ 2026 Apr 12–14, 2026 National Harbor, MD Opens tomorrow. Edelman Award Gala (Apr 13). MIP, supply chain, and GenAI tracks. Edelman winner announced Monday (Apr 13).
AGIFORS 2026 Airline Ops Apr 13–17, 2026 Barcelona, Spain Starts Monday. Airline Operations Research (OR) under disruption and tariff shock. Crew & maintenance scheduling, demand uncertainty, recovery under constraints.
Hannover Messe 2026 Apr 20–24, 2026 Hannover, Germany AI in manufacturing, physical AI, autonomous industrial operations. CargoPilot and similar constrained-agent deployments showcased here.
MIP 2026 Workshop May 18–21, 2026 Stamford, CT (UConn) Flash talk deadline: Apr 15 (5 days). Land-Doig MIP Competition. Ideal venue to benchmark integrality gap formulations.
CPAIOR 2026 May 26–29, 2026 Rabat, Morocco 23rd International Conference on the Integration of Constraint Programming (CP), Artificial Intelligence, and Operations Research. CP-MIP hybrid architectures, scheduling, routing.
CP 2026 / FLoC 2026 (SOFT'26 workshop) Jul 2026 Lisbon, Portugal Principles and Practice of Constraint Programming, part of the Federated Logic Conference (FLoC) 2026. SOFT'26 workshop (discrete optimisation and ML) abstract deadline: May 15.
Daily Synthesis

All four signals describe the same structural shift: AI has moved from the advisory layer to the execution layer. Every deployment β€” Physical AI blueprints, constrained logistics agents, tariff scenario replanning β€” follows one pattern: an offline solver computes the feasibility boundary; an online agent acts within it.

  • Microsoft names the architecture in blueprint form; UPS, FedEx, and DHL confirm it ships in production β€” the constrained-agent pattern is no longer a design proposal, it is the documented production standard at three of the world’s largest carriers.
  • C3.ai’s tariff suite enumerates the full feasible space across duty assumptions simultaneously rather than returning a single-point recommendation; Trimble’s finding that AI routing requires 40–60% of project time on integration confirms that the real work is formulation engineering, not model selection.
  • The GPU unit commitment paper and HiPO show faster linear programming (LP) solving shortens the MIP boundary refresh interval β€” but neither dissolves the integrality gap; formulation quality remains the gate between a tractable and intractable agent operating boundary.

For practitioners: Measure your integrality gap before investing in faster infrastructure: solve the LP relaxation and the full MIP on a small representative instance and compute MIP_opt ÷ LP_opt (minimisation) or LP_opt ÷ MIP_opt (maximisation) β€” result is always ≥ 1.0.

Ratio > 1.15–1.20: the LP bound is loose; branch-and-bound is exploring unnecessary tree. Reformulate first β€” valid inequalities, tighter big-M values, or Dantzig-Wolfe decomposition will outperform any hardware upgrade.

Ratio < 1.05: LP solving dominates node cost; HiPO or PDHG GPU acceleration delivers proportional speedup and directly shortens envelope refresh intervals. Note: normal ranges differ by problem class β€” well-formulated lot-sizing typically sits at 1.01–1.03; vehicle routing problems (VRPs) at fleet scale routinely run 1.30–1.60 even with tight formulations.

Decision Optimisation Radar Β· nexmindai.org

← Back to Radar Archive