← Radar Archive

Industry Signals

LLM+OR · MIP Formulation · Enterprise AI Microsoft Research · Jan 2026

Microsoft OptiMind: 20B-Parameter LLM Turns Natural Language into Solver-Ready Mixed Integer Programming Models

Released in January 2026 under the MIT license, Microsoft Research's OptiMind is a 20-billion-parameter Mixture of Experts model that converts natural language descriptions of complex decision problems directly into Gurobi-ready Mixed Integer ProgrammingAn optimisation framework where some decision variables are constrained to integer values while others are continuous. Mixed Integer Programming (MIP) is the workhorse for combinatorial scheduling, routing, and resource allocation. — first explained April 8 2026. code. OptiMind generates a reasoning trace, a formal mathematical formulation, and executable GurobiPy code, and supports a multi-turn correction loop: solver errors and logs are fed back to the model, which revises the formulation iteratively. The base architecture is a fine-tuned version of OpenAI's OSS 20B model (microsoft/OptiMind-SFT) trained on curated optimisation datasets.

Why it matters: OptiMind directly attacks the formulation bottleneck in applied OR: problem structuring and model specification consume 60-80% of project time in typical enterprise deployments. Multi-turn solver correction is the critical differentiator from prior LLM-to-code tools, because formulation errors surface at runtime, not at generation time. The key limitation, and where expert OR knowledge remains irreplaceable, is structural decomposition: OptiMind generates standard mixed integer programming (MIP) formulations well, but Column Generation, Benders Decomposition, and Lagrangian Relaxation require architectural insight about the problem's mathematical structure that current LLMs cannot yet reliably produce.

→ MarkTechPost: OptiMind Release
Supply Chain · AI Strategy · Decision Intelligence BCG · MIT · Gartner · 2026

2026 Is the Supply Chain AI ROI Accountability Year: 95% of GenAI Pilots Deliver Zero Measurable Return

BCG's 2026 supply chain planning analysis confirms that the AI investment cycle is reaching a reckoning: for procurement and supply chain leaders, 2026 is the year that separates organisations that can demonstrate return on investment from those that cannot. An MIT study from August 2025 found that 95% of generative artificial intelligence (GenAI) pilots at companies delivered zero measurable return on investment, while Gartner projects that by 2028, 50% of large enterprises will use AI-based supply chain optimisation as their primary planning method, up from 12% in 2024. BCG identifies the differentiator as how planners apply capabilities in evaluating tradeoffs, taking actions, and responding under pressure, not the technology itself.

Why it matters: The 95% zero-ROI stat for GenAI pilots sits in stark contrast with the documented impact of classical OR: Edelman Award finalists consistently show 3-7% operational improvement at enterprise scale, and the 2026 finalists (announced yesterday) span supply chain replenishment, GPU-accelerated scheduling, and national food distribution systems. The signal for practitioners: solver-backed decision systems that connect optimised outputs to measurable operational outcomes are the category that produces ROI. LLM recommendation engines without a solver layer are not. The distinction is increasingly auditable by CFOs.

→ BCG: Supply Chain Planning 2026    → SDCE: Supply Chain Teams Take Back Control
Financial Portfolio · GPU Optimisation · Solver Infrastructure NVIDIA · 2026

NVIDIA cuOpt Delivers 160x Speedup for Enterprise Portfolio Optimisation: Mean-CVaR at 10,000-Variable Scale in Real Time

NVIDIA's cuOpt linear programming (LP) solver with GPU acceleration achieves up to 160x speedups over CPU-based solutions for mean-Conditional Value at Risk (CVaR) portfolio optimisation at scales involving tens of thousands of variables and constraints. The architecture combines GPU-parallelised LP solving with scenario-based Pareto OptimalityA solution is Pareto optimal when no objective can be improved without worsening at least one other. The Pareto front is the complete set of all Pareto-optimal solutions for a multi-objective problem. — first explained April 7 2026. analysis, enabling portfolio managers to model risk-return tradeoffs across large asset universes in real time. Previously, mean-CVaR optimisation at enterprise scale required overnight batch processing; GPU acceleration moves this into live trading windows.

Why it matters: The 160x GPU speedup for portfolio LP completes a pattern visible across this week's signals: the threshold for operationally useful optimisation is always a decision-window problem. Yesterday's energy paper achieved a 100x speedup through mathematical reformulation alone. NVIDIA's cuOpt reaches comparable speedup through hardware. Both paths converge on the same result: formerly-batch optimisation becomes a real-time operational constraint. For quantitative finance teams, mean-CVaR at 10,000+ variables transitioning from overnight batch to live constraint is a structural change in how risk limits are enforced intraday.

→ NVIDIA Technical Blog: cuOpt Portfolio Optimisation
Healthcare · Workforce Scheduling · OR in Production Multiple sources · 2026

Healthcare AI Scheduling Benchmarks: 28% Reduction in Nurse Admin Tasks, 12-18% Operational Cost Savings at Scale

Across multiple 2026 healthcare deployments, artificial intelligence (AI)-driven nurse scheduling and workforce optimisation systems are delivering consistent measurable results: 12-18% operational efficiency gains within the first year of deployment, with specific benchmarks showing a 28% reduction in nurse administrative tasks. These systems use real-time demand forecasting, intelligent shift assignment, and constraint-aware scheduling that matches nurses to shifts based on certifications, rest requirements, patient acuity, and preference data. With 66% of healthcare organisations currently facing understaffing challenges, the scheduling optimisation layer is addressing a structural operational problem rather than a marginal efficiency gain.

Why it matters: Healthcare scheduling is among the most constraint-dense OR application domains in any industry: nurse certifications, legal rest periods, patient acuity levels, shift coverage minimums, union agreements, and multi-department coordination constraints create a combinatorial problem that simple heuristics cannot solve sustainably. The 28%/12% benchmarks are achievable because the underlying solver architecture for these systems is almost always constraint programming (CP) or mixed integer programming (MIP), not greedy algorithms. These results confirm the pattern from the Edelman finalists: when OR is properly engineered into operations, the gains compound rather than plateau.

→ Nature npj Health Systems: AI Reducing Healthcare Costs    → Solvice: Healthcare Workforce Scheduling

Research Papers

Research · Column Generation · Vehicle Routing arXiv:2504.08401 · Apr 2026

Graph Reduction with Unsupervised Learning in Column Generation: A Routing Application

Column Generation (CG) is widely used for large-scale combinatorial optimisation, but its pricing subproblem, which identifies the most promising variable to add to the current solution, is often the computational bottleneck. This paper proposes using unsupervised graph machine learning to reduce the size of the graph on which the pricing subproblem is solved at each CG iteration. By learning which portions of the pricing graph are irrelevant to the current restricted master problem solution, the method reduces pricing cost substantially while preserving solution quality. The routing application demonstrates that graph reduction transfers across instance sizes, providing a generalisation result critical for production deployment.

What problem it solves: The pricing subproblem in CG for vehicle routing is typically a shortest-path problem on a large graph representing all possible routes. Evaluating all edges at every iteration is computationally expensive; this paper learns to prune the graph before pricing, retaining only edges likely to appear in improving columns. The result is faster pricing without sacrificing optimality of the final master problem solution.

Why it matters: This paper lands on the same day as today's Term of the Day explanation of Column Generation, and it illustrates precisely the frontier challenge: even expert-designed CG implementations face pricing subproblem bottlenecks at scale. Machine learning is now being used not to replace CG, but to accelerate the pricing step that CG's correctness depends on. This is the collaboration pattern between ML and OR that is replacing ML-versus-OR framings: each technique handles what it does best, with ML learning problem-specific structure and OR providing the optimality guarantee.

→ View Paper on arXiv
Research · Combinatorial Optimisation · ML+OR Integration arXiv:2601.10583 · Jan 2026

Combinatorial Optimization Augmented Machine Learning: A Survey and Framework

Combinatorial Optimisation Augmented Machine Learning (COAML) is an emerging paradigm that integrates combinatorial optimisation oracles directly into machine learning pipelines, enabling the construction of policies that are simultaneously data-driven and feasibility-preserving. Unlike end-to-end learned approaches that may violate hard constraints, COAML embeds a solver layer within the learning objective, ensuring that outputs are always feasible under the combinatorial constraints of the problem. Applications span scheduling, vehicle routing, stochastic programming, and reinforcement learning, with results demonstrating that COAML policies generalise significantly better than pure ML approaches on out-of-distribution instances.

What problem it solves: Pure ML policies for combinatorial problems routinely violate hard constraints (capacity limits, time windows, regulatory requirements) because the loss function does not penalise constraint violations strongly enough during training. COAML resolves this by differentiating through the solver layer, making constraint satisfaction a structural property of the learning objective rather than a soft penalty term.

Why it matters: COAML provides the theoretical framework for the class of hybrid ML-OR systems that are now appearing across supply chain, healthcare, and transport: systems where ML handles prediction and adaptation under distribution shift, while OR handles constraint-enforcing optimisation. This survey gives practitioners a vocabulary and taxonomy for evaluating and designing such systems, at a moment when the market is producing many tools claiming to combine AI and optimisation without a principled architecture.

→ View Paper on arXiv
Research · Machine Learning · Scheduling · Survey arXiv:2512.22642 · Updated Feb 2026

Machine Learning for Scheduling: A Paradigm Shift from Solver-Centric to Data-Centric Approaches

This survey structures the rapidly growing literature on machine learning for scheduling around three recurring leverage points: learning fast surrogates for computations that are repeated across many similar problem instances; learning inner-solver strategies that steer the search process within branch-and-bound or other exact solvers; and learning algorithm-selection rules that route problem instances to the most appropriate solver or heuristic. The paper argues that the paradigm shift is not from solvers to ML, but from solver-centric design (where problem structure drives everything) to data-centric design (where empirical performance on a distribution of instances is the primary objective).

What problem it solves: The scheduling literature has fragmented into solver-centric and ML-centric camps with limited cross-pollination. This survey provides a unified framework that explains when each approach is appropriate, which hybrid architectures are most effective, and how to evaluate claims across the two paradigms on a common basis.

Why it matters: Scheduling is the application domain most directly affected by the question of whether LLM-based tools (like OptiMind) can replace classical solver methods. This survey provides the nuanced answer: for instance types well-represented in training data, data-centric approaches can match or exceed solver-centric approaches on solution speed. For novel constraint structures, distribution-shifted instances, and problems requiring provable optimality guarantees (like the Edelman-winning systems), solver-centric design remains irreplaceable. The practical implication: data-centric scheduling tools are appropriate for repetitive, well-studied problem types; solver-centric tools are required wherever novelty or optimality guarantees matter.

→ View Paper on arXiv

Term of the Day

Solution Methods · New Term First in this issue · 9 Apr 2026

Column Generation

Column Generation (CG) is an algorithmic technique for solving large-scale linear programs (LPs) and mixed integer programs (MIPs) that would be computationally intractable if all variables (columns) were included from the start. The core idea: instead of solving the full problem with all possible variables, CG works with a small subset of variables (the Restricted Master Problem, or RMP) and iteratively asks a pricing subproblem whether any excluded variable could improve the current solution. If the pricing subproblem identifies such a variable, that column is added to the RMP and the process repeats. Optimality is proven when the pricing subproblem confirms that no improving column exists, at which point the RMP solution is also optimal for the full problem.

Column Generation is the workhorse behind the largest and most practically significant OR deployments in existence. Airline crew scheduling (assigning crew members to all flights in a network) involves millions of possible crew pairings; CG prices only the pairings with potential value. Hospital nurse rostering with full-week schedule patterns, cutting stock problems for paper mills and steel plants, vehicle routing with structured routes, and portfolio optimisation with factor models all rely on CG for tractability at scale. Today's arXiv:2504.08401 uses machine learning to accelerate the pricing subproblem in routing, illustrating that CG is not a legacy technique but an active area of research improvement.

Why practitioners confuse this with other decomposition methods

Confusion 1: CG versus Benders Decomposition. Benders Decomposition splits a problem into a master problem (over integer variables) and a subproblem (over continuous variables), adding Benders cuts (rows) to the master problem from the subproblem's dual. Column Generation splits a problem into a master problem (over a subset of variables) and a pricing subproblem (that finds new variables), adding columns to the master problem. Both decompose; the direction differs. Benders adds rows; Column Generation adds columns. Confusing the two leads to misapplication: Benders works best when the master problem is over integers and the subproblem is a continuous LP; CG works best when the number of variables is large but the pricing problem is efficiently solvable.

Confusion 2: The pricing subproblem is a detail, not an afterthought. The most common practitioner error is treating the pricing subproblem as a mechanical component. In practice, the pricing subproblem is where the domain knowledge lives: it encodes the combinatorial structure of what constitutes a valid column (a legal crew pairing, a feasible route, a valid cutting pattern). Designing an efficient pricing subproblem often requires deeper OR expertise than formulating the master problem. LLMs like OptiMind can draft the master problem formulation; they cannot yet reliably design the pricing subproblem for novel problem structures.

Confusion 3: CG is only for "exponentially many variables." Practitioners assume CG applies only when the number of variables is inherently astronomical (all possible routes, all possible schedules). In fact, CG applies whenever the LP relaxation has a tractable pricing subproblem, even for problems with a moderate number of variables, because starting with a small feasible basis and pricing incrementally can be computationally superior to solving the full LP from scratch when the full LP is large but sparse.

Branch-and-Price: When CG is applied within a branch-and-boundA general algorithmic framework for solving combinatorial optimisation problems exactly, working by recursively partitioning the feasible region and computing bounds to prune branches. Branch and Bound (B&B) is the core engine in all modern MIP solvers. — first explained April 8 2026. tree to solve the integer version of the problem, the technique is called Branch-and-Price. At each node of the branch-and-bound tree, CG is used to solve the LP relaxation, and branching occurs on fractional integer variables. Branch-and-Price is the state-of-the-art method for the largest crew scheduling, vehicle routing, and cutting stock problems solved in production today.

Related:

Mixed Integer Programming · Benders Decomposition · Lagrangian Relaxation · Branch and Bound
→ arXiv:2504.08401 — Column Generation with graph reduction (today's paper)    → Reference: Book on Column Generation (Optimization Online)

Upcoming Conferences

Conference Dates Location Key Track / Deadline
INFORMS Analytics+ 2026 Apr 12–14, 2026 National Harbor, MD Edelman Award Gala (Apr 13), Mixed Integer Programming (MIP) Panel, Supply Chain, GenAI
MIP 2026 Workshop May 18–21, 2026 Stamford, CT (UConn) Flash talk deadline: Apr 15. Land-Doig MIP Competition. Single-track, invitation-only speakers.
IPCO 2026 Jun 17–19, 2026 Padova, Italy 27th Conference on Integer Programming and Combinatorial Optimisation. Poster deadline extended to Apr 6.
VeRoLog 2026 Jul 6–9, 2026 Bath, UK 10th EURO Working Group on Vehicle Routing and Logistics Optimisation. Covers routing, Column Generation, metaheuristics.
Daily Synthesis

The simultaneous arrival of LLM-based formulation tools and a resurgent focus on measurable OR ROI is producing a clarifying moment: LLMs lower the barrier to enter MIP formulation, while Column Generation and other structural decompositions remain the barrier to exit commodity OR. The organisations winning the Edelman Award this week, and the ones delivering 28% admin reductions in healthcare, all built systems that required architectural decomposition decisions no LLM made for them.

  • Microsoft OptiMind democratises formulation; it does not democratise decomposition. The pricing subproblem in Column Generation, the Benders cut generation in large two-stage programs, and the Lagrangian multiplier update rule in capacity-constrained routing all require problem-specific structural insight. This is the knowledge that experienced OR engineers hold and that LLMs do not yet reliably replicate.
  • The 95% zero-ROI stat for GenAI pilots and the NVIDIA cuOpt 160x GPU speedup for portfolio optimisation are two sides of the same problem: GenAI without a solver layer generates recommendations; solver-backed systems generate decisions. The ROI accountability gap in 2026 is fundamentally a gap between recommendation-generating AI and decision-executing AI. Column Generation systems belong firmly in the second category.
  • Today's arXiv:2504.08401, using unsupervised graph learning to accelerate the Column Generation pricing step, is the correct architecture for the ML-OR hybrid era: ML learns structure, OR guarantees feasibility. The Combinatorial Optimisation Augmented Machine Learning (COAML) framework in arXiv:2601.10583 formalises this pattern. Practitioners building hybrid systems should use this framework to evaluate whether their ML layer is accelerating OR or bypassing it.
  • With INFORMS Analytics+ opening Saturday (April 12) and the MIP 2026 Workshop flash-talk deadline on Tuesday (April 15), the next ten days are the highest-signal stretch of Q2 2026 for the OR and decision intelligence community. The Edelman winner announced Monday will set the benchmark for what applied OR impact looks like at scale in 2026.

For practitioners: If you are evaluating whether to apply Column Generation to a problem, ask two diagnostic questions: (1) Can you enumerate, in principle, all possible values of the decision variable, even if the number is exponential? (2) Can you efficiently find the most promising unexplored value using a shortest-path, knapsack, or similar subproblem? If both answers are yes, Column Generation almost certainly applies and is likely the difference between a model that scales and one that does not. The MIP 2026 Workshop (May 18-21, Stamford) includes the Land-Doig MIP Competition, an ideal venue to benchmark CG implementations against state-of-the-art approaches.

Decision Optimisation Radar · nexmindai.org

← Back to Radar Archive