September 19, 2025
Adaptive robust optimization (ARO) extends static robust optimization by allowing decisions to depend on the realized uncertainty — weakly dominating static solutions within the modeled uncertainty set. However, ARO makes previous constraints that were independent of uncertainty now dependent, making it vulnerable to additional infeasibilities when realizations fall outside the uncertainty set. This phenomenon of adaptive policies being brittle is analogous to overfitting in machine learning. To mitigate against this, we propose assigning constraint-specific uncertainty set sizes, with harder constraints given stronger probabilistic guarantees. Interpreted through the overfitting lens, this acts as regularization: tighter guarantees shrink adaptive coefficients to ensure stability, while looser ones preserve useful flexibility. This view motivates a principled approach to designing uncertainty sets that balances robustness and adaptivity.
Robust optimization (RO) is one of the main frameworks (along with stochastic programming) for decision-making under uncertainty. RO produces a robust solution by requiring feasibility across all realizations within a specified uncertainty set \(\mathcal{U}\). In this paper, we consider a robust linear optimization problem with \(m\) robust constraints and uncertainty on the right-hand side (RHS): \[\begin{align} \max_{\boldsymbol{x} \in \mathbb{R}^n} \quad & \boldsymbol{c}^\top \boldsymbol{x} \\ \text{s.t.} \quad& \boldsymbol{a}_i^\top \boldsymbol{x} \le b_i + \boldsymbol{d}_i^\top \boldsymbol{u}, \quad \forall \boldsymbol{u} \in \mathcal{U}, \quad \forall i \in [m], \\ & \boldsymbol{x} \ge \boldsymbol{0}, \end{align}\] where for parameters we have \(\boldsymbol{c}\in\mathbb{R}^n\), and for each constraint \(i\), we have \(\boldsymbol{a}_i\in\mathbb{R}^n, b_i\in\mathbb{R}, \boldsymbol{d}_i \in\mathbb{R}^p\). Finally, we have the perturbation parameter \(\boldsymbol{u} \in\mathbb{R}^p\) which is uncertain but assumed to lie within \(\mathcal{U}\). Here, \(\boldsymbol{x}\) is fixed before \(u\) is realized, yielding a static robust solution.
Adaptive robust optimization (ARO) extends this framework by allowing some decisions to adjust after observing \(\boldsymbol{u}\) [1]. A common tractable restriction is to use affine decision rules \[\label{eq:adr} \boldsymbol{x}(\boldsymbol{u}) = \boldsymbol{z} + V \boldsymbol{u}\tag{1}\] with decision variables \(\boldsymbol{z}\in\mathbb{R}^n\) and \(V\in\mathbb{R}^{n\times p}\). Static solutions correspond to \({V}=\boldsymbol{0}\), so adaptive policies always weakly dominate static ones within the set.
The size and geometry of \(\mathcal{U}\) are chosen to balance feasibility and optimality. Common choices include the box (\(\ell_\infty\) norm) [2], ellipsoid (\(\ell_2\) norm) [3], and budget sets (intersection of \(\ell_\infty\) and \(\ell_1\) norms) [4]. A well-chosen uncertainty set should not contain all possible realizations of \(\boldsymbol{u}\), as this would be overly conservative. Therefore, when evaluating solutions in simulation, one should not draw samples directly from the uncertainty set, but rather from an underlying distribution.
A key, under-discussed drawback of ARO is that constraints that were originally independent of uncertainty become dependent. While static robust solutions often remain feasible outside the modeled set, adaptive policies may fail catastrophically—for example, violating variable non-negativity constraints. In this sense, adaptive policies are brittle: they achieve superior performance within \(\mathcal{U}\) but generalize poorly beyond it. This brittleness is directly analogous to overfitting in machine learning, where models with higher flexibility fit training data well but perform poorly out-of-sample.
This perspective motivates a rethinking of how uncertainty sets are specified in ARO. Not all constraints are equally critical: hard constraints demand stronger guarantees, while softer constraints may tolerate limited violations. Interpreted through the overfitting lens, this leads to a regularization view: tighter guarantees shrink adaptive coefficients to improve stability, while looser guarantees preserve flexibility. Although the equivalence of robustness and regularization has been studied extensively [5], [6], to our knowledge, there have been no similar discussions of this insight for ARO. In this note we illustrate brittleness using a renewable generation toy example, propose constraint-dependent uncertainty set sizes as a principled remedy for this brittleness, and show how robust counterparts (RC) impose regularization on adaptive coefficients.
We illustrate brittleness with a toy model of production planning under renewables: \[\begin{align} \max_{x,s} \quad & x_1 + x_2 + y_1 + y_2 - 100s_1 - 100s_2 \\ \text{s.t.} \quad & x_1 + y_1 - s_1 \leq 2 + u_1, \quad \forall (u_1,u_2)\in\mathcal{U}, \\ & x_2 + y_2 - s_2 \leq 2 + u_2, \quad \forall (u_1,u_2)\in\mathcal{U}, \\ & x_1,x_2,y_1,y_2,s_1,s_2 \geq 0. \end{align}\] Here a forecast of \(2\) units of renewable power is available in each period. Decisions \(x_1,y_1\) allocate this supply in period 1, and \(x_2,y_2\) in period 2. Shortfalls are covered by grid imports \(s_1,s_2\), which are costly and non-renewable. Uncertainty in renewable availability is represented by a budget uncertainty set \[\mathcal{U} = \{(u_1,u_2): \|\boldsymbol{u}\|_\infty \le \rho,\;\|\boldsymbol{u}\|_1 \le \Gamma\}.\] An optimal solution to the nominal problem (where \(\mathcal{U} = \{\boldsymbol{0}\}\)) is \(x_i=y_i=1, \;s_i= 0\) for \(i =1,2\) with an objective value of 4, but any shortfall in renewables forces costly grid usage.
An optimal solution to the static robust problem with \((\rho, \Gamma) = (1,1)\) is \(x_i=1, y_i =s_i =0,\) for \(i=1,2\), with an objective value of 2. Notice that although the \(l_1\) ball that defines \(\mathcal{U}\) couples the uncertainties across constraints sets in \(\mathcal{U}\), the static robust problem in general cannot exploit this coupling [7], giving an overly conservative solution that is equivalent to the uncertainty set being a box \(\mathcal{U} = \{(u_1,u_2) : \|\boldsymbol{u}\|_\infty \le \rho\}.\)
In contrast, ARO is able to exploit this coupling [8]. Suppose we introduce a restricted affine decision rules that made the flows adapt to the current period of uncertainty: \[x_i(u_i) = z^x_i + V^x_{i,i} u_i, \quad y_i(u_i) = z^y_i + V^y_{i,i} u_i, \quad i =1,2.\] An optimal adaptive policy with \((\rho, \Gamma)=(1,1)\) is \(x_i(u_i)=1,\;y_i(u_i)=1+u_i\), for \(i=1,2\) with an objective value of 3 – significantly outperforming the static policy within \(\mathcal{U}\). However, the adaptive policy is brittle: the non-negativity constraints, originally independent of \(u\), now depend on \(u\) and may be violated. For instance, if \(u_1<-1\), then \(y_1(u_1)<0\), which is physically infeasible. Shortages can be met with imports, but negative flows cannot be implemented.
A natural remedy is to assign larger uncertainty radii to hard constraints (e.g., non-negativity) and smaller ones to softer constraints (e.g., grid imports). For example, with \((\rho,\Gamma)=(2,2)\) on the non-negativity constraints only, an optimal adaptive solution is \(x_i=y_i=1+\tfrac12 u_i\), for \(i=1,2\) and maintains an objective value of 3 and remains feasible for all \(u\ge -2\). In our example, only a non-negative RHS of renewable supply makes sense, so \(u_1,u_2\) naturally both have a lower bound of \(-2\). So the enlarged set recovers the guaranteed feasibility previously enjoyed by the static policy while retaining the benefits from ARO and uncertainty coupling.
From the toy example we saw that a uniform uncertainty set \(\mathcal{U}\) is not appropriate in ARO, since many constraints that were previously independent of uncertainty become dependent once adaptivity is introduced. This motivates the use of constraint-specific uncertainty sets \(\mathcal{U}_i\) for each constraint \(i \in [m]\), sized according to its criticality. We distinguish two types:
Hard constraints (e.g. flow non-negativity): must be satisfied in all realizations. These require deterministic guarantees, with uncertainty sets that fully cover the support of \(\boldsymbol{u}\).
Softer constraints (e.g., renewable allocations with non-renewable grid imports): can tolerate limited violations if backup resources exist. For these, probabilistic guarantees are appropriate, where ellipsoidal or budget sets provide explicit confidence levels while preserving adaptive flexibility.
We now provide quantitative prescriptions. First, we recall probabilistic guarantees under Gaussian and distribution-free assumptions from the RO literature [9], suited for softer constraints. Then, we present deterministic robust counterparts for bounded and semi-bounded supports, which are appropriate for hard constraints that must hold almost surely.
For each constraint \(i\in[m]\), we may require \[\label{eq:prob-guarantee} \mathbb{P}\!\left[\boldsymbol{a}_i^\top \boldsymbol{x}(\boldsymbol{u}) \;\le\; b_i + \boldsymbol{d}_i^\top \boldsymbol{u}\right]\;\ge\;1-\epsilon.\tag{2}\] Different uncertainty sets \(\mathcal{U}_i\) provide different sufficient conditions for 2 .
Gaussian case. If \(\boldsymbol{u}\sim\mathcal{N}(\boldsymbol{\mu},\Sigma)\) and \(\boldsymbol{x}(\boldsymbol{u})=\boldsymbol{z}+V\boldsymbol{u}\), then 2 holds iff \[\boldsymbol{a}_i^\top \boldsymbol{z} + (V^\top \boldsymbol{a}_i - \boldsymbol{d}_i)^\top \boldsymbol{\mu} + \rho_{1-\epsilon}\,\bigl\|\Sigma^{1/2}(V^\top \boldsymbol{a}_i - \boldsymbol{d}_i)\bigr\|_2 \;\le\; b_i,\] where \(\rho_{1-\epsilon}\) is the \((1-\epsilon)\)-quantile of the standard normal. Equivalently, this is the RC with ellipsoidal uncertainty set \[\mathcal{U}_i=\left\{\boldsymbol{u}:\; \bigl\|\Sigma^{-1/2}(\boldsymbol{u}-\boldsymbol{\mu})\bigr\|_2 \le \rho_{1-\epsilon}\right\}.\]
Distribution-free case. If \(u_1,\dots,u_p\) are independent, zero-mean, with support \([-1,1]\), then:
Ball-box uncertainty set. For \(\mathcal{U}_i=\{\boldsymbol{u}:\|\boldsymbol{u}\|_2\le \rho,\;\|\boldsymbol{u}\|_\infty\le 1\}\), the RC is \[\label{eq:rc-ball} \boldsymbol{a}_i^\top \boldsymbol{z} + \rho\,\|V^\top \boldsymbol{a}_i - \boldsymbol{d}_i\|_2 \;\le\; b_i,\tag{3}\] which ensures \(\mathbb{P}[\boldsymbol{a}_i^\top \boldsymbol{x}(\boldsymbol{u})\le b_i + \boldsymbol{d}_i^\top \boldsymbol{u}] \ge 1-\exp(-\rho^2/2)\). Choosing \(\rho=\sqrt{2\ln(1/\epsilon)}\) yields 2 .
Budget uncertainty set. For \(\mathcal{U}_i =\{\boldsymbol{u}:\|\boldsymbol{u}\|_1\le \Gamma,\;\|\boldsymbol{u}\|_\infty\le 1\}\), the RC is \[\boldsymbol{a}_i^\top \boldsymbol{z} + \Gamma\,\|V^\top \boldsymbol{a}_i - \boldsymbol{d}_i - \boldsymbol{y}_i\|_\infty + \|\boldsymbol{y}_i\|_1 \;\le\; b_i, \quad \boldsymbol{y}_i\in\mathbb{R}^p,\] which ensures \(\mathbb{P}[\boldsymbol{a}_i^\top \boldsymbol{x}(\boldsymbol{u})\le b_i + \boldsymbol{d}_i^\top \boldsymbol{u}] \ge 1-\exp(-\Gamma^2/(2p))\). Choosing \(\Gamma=\sqrt{2\ln(1/\epsilon)}\sqrt{p}\) yields 2 .
For hard constraints that must hold without exception, probabilistic guarantees are insufficient. In these cases, one must construct an uncertainty set that fully contains the support of \(\boldsymbol{u}\) and enforce feasibility deterministically. This leads to robust counterparts over polyhedral box uncertainty sets, whose derivation follows standard duality arguments (see [10] for a tutorial). We state the results directly below.
Proposition 1 (Bounded support). Suppose \(u_k\) has support \([L_k,U_k] \;\forall k \in [p]\), and \(\boldsymbol{x}(\boldsymbol{u})\) is affinely adaptive (1 ). Then the following RC satisfies \(\boldsymbol{a}_i^\top \boldsymbol{x}(\boldsymbol{u})\le b_i + \boldsymbol{d}_i^\top \boldsymbol{u}\) w.p. 1: \[\boldsymbol{a}_i^\top \boldsymbol{z} + \boldsymbol{U}^\top \boldsymbol{\beta} - \boldsymbol{L}^\top \boldsymbol{\alpha} \;\le b_i, \quad \boldsymbol{\beta} - \boldsymbol{\alpha} = \boldsymbol{V}^\top \boldsymbol{a}_i - \boldsymbol{d}_i, \quad \boldsymbol{\alpha,\beta} \ge \boldsymbol{0},\] where \(\boldsymbol{L} = (L_k)_{k=1}^p, \boldsymbol{U} = (U_k)_{k=1}^p\) and \(\boldsymbol{\alpha}, \boldsymbol{\beta} \in \mathbb{R}^p\).
Corollary 1 (Semi-bounded support). Under the same setup as Proposition 1, but instead if \(u_k\) is unbounded from below (resp. above), for each \(k \in [p]\), the RC from Proposition 1 along with \(\alpha_k = 0 \;(\text{resp. } \beta_k =0 )\) satisfies \(\boldsymbol{a}_i^\top \boldsymbol{x}(\boldsymbol{u})\le b_i + \boldsymbol{d}_i^\top \boldsymbol{u}\) w.p. 1.
We now interpret robust constraints through the lens of regularization, providing a mitigation for the brittleness of ARO in line with the overfitting analogy. The RCs of affine rules typically rearrange into explicit norm bounds on the adaptive coefficients. For example, (3 ) can be rewritten as \[\|V^\top \boldsymbol{a}_i - \boldsymbol{d}_i\|_2 \;\le\; \frac{\boldsymbol{b}_i - \boldsymbol{a}_i^\top \boldsymbol{z}}{\rho}, \qquad \rho > 0.\] Here the probabilistic guarantee \(1-\epsilon\) serves as a regularization parameter: requiring stronger guarantees (smaller \(\epsilon\)) tightens the constraint on \(V\). In the limit \(\epsilon \to 0\), we obtain \(V=\boldsymbol{0}\), recovering the static policy, which avoids the brittleness of adaptive rules. The RC directly parallels ridge regression in machine learning. In both cases, the regularization strength is inversely related to the allowed coefficient magnitude.
For illustration, we set \(\boldsymbol{d}_i=0\) and normalize \(\boldsymbol{b}_i - \boldsymbol{a}_i^\top \boldsymbol{z} = 1\), and plot the resulting upper bound on \(\|V^\top \boldsymbol{a}_i\|_2\) against the probabilistic guarantee \(1-\epsilon\). Figure 1 compares the bounds derived under a Gaussian assumption with those obtained from the distribution-free guarantee.
As the guarantee level \(1-\epsilon\) approaches one, the maximum feasible adaptivity shrinks toward zero. This behavior directly mirrors the bias–variance trade-off in machine learning: greater flexibility yields better in-sample fit but poorer generalization, whereas stronger regularization reduces capacity and stabilizes out-of-sample performance.
Our analysis suggests several lessons for modeling with ARO.
Simulate beyond the set. Evaluate performance on distributions whose support exceeds the modeled uncertainty set, just as machine learning models must be tested out-of-sample.
Differentiate constraints. Treat some as soft, where limited violations are tolerable (e.g., supply–demand balance), and others as hard, where violations are catastrophic (e.g., flow non-negativity).
Use constraint-dependent uncertainty set sizes. Assign larger sets to hard constraints, effectively regularizing their adaptive coefficients more strongly, while smaller sizes suffice for softer ones.
Overall, ARO provides flexibility but can overfit to the specified uncertainty sets, producing brittle out-of-set behavior. Interpreting robust counterparts as implicit regularization motivates constraint-dependent uncertainty sets as a principled way to balance adaptivity and stability.