Browsing by Author "Bayraksan, Guezin"
Now showing 1 - 2 of 2
Results Per Page
Sort Options
- Data-Driven Sample Average Approximation with Covariate InformationKannan, Rohit; Bayraksan, Guezin; Luedtke, James R. (INFORMS, 2025-01-06)We study optimization for data-driven decision-making when we have observations of the uncertain parameters within an optimization model together with concurrent observations of covariates. The goal is to choose a decision that minimizes the expected cost conditioned on a new covariate observation. We investigate two data-driven frameworks that integrate a machine learning prediction model within a stochastic programming sample average approximation (SAA) for approximating the solution to this problem. One SAA framework is new and uses leave-one-out residuals for scenario generation. The frameworks we investigate are flexible and accommodate parametric, nonparametric, and semiparametric regression techniques. We derive conditions on the data generation process, the prediction model, and the stochastic program under which solutions of these data-driven SAAs are consistent and asymptotically optimal, and also derive finite sample guarantees. Computational experiments validate our theoretical results, demonstrate examples where our datadriven formulations have advantages over existing approaches (even if the prediction model is misspecified), and illustrate the benefits of our data-driven formulations in the limited data regime.
- Residuals-based distributionally robust optimization with covariate informationKannan, Rohit; Bayraksan, Guezin; Luedtke, James R. (Springer, 2023-09-26)We consider data-driven approaches that integrate a machine learning prediction model within distributionally robust optimization (DRO) given limited joint observations of uncertain parameters and covariates. Our framework is flexible in the sense that it can accommodate a variety of regression setups and DRO ambiguity sets. We investigate asymptotic and finite sample properties of solutions obtained using Wasserstein, sample robust optimization, and phi-divergence-based ambiguity sets within our DRO formulations, and explore cross-validation approaches for sizing these ambiguity sets. Through numerical experiments, we validate our theoretical results, study the effectiveness of our approaches for sizing ambiguity sets, and illustrate the benefits of our DRO formulations in the limited data regime even when the prediction model is misspecified.