Learning in Combinatorial Optimization: What and How to Explore
Author
dc.contributor.author
Modaresi, Sajad
Author
dc.contributor.author
Sauré Valenzuela, Denis
Author
dc.contributor.author
Vielma, Juan Pablo
Admission date
dc.date.accessioned
2021-03-22T21:07:34Z
Available date
dc.date.available
2021-03-22T21:07:34Z
Publication date
dc.date.issued
2020
Cita de ítem
dc.identifier.citation
Operations Research Volumen: 68 Número: 5 Páginas: 1585-1604 Sep-Oct 2020
es_ES
Identifier
dc.identifier.other
10.1287/opre.2019.1926
Identifier
dc.identifier.uri
https://repositorio.uchile.cl/handle/2250/178743
Abstract
dc.description.abstract
We study dynamic decision making under uncertainty when, at each period, a decision maker implements a solution to a combinatorial optimization problem. The objective coefficient vectors of said problem, which are unobserved before implementation, vary from period to period. These vectors, however, are known to be random draws from an initially unknown distribution with known range. By implementing different solutions, the decision maker extracts information about the underlying distribution but at the same time experiences the cost associated with said solutions. We show that resolving the implied exploration versus exploitation tradeoff efficiently is related to solving a lower-bound problem (LBP), which simultaneously answers the questions of what to explore and how to do so. We establish a fundamental limit on the asymptotic performance of any admissible policy that is proportional to the optimal objective value of the LBP problem. We show that such a lower bound might be asymptotically attained by policies that adaptively reconstruct and solve the LBP at an exponentially decreasing frequency. Because the LBP is likely intractable in practice, we propose policies that instead reconstruct and solve a proxy for the LBP, which we call the optimality cover problem (OCP). We provide strong evidence of the practical tractability of the OCP, which implies that the proposed policies can be implemented in real time. We test the performance of the proposed policies through extensive numerical experiments, and we show that they significantly outperform relevant benchmarks in the long-term and are competitive in the short-term.
es_ES
Patrocinador
dc.description.sponsorship
National Science Foundation (NSF)
NSF - Directorate for Engineering (ENG)
1233441
Complex Engineering Systems Institute
CONICYT: PIA FB0816