Learning to Optimize at Scale: A Benders Decomposition-TransfORmers Framework for Stochastic Combinatorial Optimization
Abstract
We introduce the first integration of transformer-based learning within a dynamic decomposition framework to solve large-scale two-stage stochastic mixed-integer programs. We focus on the stochastic capacitated lot-sizing problem under demand uncertainty. Our approach reformulates the problem to incorporate stochasticity and partitions the problem into smaller, tractable subproblems using a dynamic decomposition. These subproblems are solved by a transfORmer model, providing a novel combination of learned combinatorial and relaxation-based cuts to the master problem. This hybrid learning-optimization strategy bridges deep learning with exact mathematical programming, achieving fast but high-quality solutions. For the test set considered, our method outperforms traditional decomposition and direct solvers in both runtime and scalability, demonstrating the potential of transformers as surrogate optimizers embedded within structured solution algorithms for stochastic combinatorial problems.