A neural network-based distributional constraint learning methodology for mixed-integer stochastic optimization
The use of machine-learning methods helps to improve decision-making in different fields. In particular, the idea of bridging predictions (predictive models) and prescriptions (optimization problems) is gaining attention within the scientific community. One of the main ideas to address this trade-of...
        Saved in:
      
    
          | Published in | Expert systems with applications Vol. 232; p. 120895 | 
|---|---|
| Main Authors | , | 
| Format | Journal Article | 
| Language | English | 
| Published | 
            Elsevier Ltd
    
        01.12.2023
     | 
| Subjects | |
| Online Access | Get full text | 
| ISSN | 0957-4174 1873-6793 1873-6793  | 
| DOI | 10.1016/j.eswa.2023.120895 | 
Cover
| Summary: | The use of machine-learning methods helps to improve decision-making in different fields. In particular, the idea of bridging predictions (predictive models) and prescriptions (optimization problems) is gaining attention within the scientific community. One of the main ideas to address this trade-off is the Constraint Learning (CL) methodology, where the structure of the machine learning model can be treated as a set of constraints to be embedded within the optimization problem, establishing the relationship between a direct decision variable x and a response variable y. However, most CL approaches have focused on making point predictions, not considering the statistical and external uncertainty faced in the modeling process. In this paper, we extend the CL methodology to deal with uncertainty in the response variable y. The novel Distributional Constraint Learning (DCL) methodology makes use of a piece-wise linearizable neural network-based model to estimate the parameters of the conditional distribution of y (dependent on decisions x and contextual information), which can be embedded within mixed-integer optimization problems. In particular, we formulate a stochastic optimization problem by sampling random values from the estimated distribution by using a linear set of constraints. In this sense, DCL combines both the predictive performance of the neural network method and the possibility of generating scenarios to account for uncertainty within a tractable optimization model. The behavior of the proposed methodology is tested in the context of electricity systems, where a Virtual Power Plant seeks to optimize its operation, subject to different forms of uncertainty, and with price-responsive consumers.
•Extent CL methodology in MIO addressing uncertainty in response variables.•DCL uses neural nets to estimate and embed conditional distributions in MIO.•Develop open access (DistCL) tool to employ the DCL methodology.•Test the validity of the approach in a Virtual Power Plant problem. | 
|---|---|
| ISSN: | 0957-4174 1873-6793 1873-6793  | 
| DOI: | 10.1016/j.eswa.2023.120895 |