Solving Mixed Integer Programs Using Neural Networks. Nair, V., Bartunov, S., Gimeno, F., von Glehn, I., Lichocki, P., Lobov, I., O'Donoghue, B., Sonnerat, N., Tjandraatmadja, C., Wang, P., Addanki, R., Hapuarachchi, T., Keck, T., Keeling, J., Kohli, P., Ktena, I., Li, Y., Vinyals, O., & Zwols, Y. 2020. Paper Website abstract bibtex Mixed Integer Programming (MIP) solvers rely on an array of sophisticated heuristics developed with decades of research to solve large-scale MIP instances encountered in practice. Machine learning offers to automatically construct better heuristics from data by exploiting shared structure among instances in the data. This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one. Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP. Neural Diving learns a deep neural network to generate multiple partial assignments for its integer variables, and the resulting smaller MIPs for un-assigned variables are solved with SCIP to construct high quality joint assignments. Neural Branching learns a deep neural network to make variable selection decisions in branch-and-bound to bound the objective value gap with a small tree. This is done by imitating a new variant of Full Strong Branching we propose that scales to large instances using GPUs. We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each. Most instances in all the datasets combined have $10^3-10^6$ variables and constraints after presolve, which is significantly larger than previous learning approaches. Comparing solvers with respect to primal-dual gap averaged over a held-out set of instances, the learning-augmented SCIP is 2x to 10x better on all datasets except one on which it is $10^5$x better, at large time limits. To the best of our knowledge, ours is the first learning approach to demonstrate such large improvements over SCIP on both large-scale real-world application datasets and MIPLIB.
@article{
title = {Solving Mixed Integer Programs Using Neural Networks},
type = {article},
year = {2020},
keywords = {deep learning,discrete optimization,first version december 2020,graph networks,history,mixed integer programming},
pages = {1-57},
websites = {http://arxiv.org/abs/2012.13349},
id = {4363b46f-6a12-37f4-b0bd-5ca25d3e34b0},
created = {2021-07-12T10:19:36.836Z},
file_attached = {true},
profile_id = {ad172e55-c0e8-3aa4-8465-09fac4d5f5c8},
group_id = {1ff583c0-be37-34fa-9c04-73c69437d354},
last_modified = {2021-07-12T10:19:45.945Z},
read = {false},
starred = {false},
authored = {false},
confirmed = {true},
hidden = {false},
folder_uuids = {20ccb950-fef9-4ee1-800c-a60ba9f1df16},
private_publication = {false},
abstract = {Mixed Integer Programming (MIP) solvers rely on an array of sophisticated heuristics developed with decades of research to solve large-scale MIP instances encountered in practice. Machine learning offers to automatically construct better heuristics from data by exploiting shared structure among instances in the data. This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one. Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP. Neural Diving learns a deep neural network to generate multiple partial assignments for its integer variables, and the resulting smaller MIPs for un-assigned variables are solved with SCIP to construct high quality joint assignments. Neural Branching learns a deep neural network to make variable selection decisions in branch-and-bound to bound the objective value gap with a small tree. This is done by imitating a new variant of Full Strong Branching we propose that scales to large instances using GPUs. We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each. Most instances in all the datasets combined have $10^3-10^6$ variables and constraints after presolve, which is significantly larger than previous learning approaches. Comparing solvers with respect to primal-dual gap averaged over a held-out set of instances, the learning-augmented SCIP is 2x to 10x better on all datasets except one on which it is $10^5$x better, at large time limits. To the best of our knowledge, ours is the first learning approach to demonstrate such large improvements over SCIP on both large-scale real-world application datasets and MIPLIB.},
bibtype = {article},
author = {Nair, Vinod and Bartunov, Sergey and Gimeno, Felix and von Glehn, Ingrid and Lichocki, Pawel and Lobov, Ivan and O'Donoghue, Brendan and Sonnerat, Nicolas and Tjandraatmadja, Christian and Wang, Pengming and Addanki, Ravichandra and Hapuarachchi, Tharindi and Keck, Thomas and Keeling, James and Kohli, Pushmeet and Ktena, Ira and Li, Yujia and Vinyals, Oriol and Zwols, Yori}
}
Downloads: 0
{"_id":"6jWy5a37Mn4nLdMQG","bibbaseid":"nair-bartunov-gimeno-vonglehn-lichocki-lobov-odonoghue-sonnerat-etal-solvingmixedintegerprogramsusingneuralnetworks-2020","author_short":["Nair, V.","Bartunov, S.","Gimeno, F.","von Glehn, I.","Lichocki, P.","Lobov, I.","O'Donoghue, B.","Sonnerat, N.","Tjandraatmadja, C.","Wang, P.","Addanki, R.","Hapuarachchi, T.","Keck, T.","Keeling, J.","Kohli, P.","Ktena, I.","Li, Y.","Vinyals, O.","Zwols, Y."],"bibdata":{"title":"Solving Mixed Integer Programs Using Neural Networks","type":"article","year":"2020","keywords":"deep learning,discrete optimization,first version december 2020,graph networks,history,mixed integer programming","pages":"1-57","websites":"http://arxiv.org/abs/2012.13349","id":"4363b46f-6a12-37f4-b0bd-5ca25d3e34b0","created":"2021-07-12T10:19:36.836Z","file_attached":"true","profile_id":"ad172e55-c0e8-3aa4-8465-09fac4d5f5c8","group_id":"1ff583c0-be37-34fa-9c04-73c69437d354","last_modified":"2021-07-12T10:19:45.945Z","read":false,"starred":false,"authored":false,"confirmed":"true","hidden":false,"folder_uuids":"20ccb950-fef9-4ee1-800c-a60ba9f1df16","private_publication":false,"abstract":"Mixed Integer Programming (MIP) solvers rely on an array of sophisticated heuristics developed with decades of research to solve large-scale MIP instances encountered in practice. Machine learning offers to automatically construct better heuristics from data by exploiting shared structure among instances in the data. This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one. Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP. Neural Diving learns a deep neural network to generate multiple partial assignments for its integer variables, and the resulting smaller MIPs for un-assigned variables are solved with SCIP to construct high quality joint assignments. Neural Branching learns a deep neural network to make variable selection decisions in branch-and-bound to bound the objective value gap with a small tree. This is done by imitating a new variant of Full Strong Branching we propose that scales to large instances using GPUs. We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each. Most instances in all the datasets combined have $10^3-10^6$ variables and constraints after presolve, which is significantly larger than previous learning approaches. Comparing solvers with respect to primal-dual gap averaged over a held-out set of instances, the learning-augmented SCIP is 2x to 10x better on all datasets except one on which it is $10^5$x better, at large time limits. To the best of our knowledge, ours is the first learning approach to demonstrate such large improvements over SCIP on both large-scale real-world application datasets and MIPLIB.","bibtype":"article","author":"Nair, Vinod and Bartunov, Sergey and Gimeno, Felix and von Glehn, Ingrid and Lichocki, Pawel and Lobov, Ivan and O'Donoghue, Brendan and Sonnerat, Nicolas and Tjandraatmadja, Christian and Wang, Pengming and Addanki, Ravichandra and Hapuarachchi, Tharindi and Keck, Thomas and Keeling, James and Kohli, Pushmeet and Ktena, Ira and Li, Yujia and Vinyals, Oriol and Zwols, Yori","bibtex":"@article{\n title = {Solving Mixed Integer Programs Using Neural Networks},\n type = {article},\n year = {2020},\n keywords = {deep learning,discrete optimization,first version december 2020,graph networks,history,mixed integer programming},\n pages = {1-57},\n websites = {http://arxiv.org/abs/2012.13349},\n id = {4363b46f-6a12-37f4-b0bd-5ca25d3e34b0},\n created = {2021-07-12T10:19:36.836Z},\n file_attached = {true},\n profile_id = {ad172e55-c0e8-3aa4-8465-09fac4d5f5c8},\n group_id = {1ff583c0-be37-34fa-9c04-73c69437d354},\n last_modified = {2021-07-12T10:19:45.945Z},\n read = {false},\n starred = {false},\n authored = {false},\n confirmed = {true},\n hidden = {false},\n folder_uuids = {20ccb950-fef9-4ee1-800c-a60ba9f1df16},\n private_publication = {false},\n abstract = {Mixed Integer Programming (MIP) solvers rely on an array of sophisticated heuristics developed with decades of research to solve large-scale MIP instances encountered in practice. Machine learning offers to automatically construct better heuristics from data by exploiting shared structure among instances in the data. This paper applies learning to the two key sub-tasks of a MIP solver, generating a high-quality joint variable assignment, and bounding the gap in objective value between that assignment and an optimal one. Our approach constructs two corresponding neural network-based components, Neural Diving and Neural Branching, to use in a base MIP solver such as SCIP. Neural Diving learns a deep neural network to generate multiple partial assignments for its integer variables, and the resulting smaller MIPs for un-assigned variables are solved with SCIP to construct high quality joint assignments. Neural Branching learns a deep neural network to make variable selection decisions in branch-and-bound to bound the objective value gap with a small tree. This is done by imitating a new variant of Full Strong Branching we propose that scales to large instances using GPUs. We evaluate our approach on six diverse real-world datasets, including two Google production datasets and MIPLIB, by training separate neural networks on each. Most instances in all the datasets combined have $10^3-10^6$ variables and constraints after presolve, which is significantly larger than previous learning approaches. Comparing solvers with respect to primal-dual gap averaged over a held-out set of instances, the learning-augmented SCIP is 2x to 10x better on all datasets except one on which it is $10^5$x better, at large time limits. To the best of our knowledge, ours is the first learning approach to demonstrate such large improvements over SCIP on both large-scale real-world application datasets and MIPLIB.},\n bibtype = {article},\n author = {Nair, Vinod and Bartunov, Sergey and Gimeno, Felix and von Glehn, Ingrid and Lichocki, Pawel and Lobov, Ivan and O'Donoghue, Brendan and Sonnerat, Nicolas and Tjandraatmadja, Christian and Wang, Pengming and Addanki, Ravichandra and Hapuarachchi, Tharindi and Keck, Thomas and Keeling, James and Kohli, Pushmeet and Ktena, Ira and Li, Yujia and Vinyals, Oriol and Zwols, Yori}\n}","author_short":["Nair, V.","Bartunov, S.","Gimeno, F.","von Glehn, I.","Lichocki, P.","Lobov, I.","O'Donoghue, B.","Sonnerat, N.","Tjandraatmadja, C.","Wang, P.","Addanki, R.","Hapuarachchi, T.","Keck, T.","Keeling, J.","Kohli, P.","Ktena, I.","Li, Y.","Vinyals, O.","Zwols, Y."],"urls":{"Paper":"https://bibbase.org/service/mendeley/bfbbf840-4c42-3914-a463-19024f50b30c/file/7041ee66-a508-33b9-456a-d219f23ae902/201213349.pdf.pdf","Website":"http://arxiv.org/abs/2012.13349"},"biburl":"https://bibbase.org/service/mendeley/bfbbf840-4c42-3914-a463-19024f50b30c","bibbaseid":"nair-bartunov-gimeno-vonglehn-lichocki-lobov-odonoghue-sonnerat-etal-solvingmixedintegerprogramsusingneuralnetworks-2020","role":"author","keyword":["deep learning","discrete optimization","first version december 2020","graph networks","history","mixed integer programming"],"metadata":{"authorlinks":{}}},"bibtype":"article","biburl":"https://bibbase.org/service/mendeley/bfbbf840-4c42-3914-a463-19024f50b30c","dataSources":["C8ZTSgdcqKrDKQsFr","ya2CyA73rpZseyrZ8","2252seNhipfTmjEBQ"],"keywords":["deep learning","discrete optimization","first version december 2020","graph networks","history","mixed integer programming"],"search_terms":["solving","mixed","integer","programs","using","neural","networks","nair","bartunov","gimeno","von glehn","lichocki","lobov","o'donoghue","sonnerat","tjandraatmadja","wang","addanki","hapuarachchi","keck","keeling","kohli","ktena","li","vinyals","zwols"],"title":"Solving Mixed Integer Programs Using Neural Networks","year":2020}