Optimal decision diagrams for classification. Florio, A., Martins, P., Schiffer, M., Serra, T., & Vidal, T. In Williams, B., Chen, Y., & Neville, J., editors, AAAI '23, pages 7577-7585, 2023. AAAI Press.
Website abstract bibtex Decision diagrams for classification have some notable advantages over decision trees, as their internal connections can be determined at training time and their width is not bound to grow exponentially with their depth. Accordingly, decision diagrams are usually less prone to data fragmentation in internal nodes. However, the inherent complexity of training these classifiers acted as a long-standing barrier to their widespread adoption. In this context, we study the training of optimal decision diagrams (ODDs) from a mathematical programming perspective. We introduce a novel mixed-integer linear programming model for training and demonstrate its applicability for many datasets of practical importance. Further, we show how this model can be easily extended for fairness, parsimony, and stability notions. We present numerical analyses showing that our model allows training ODDs in short computational times, and that ODDs achieve better accuracy than optimal decision trees, while allowing for improved stability without significant accuracy losses.
@inproceedings{
title = {Optimal decision diagrams for classification},
type = {inproceedings},
year = {2023},
pages = {7577-7585},
websites = {https://arxiv.org/pdf/2205.14500.pdf},
publisher = {AAAI Press},
city = {Washington, D.C.},
institution = {arXiv:2205.14500},
id = {035ed1bc-c913-38a8-9a9c-e1a2a4022c5f},
created = {2022-05-31T15:40:32.109Z},
file_attached = {true},
profile_id = {5e3d1dc4-cb58-3af5-aff1-4d943d2eaf6a},
last_modified = {2025-02-16T14:45:05.185Z},
read = {true},
starred = {false},
authored = {true},
confirmed = {true},
hidden = {false},
citation_key = {Florio2023a},
private_publication = {false},
abstract = {Decision diagrams for classification have some notable advantages over decision trees, as their internal connections can be determined at training time and their width is not bound to grow exponentially with their depth. Accordingly, decision diagrams are usually less prone to data fragmentation in internal nodes. However, the inherent complexity of training these classifiers acted as a long-standing barrier to their widespread adoption. In this context, we study the training of optimal decision diagrams (ODDs) from a mathematical programming perspective. We introduce a novel mixed-integer linear programming model for training and demonstrate its applicability for many datasets of practical importance. Further, we show how this model can be easily extended for fairness, parsimony, and stability notions. We present numerical analyses showing that our model allows training ODDs in short computational times, and that ODDs achieve better accuracy than optimal decision trees, while allowing for improved stability without significant accuracy losses.},
bibtype = {inproceedings},
author = {Florio, A.M. and Martins, P. and Schiffer, M. and Serra, T. and Vidal, T.},
editor = {Williams, B. and Chen, Y. and Neville, J.},
booktitle = {AAAI '23}
}
Downloads: 0
{"_id":"5Krbqk9E6st7bLrxy","bibbaseid":"florio-martins-schiffer-serra-vidal-optimaldecisiondiagramsforclassification-2023","author_short":["Florio, A.","Martins, P.","Schiffer, M.","Serra, T.","Vidal, T."],"bibdata":{"title":"Optimal decision diagrams for classification","type":"inproceedings","year":"2023","pages":"7577-7585","websites":"https://arxiv.org/pdf/2205.14500.pdf","publisher":"AAAI Press","city":"Washington, D.C.","institution":"arXiv:2205.14500","id":"035ed1bc-c913-38a8-9a9c-e1a2a4022c5f","created":"2022-05-31T15:40:32.109Z","file_attached":"true","profile_id":"5e3d1dc4-cb58-3af5-aff1-4d943d2eaf6a","last_modified":"2025-02-16T14:45:05.185Z","read":"true","starred":false,"authored":"true","confirmed":"true","hidden":false,"citation_key":"Florio2023a","private_publication":false,"abstract":"Decision diagrams for classification have some notable advantages over decision trees, as their internal connections can be determined at training time and their width is not bound to grow exponentially with their depth. Accordingly, decision diagrams are usually less prone to data fragmentation in internal nodes. However, the inherent complexity of training these classifiers acted as a long-standing barrier to their widespread adoption. In this context, we study the training of optimal decision diagrams (ODDs) from a mathematical programming perspective. We introduce a novel mixed-integer linear programming model for training and demonstrate its applicability for many datasets of practical importance. Further, we show how this model can be easily extended for fairness, parsimony, and stability notions. We present numerical analyses showing that our model allows training ODDs in short computational times, and that ODDs achieve better accuracy than optimal decision trees, while allowing for improved stability without significant accuracy losses.","bibtype":"inproceedings","author":"Florio, A.M. and Martins, P. and Schiffer, M. and Serra, T. and Vidal, T.","editor":"Williams, B. and Chen, Y. and Neville, J.","booktitle":"AAAI '23","bibtex":"@inproceedings{\n title = {Optimal decision diagrams for classification},\n type = {inproceedings},\n year = {2023},\n pages = {7577-7585},\n websites = {https://arxiv.org/pdf/2205.14500.pdf},\n publisher = {AAAI Press},\n city = {Washington, D.C.},\n institution = {arXiv:2205.14500},\n id = {035ed1bc-c913-38a8-9a9c-e1a2a4022c5f},\n created = {2022-05-31T15:40:32.109Z},\n file_attached = {true},\n profile_id = {5e3d1dc4-cb58-3af5-aff1-4d943d2eaf6a},\n last_modified = {2025-02-16T14:45:05.185Z},\n read = {true},\n starred = {false},\n authored = {true},\n confirmed = {true},\n hidden = {false},\n citation_key = {Florio2023a},\n private_publication = {false},\n abstract = {Decision diagrams for classification have some notable advantages over decision trees, as their internal connections can be determined at training time and their width is not bound to grow exponentially with their depth. Accordingly, decision diagrams are usually less prone to data fragmentation in internal nodes. However, the inherent complexity of training these classifiers acted as a long-standing barrier to their widespread adoption. In this context, we study the training of optimal decision diagrams (ODDs) from a mathematical programming perspective. We introduce a novel mixed-integer linear programming model for training and demonstrate its applicability for many datasets of practical importance. Further, we show how this model can be easily extended for fairness, parsimony, and stability notions. We present numerical analyses showing that our model allows training ODDs in short computational times, and that ODDs achieve better accuracy than optimal decision trees, while allowing for improved stability without significant accuracy losses.},\n bibtype = {inproceedings},\n author = {Florio, A.M. and Martins, P. and Schiffer, M. and Serra, T. and Vidal, T.},\n editor = {Williams, B. and Chen, Y. and Neville, J.},\n booktitle = {AAAI '23}\n}","author_short":["Florio, A.","Martins, P.","Schiffer, M.","Serra, T.","Vidal, T."],"editor_short":["Williams, B.","Chen, Y.","Neville, J."],"urls":{"Website":"https://arxiv.org/pdf/2205.14500.pdf"},"biburl":"https://bibbase.org/service/mendeley/1465671","bibbaseid":"florio-martins-schiffer-serra-vidal-optimaldecisiondiagramsforclassification-2023","role":"author","metadata":{"authorlinks":{}}},"bibtype":"inproceedings","biburl":"https://bibbase.org/service/mendeley/1465671","dataSources":["yinfondEAJRbDM9sJ","sempRA6PhmAdGk3yG","2252seNhipfTmjEBQ"],"keywords":[],"search_terms":["optimal","decision","diagrams","classification","florio","martins","schiffer","serra","vidal"],"title":"Optimal decision diagrams for classification","year":2023}