{"_id":"nuz6bYPD9wCmw4GCm","bibbaseid":"ulabdeen-yin-kekatos-jin-learningneuralnetworksunderinputoutputspecifications-2022","author_short":["ul Abdeen, Z.","Yin, H.","Kekatos, V.","Jin, M."],"bibdata":{"bibtype":"inproceedings","type":"inproceedings","title":"Learning Neural Networks under Input-Output Specifications","author":[{"firstnames":["Zain"],"propositions":["ul"],"lastnames":["Abdeen"],"suffixes":[]},{"firstnames":["He"],"propositions":[],"lastnames":["Yin"],"suffixes":[]},{"firstnames":["Vassilis"],"propositions":[],"lastnames":["Kekatos"],"suffixes":[]},{"firstnames":["Ming"],"propositions":[],"lastnames":["Jin"],"suffixes":[]}],"booktitle":"American Control Conference","pages":"","year":"2022","url_pdf":"LearnNNSpecs.pdf","keywords":"Machine learning, Optimization","abstract":"In this paper, we examine an important problem of learning neural networks that certifiably meet certain specifications on input-output behaviors. Our strategy is to find an inner approximation of the set of admissible policy parameters, which is convex in a transformed space. To this end, we address the key technical challenge of convexifying the verification condition for neural networks, which is derived by abstracting the nonlinear specifications and activation functions with quadratic constraints. In particular, we propose a reparametrization scheme of the original neural network based on loop transformation, which leads to a convex condition that can be computationally enforced during learning. The theoretical construction is validated in an experiment that specifies reachable sets for different regions of inputs. ","bibtex":"@inproceedings{2022_2C_learning,\n title={Learning Neural Networks under Input-Output Specifications},\n author={Zain ul Abdeen and He Yin and Vassilis Kekatos and Ming Jin},\n booktitle={American Control Conference},\n pages={},\n year={2022},\n url_pdf={LearnNNSpecs.pdf},\n keywords = {Optimization, Machine Learning},\n abstract={In this paper, we examine an important problem of learning neural networks that certifiably meet certain specifications on input-output behaviors. Our strategy is to find an inner approximation of the set of admissible policy parameters, which is convex in a transformed space. To this end, we address the key technical challenge of convexifying the verification condition for neural networks, which is derived by abstracting the nonlinear specifications and activation functions with quadratic constraints. In particular, we propose a reparametrization scheme of the original neural network based on loop transformation, which leads to a convex condition that can be computationally enforced during learning. The theoretical construction is validated in an experiment that specifies reachable sets for different regions of inputs. },\n keywords={Machine learning, Optimization}\n}\n\n\n","author_short":["ul Abdeen, Z.","Yin, H.","Kekatos, V.","Jin, M."],"key":"2022_2C_learning","id":"2022_2C_learning","bibbaseid":"ulabdeen-yin-kekatos-jin-learningneuralnetworksunderinputoutputspecifications-2022","role":"author","urls":{" pdf":"http://www.jinming.tech/papers/LearnNNSpecs.pdf"},"keyword":["Machine learning","Optimization"],"metadata":{"authorlinks":{}},"downloads":8},"bibtype":"inproceedings","biburl":"http://www.jinming.tech/papers/myref.bib","dataSources":["sTzDHHaipTZWjp8oe","Y64tp2HnDCfXgLdc5"],"keywords":["machine learning","optimization"],"search_terms":["learning","neural","networks","under","input","output","specifications","ul abdeen","yin","kekatos","jin"],"title":"Learning Neural Networks under Input-Output Specifications","year":2022,"downloads":8}