Rethinking Reinforcement Learning based Logic Synthesis. Wang, C., Chen, C., Li, D., & Wang, B. June, 2022. arXiv:2205.07614 [cs]Paper abstract bibtex Recently, reinforcement learning has been used to address logic synthesis by formulating the operator sequence optimization problem as a Markov decision process. However, through extensive experiments, we find out that the learned policy makes decisions independent from the circuit features (i.e., states) and yields an operator sequence that is permutation invariant to some extent in terms of operators. Based on these findings, we develop a new RL-based method that can automatically recognize critical operators and generate common operator sequences generalizable to unseen circuits. Our algorithm is verified on both the EPFL benchmark, a private dataset and a circuit at industrial scale. Experimental results demonstrate that it achieves a good balance among delay, area and runtime, and is practical for industrial usage.
@misc{wang_rethinking_2022,
title = {Rethinking {Reinforcement} {Learning} based {Logic} {Synthesis}},
url = {http://arxiv.org/abs/2205.07614},
abstract = {Recently, reinforcement learning has been used to address logic synthesis by formulating the operator sequence optimization problem as a Markov decision process. However, through extensive experiments, we find out that the learned policy makes decisions independent from the circuit features (i.e., states) and yields an operator sequence that is permutation invariant to some extent in terms of operators. Based on these findings, we develop a new RL-based method that can automatically recognize critical operators and generate common operator sequences generalizable to unseen circuits. Our algorithm is verified on both the EPFL benchmark, a private dataset and a circuit at industrial scale. Experimental results demonstrate that it achieves a good balance among delay, area and runtime, and is practical for industrial usage.},
urldate = {2022-10-09},
publisher = {arXiv},
author = {Wang, Chao and Chen, Chen and Li, Dong and Wang, Bin},
month = jun,
year = {2022},
note = {arXiv:2205.07614 [cs]},
keywords = {Computer Science - Hardware Architecture, Computer Science - Machine Learning},
}
Downloads: 0
{"_id":"9PrFDqmS22PRw8vhp","bibbaseid":"wang-chen-li-wang-rethinkingreinforcementlearningbasedlogicsynthesis-2022","author_short":["Wang, C.","Chen, C.","Li, D.","Wang, B."],"bibdata":{"bibtype":"misc","type":"misc","title":"Rethinking Reinforcement Learning based Logic Synthesis","url":"http://arxiv.org/abs/2205.07614","abstract":"Recently, reinforcement learning has been used to address logic synthesis by formulating the operator sequence optimization problem as a Markov decision process. However, through extensive experiments, we find out that the learned policy makes decisions independent from the circuit features (i.e., states) and yields an operator sequence that is permutation invariant to some extent in terms of operators. Based on these findings, we develop a new RL-based method that can automatically recognize critical operators and generate common operator sequences generalizable to unseen circuits. Our algorithm is verified on both the EPFL benchmark, a private dataset and a circuit at industrial scale. Experimental results demonstrate that it achieves a good balance among delay, area and runtime, and is practical for industrial usage.","urldate":"2022-10-09","publisher":"arXiv","author":[{"propositions":[],"lastnames":["Wang"],"firstnames":["Chao"],"suffixes":[]},{"propositions":[],"lastnames":["Chen"],"firstnames":["Chen"],"suffixes":[]},{"propositions":[],"lastnames":["Li"],"firstnames":["Dong"],"suffixes":[]},{"propositions":[],"lastnames":["Wang"],"firstnames":["Bin"],"suffixes":[]}],"month":"June","year":"2022","note":"arXiv:2205.07614 [cs]","keywords":"Computer Science - Hardware Architecture, Computer Science - Machine Learning","bibtex":"@misc{wang_rethinking_2022,\n\ttitle = {Rethinking {Reinforcement} {Learning} based {Logic} {Synthesis}},\n\turl = {http://arxiv.org/abs/2205.07614},\n\tabstract = {Recently, reinforcement learning has been used to address logic synthesis by formulating the operator sequence optimization problem as a Markov decision process. However, through extensive experiments, we find out that the learned policy makes decisions independent from the circuit features (i.e., states) and yields an operator sequence that is permutation invariant to some extent in terms of operators. Based on these findings, we develop a new RL-based method that can automatically recognize critical operators and generate common operator sequences generalizable to unseen circuits. Our algorithm is verified on both the EPFL benchmark, a private dataset and a circuit at industrial scale. Experimental results demonstrate that it achieves a good balance among delay, area and runtime, and is practical for industrial usage.},\n\turldate = {2022-10-09},\n\tpublisher = {arXiv},\n\tauthor = {Wang, Chao and Chen, Chen and Li, Dong and Wang, Bin},\n\tmonth = jun,\n\tyear = {2022},\n\tnote = {arXiv:2205.07614 [cs]},\n\tkeywords = {Computer Science - Hardware Architecture, Computer Science - Machine Learning},\n}\n\n","author_short":["Wang, C.","Chen, C.","Li, D.","Wang, B."],"key":"wang_rethinking_2022","id":"wang_rethinking_2022","bibbaseid":"wang-chen-li-wang-rethinkingreinforcementlearningbasedlogicsynthesis-2022","role":"author","urls":{"Paper":"http://arxiv.org/abs/2205.07614"},"keyword":["Computer Science - Hardware Architecture","Computer Science - Machine Learning"],"metadata":{"authorlinks":{}},"downloads":0,"html":""},"bibtype":"misc","biburl":"https://bibbase.org/zotero/bxt101","dataSources":["Wsv2bQ4jPuc7qme8R"],"keywords":["computer science - hardware architecture","computer science - machine learning"],"search_terms":["rethinking","reinforcement","learning","based","logic","synthesis","wang","chen","li","wang"],"title":"Rethinking Reinforcement Learning based Logic Synthesis","year":2022}