E.T.: re-thinking self-attention for transformer models on GPUs. Chen, S., Huang, S., Pandey, S., Li, B., Gao, G. R., Zheng, L., Ding, C., & Liu, H. In de Supinski, B. R., Hall, M. W., & Gamblin, T., editors, International Conference for High Performance Computing, Networking, Storage and Analysis, SC 2021, St. Louis, Missouri, USA, November 14-19, 2021, pages 25, 2021. ACM.
E.T.: re-thinking self-attention for transformer models on GPUs [link]Paper  doi  bibtex   
@inproceedings{DBLP:conf/sc/ChenHPLG0D021,
  author       = {Shiyang Chen and
                  Shaoyi Huang and
                  Santosh Pandey and
                  Bingbing Li and
                  Guang R. Gao and
                  Long Zheng and
                  Caiwen Ding and
                  Hang Liu},
  editor       = {Bronis R. de Supinski and
                  Mary W. Hall and
                  Todd Gamblin},
  title        = {{E.T.:} re-thinking self-attention for transformer models on GPUs},
  booktitle    = {International Conference for High Performance Computing, Networking,
                  Storage and Analysis, {SC} 2021, St. Louis, Missouri, USA, November
                  14-19, 2021},
  pages        = {25},
  publisher    = {{ACM}},
  year         = {2021},
  url          = {https://doi.org/10.1145/3458817.3476138},
  doi          = {10.1145/3458817.3476138},
  timestamp    = {Mon, 05 Feb 2024 00:00:00 +0100},
  biburl       = {https://dblp.org/rec/conf/sc/ChenHPLG0D021.bib},
  bibsource    = {dblp computer science bibliography, https://dblp.org}
}

Downloads: 0