BitNet: Scaling 1-bit Transformers for Large Language Models. Wang, H., Ma, S., Dong, L., Huang, S., Wang, H., Ma, L., Yang, F., Wang, R., Wu, Y., & Wei, F. CoRR, 2023.
BitNet: Scaling 1-bit Transformers for Large Language Models [link]Paper  doi  bibtex   
@article{DBLP:journals/corr/abs-2310-11453,
  author       = {Hongyu Wang and
                  Shuming Ma and
                  Li Dong and
                  Shaohan Huang and
                  Huaijie Wang and
                  Lingxiao Ma and
                  Fan Yang and
                  Ruiping Wang and
                  Yi Wu and
                  Furu Wei},
  title        = {BitNet: Scaling 1-bit Transformers for Large Language Models},
  journal      = {CoRR},
  volume       = {abs/2310.11453},
  year         = {2023},
  url          = {https://doi.org/10.48550/arXiv.2310.11453},
  doi          = {10.48550/ARXIV.2310.11453},
  eprinttype   = {arXiv},
  eprint       = {2310.11453},
  timestamp    = {Tue, 16 Jul 2024 01:00:00 +0200},
  biburl       = {https://dblp.org/rec/journals/corr/abs-2310-11453.bib},
  bibsource    = {dblp computer science bibliography, https://dblp.org}
}

Downloads: 0