PaLM 2 Technical Report. Anil, R., Dai, A. M., Firat, O., Johnson, M., Lepikhin, D., Passos, A., Shakeri, S., Taropa, E., Bailey, P., Chen, Z., Chu, E., Clark, J. H., Shafey, L. E., Huang, Y., Meier-Hellstern, K., Mishra, G., Moreira, E., Omernick, M., Robinson, K., Ruder, S., Tay, Y., Xiao, K., Xu, Y., Zhang, Y., Abrego, G. H., Ahn, J., Austin, J., Barham, P., Botha, J., Bradbury, J., Brahma, S., Brooks, K., Catasta, M., Cheng, Y., Cherry, C., Choquette-Choo, C. A., Chowdhery, A., Crepy, C., Dave, S., Dehghani, M., Dev, S., Devlin, J., Díaz, M., Du, N., Dyer, E., Feinberg, V., Feng, F., Fienber, V., Freitag, M., Garcia, X., Gehrmann, S., Gonzalez, L., Gur-Ari, G., Hand, S., Hashemi, H., Hou, L., Howland, J., Hu, A., Hui, J., Hurwitz, J., Isard, M., Ittycheriah, A., Jagielski, M., Jia, W., Kenealy, K., Krikun, M., Kudugunta, S., Lan, C., Lee, K., Lee, B., Li, E., Li, M., Li, W., Li, Y., Li, J., Lim, H., Lin, H., Liu, Z., Liu, F., Maggioni, M., Mahendru, A., Maynez, J., Misra, V., Moussalem, M., Nado, Z., Nham, J., Ni, E., Nystrom, A., Parrish, A., Pellat, M., Polacek, M., Polozov, A., Pope, R., Qiao, S., Reif, E., Richter, B., Riley, P., Ros, A. C., Roy, A., Saeta, B., Samuel, R., Shelby, R., Slone, A., Smilkov, D., So, D. R., Sohn, D., Tokumine, S., Valter, D., Vasudevan, V., Vodrahalli, K., Wang, X., Wang, P., Wang, Z., Wang, T., Wieting, J., Wu, Y., Xu, K., Xu, Y., Xue, L., Yin, P., Yu, J., Zhang, Q., Zheng, S., Zheng, C., Zhou, W., Zhou, D., Petrov, S., & Wu, Y. September, 2023. arXiv:2305.10403 [cs]
Paper abstract bibtex We introduce PaLM 2, a new state-of-the-art language model that has better multilingual and reasoning capabilities and is more compute-efficient than its predecessor PaLM. PaLM 2 is a Transformer-based model trained using a mixture of objectives. Through extensive evaluations on English and multilingual language, and reasoning tasks, we demonstrate that PaLM 2 has significantly improved quality on downstream tasks across different model sizes, while simultaneously exhibiting faster and more efficient inference compared to PaLM. This improved efficiency enables broader deployment while also allowing the model to respond faster, for a more natural pace of interaction. PaLM 2 demonstrates robust reasoning capabilities exemplified by large improvements over PaLM on BIG-Bench and other reasoning tasks. PaLM 2 exhibits stable performance on a suite of responsible AI evaluations, and enables inference-time control over toxicity without additional overhead or impact on other capabilities. Overall, PaLM 2 achieves state-of-the-art performance across a diverse set of tasks and capabilities. When discussing the PaLM 2 family, it is important to distinguish between pre-trained models (of various sizes), fine-tuned variants of these models, and the user-facing products that use these models. In particular, user-facing products typically include additional pre- and post-processing steps. Additionally, the underlying models may evolve over time. Therefore, one should not expect the performance of user-facing products to exactly match the results reported in this report.
@misc{anil_palm_2023,
title = {{PaLM} 2 {Technical} {Report}},
url = {http://arxiv.org/abs/2305.10403},
abstract = {We introduce PaLM 2, a new state-of-the-art language model that has better multilingual and reasoning capabilities and is more compute-efficient than its predecessor PaLM. PaLM 2 is a Transformer-based model trained using a mixture of objectives. Through extensive evaluations on English and multilingual language, and reasoning tasks, we demonstrate that PaLM 2 has significantly improved quality on downstream tasks across different model sizes, while simultaneously exhibiting faster and more efficient inference compared to PaLM. This improved efficiency enables broader deployment while also allowing the model to respond faster, for a more natural pace of interaction. PaLM 2 demonstrates robust reasoning capabilities exemplified by large improvements over PaLM on BIG-Bench and other reasoning tasks. PaLM 2 exhibits stable performance on a suite of responsible AI evaluations, and enables inference-time control over toxicity without additional overhead or impact on other capabilities. Overall, PaLM 2 achieves state-of-the-art performance across a diverse set of tasks and capabilities. When discussing the PaLM 2 family, it is important to distinguish between pre-trained models (of various sizes), fine-tuned variants of these models, and the user-facing products that use these models. In particular, user-facing products typically include additional pre- and post-processing steps. Additionally, the underlying models may evolve over time. Therefore, one should not expect the performance of user-facing products to exactly match the results reported in this report.},
urldate = {2024-11-15},
publisher = {arXiv},
author = {Anil, Rohan and Dai, Andrew M. and Firat, Orhan and Johnson, Melvin and Lepikhin, Dmitry and Passos, Alexandre and Shakeri, Siamak and Taropa, Emanuel and Bailey, Paige and Chen, Zhifeng and Chu, Eric and Clark, Jonathan H. and Shafey, Laurent El and Huang, Yanping and Meier-Hellstern, Kathy and Mishra, Gaurav and Moreira, Erica and Omernick, Mark and Robinson, Kevin and Ruder, Sebastian and Tay, Yi and Xiao, Kefan and Xu, Yuanzhong and Zhang, Yujing and Abrego, Gustavo Hernandez and Ahn, Junwhan and Austin, Jacob and Barham, Paul and Botha, Jan and Bradbury, James and Brahma, Siddhartha and Brooks, Kevin and Catasta, Michele and Cheng, Yong and Cherry, Colin and Choquette-Choo, Christopher A. and Chowdhery, Aakanksha and Crepy, Clément and Dave, Shachi and Dehghani, Mostafa and Dev, Sunipa and Devlin, Jacob and Díaz, Mark and Du, Nan and Dyer, Ethan and Feinberg, Vlad and Feng, Fangxiaoyu and Fienber, Vlad and Freitag, Markus and Garcia, Xavier and Gehrmann, Sebastian and Gonzalez, Lucas and Gur-Ari, Guy and Hand, Steven and Hashemi, Hadi and Hou, Le and Howland, Joshua and Hu, Andrea and Hui, Jeffrey and Hurwitz, Jeremy and Isard, Michael and Ittycheriah, Abe and Jagielski, Matthew and Jia, Wenhao and Kenealy, Kathleen and Krikun, Maxim and Kudugunta, Sneha and Lan, Chang and Lee, Katherine and Lee, Benjamin and Li, Eric and Li, Music and Li, Wei and Li, YaGuang and Li, Jian and Lim, Hyeontaek and Lin, Hanzhao and Liu, Zhongtao and Liu, Frederick and Maggioni, Marcello and Mahendru, Aroma and Maynez, Joshua and Misra, Vedant and Moussalem, Maysam and Nado, Zachary and Nham, John and Ni, Eric and Nystrom, Andrew and Parrish, Alicia and Pellat, Marie and Polacek, Martin and Polozov, Alex and Pope, Reiner and Qiao, Siyuan and Reif, Emily and Richter, Bryan and Riley, Parker and Ros, Alex Castro and Roy, Aurko and Saeta, Brennan and Samuel, Rajkumar and Shelby, Renee and Slone, Ambrose and Smilkov, Daniel and So, David R. and Sohn, Daniel and Tokumine, Simon and Valter, Dasha and Vasudevan, Vijay and Vodrahalli, Kiran and Wang, Xuezhi and Wang, Pidong and Wang, Zirui and Wang, Tao and Wieting, John and Wu, Yuhuai and Xu, Kelvin and Xu, Yunhan and Xue, Linting and Yin, Pengcheng and Yu, Jiahui and Zhang, Qiao and Zheng, Steven and Zheng, Ce and Zhou, Weikang and Zhou, Denny and Petrov, Slav and Wu, Yonghui},
month = sep,
year = {2023},
note = {arXiv:2305.10403 [cs]},
keywords = {Computer Science - Artificial Intelligence, Computer Science - Computation and Language},
}
Downloads: 0
{"_id":"zKBjq8xpG4YpMrD2P","bibbaseid":"anil-dai-firat-johnson-lepikhin-passos-shakeri-taropa-etal-palm2technicalreport-2023","author_short":["Anil, R.","Dai, A. M.","Firat, O.","Johnson, M.","Lepikhin, D.","Passos, A.","Shakeri, S.","Taropa, E.","Bailey, P.","Chen, Z.","Chu, E.","Clark, J. H.","Shafey, L. E.","Huang, Y.","Meier-Hellstern, K.","Mishra, G.","Moreira, E.","Omernick, M.","Robinson, K.","Ruder, S.","Tay, Y.","Xiao, K.","Xu, Y.","Zhang, Y.","Abrego, G. H.","Ahn, J.","Austin, J.","Barham, P.","Botha, J.","Bradbury, J.","Brahma, S.","Brooks, K.","Catasta, M.","Cheng, Y.","Cherry, C.","Choquette-Choo, C. A.","Chowdhery, A.","Crepy, C.","Dave, S.","Dehghani, M.","Dev, S.","Devlin, J.","Díaz, M.","Du, N.","Dyer, E.","Feinberg, V.","Feng, F.","Fienber, V.","Freitag, M.","Garcia, X.","Gehrmann, S.","Gonzalez, L.","Gur-Ari, G.","Hand, S.","Hashemi, H.","Hou, L.","Howland, J.","Hu, A.","Hui, J.","Hurwitz, J.","Isard, M.","Ittycheriah, A.","Jagielski, M.","Jia, W.","Kenealy, K.","Krikun, M.","Kudugunta, S.","Lan, C.","Lee, K.","Lee, B.","Li, E.","Li, M.","Li, W.","Li, Y.","Li, J.","Lim, H.","Lin, H.","Liu, Z.","Liu, F.","Maggioni, M.","Mahendru, A.","Maynez, J.","Misra, V.","Moussalem, M.","Nado, Z.","Nham, J.","Ni, E.","Nystrom, A.","Parrish, A.","Pellat, M.","Polacek, M.","Polozov, A.","Pope, R.","Qiao, S.","Reif, E.","Richter, B.","Riley, P.","Ros, A. C.","Roy, A.","Saeta, B.","Samuel, R.","Shelby, R.","Slone, A.","Smilkov, D.","So, D. R.","Sohn, D.","Tokumine, S.","Valter, D.","Vasudevan, V.","Vodrahalli, K.","Wang, X.","Wang, P.","Wang, Z.","Wang, T.","Wieting, J.","Wu, Y.","Xu, K.","Xu, Y.","Xue, L.","Yin, P.","Yu, J.","Zhang, Q.","Zheng, S.","Zheng, C.","Zhou, W.","Zhou, D.","Petrov, S.","Wu, Y."],"bibdata":{"bibtype":"misc","type":"misc","title":"PaLM 2 Technical Report","url":"http://arxiv.org/abs/2305.10403","abstract":"We introduce PaLM 2, a new state-of-the-art language model that has better multilingual and reasoning capabilities and is more compute-efficient than its predecessor PaLM. PaLM 2 is a Transformer-based model trained using a mixture of objectives. Through extensive evaluations on English and multilingual language, and reasoning tasks, we demonstrate that PaLM 2 has significantly improved quality on downstream tasks across different model sizes, while simultaneously exhibiting faster and more efficient inference compared to PaLM. This improved efficiency enables broader deployment while also allowing the model to respond faster, for a more natural pace of interaction. PaLM 2 demonstrates robust reasoning capabilities exemplified by large improvements over PaLM on BIG-Bench and other reasoning tasks. PaLM 2 exhibits stable performance on a suite of responsible AI evaluations, and enables inference-time control over toxicity without additional overhead or impact on other capabilities. Overall, PaLM 2 achieves state-of-the-art performance across a diverse set of tasks and capabilities. When discussing the PaLM 2 family, it is important to distinguish between pre-trained models (of various sizes), fine-tuned variants of these models, and the user-facing products that use these models. In particular, user-facing products typically include additional pre- and post-processing steps. Additionally, the underlying models may evolve over time. Therefore, one should not expect the performance of user-facing products to exactly match the results reported in this report.","urldate":"2024-11-15","publisher":"arXiv","author":[{"propositions":[],"lastnames":["Anil"],"firstnames":["Rohan"],"suffixes":[]},{"propositions":[],"lastnames":["Dai"],"firstnames":["Andrew","M."],"suffixes":[]},{"propositions":[],"lastnames":["Firat"],"firstnames":["Orhan"],"suffixes":[]},{"propositions":[],"lastnames":["Johnson"],"firstnames":["Melvin"],"suffixes":[]},{"propositions":[],"lastnames":["Lepikhin"],"firstnames":["Dmitry"],"suffixes":[]},{"propositions":[],"lastnames":["Passos"],"firstnames":["Alexandre"],"suffixes":[]},{"propositions":[],"lastnames":["Shakeri"],"firstnames":["Siamak"],"suffixes":[]},{"propositions":[],"lastnames":["Taropa"],"firstnames":["Emanuel"],"suffixes":[]},{"propositions":[],"lastnames":["Bailey"],"firstnames":["Paige"],"suffixes":[]},{"propositions":[],"lastnames":["Chen"],"firstnames":["Zhifeng"],"suffixes":[]},{"propositions":[],"lastnames":["Chu"],"firstnames":["Eric"],"suffixes":[]},{"propositions":[],"lastnames":["Clark"],"firstnames":["Jonathan","H."],"suffixes":[]},{"propositions":[],"lastnames":["Shafey"],"firstnames":["Laurent","El"],"suffixes":[]},{"propositions":[],"lastnames":["Huang"],"firstnames":["Yanping"],"suffixes":[]},{"propositions":[],"lastnames":["Meier-Hellstern"],"firstnames":["Kathy"],"suffixes":[]},{"propositions":[],"lastnames":["Mishra"],"firstnames":["Gaurav"],"suffixes":[]},{"propositions":[],"lastnames":["Moreira"],"firstnames":["Erica"],"suffixes":[]},{"propositions":[],"lastnames":["Omernick"],"firstnames":["Mark"],"suffixes":[]},{"propositions":[],"lastnames":["Robinson"],"firstnames":["Kevin"],"suffixes":[]},{"propositions":[],"lastnames":["Ruder"],"firstnames":["Sebastian"],"suffixes":[]},{"propositions":[],"lastnames":["Tay"],"firstnames":["Yi"],"suffixes":[]},{"propositions":[],"lastnames":["Xiao"],"firstnames":["Kefan"],"suffixes":[]},{"propositions":[],"lastnames":["Xu"],"firstnames":["Yuanzhong"],"suffixes":[]},{"propositions":[],"lastnames":["Zhang"],"firstnames":["Yujing"],"suffixes":[]},{"propositions":[],"lastnames":["Abrego"],"firstnames":["Gustavo","Hernandez"],"suffixes":[]},{"propositions":[],"lastnames":["Ahn"],"firstnames":["Junwhan"],"suffixes":[]},{"propositions":[],"lastnames":["Austin"],"firstnames":["Jacob"],"suffixes":[]},{"propositions":[],"lastnames":["Barham"],"firstnames":["Paul"],"suffixes":[]},{"propositions":[],"lastnames":["Botha"],"firstnames":["Jan"],"suffixes":[]},{"propositions":[],"lastnames":["Bradbury"],"firstnames":["James"],"suffixes":[]},{"propositions":[],"lastnames":["Brahma"],"firstnames":["Siddhartha"],"suffixes":[]},{"propositions":[],"lastnames":["Brooks"],"firstnames":["Kevin"],"suffixes":[]},{"propositions":[],"lastnames":["Catasta"],"firstnames":["Michele"],"suffixes":[]},{"propositions":[],"lastnames":["Cheng"],"firstnames":["Yong"],"suffixes":[]},{"propositions":[],"lastnames":["Cherry"],"firstnames":["Colin"],"suffixes":[]},{"propositions":[],"lastnames":["Choquette-Choo"],"firstnames":["Christopher","A."],"suffixes":[]},{"propositions":[],"lastnames":["Chowdhery"],"firstnames":["Aakanksha"],"suffixes":[]},{"propositions":[],"lastnames":["Crepy"],"firstnames":["Clément"],"suffixes":[]},{"propositions":[],"lastnames":["Dave"],"firstnames":["Shachi"],"suffixes":[]},{"propositions":[],"lastnames":["Dehghani"],"firstnames":["Mostafa"],"suffixes":[]},{"propositions":[],"lastnames":["Dev"],"firstnames":["Sunipa"],"suffixes":[]},{"propositions":[],"lastnames":["Devlin"],"firstnames":["Jacob"],"suffixes":[]},{"propositions":[],"lastnames":["Díaz"],"firstnames":["Mark"],"suffixes":[]},{"propositions":[],"lastnames":["Du"],"firstnames":["Nan"],"suffixes":[]},{"propositions":[],"lastnames":["Dyer"],"firstnames":["Ethan"],"suffixes":[]},{"propositions":[],"lastnames":["Feinberg"],"firstnames":["Vlad"],"suffixes":[]},{"propositions":[],"lastnames":["Feng"],"firstnames":["Fangxiaoyu"],"suffixes":[]},{"propositions":[],"lastnames":["Fienber"],"firstnames":["Vlad"],"suffixes":[]},{"propositions":[],"lastnames":["Freitag"],"firstnames":["Markus"],"suffixes":[]},{"propositions":[],"lastnames":["Garcia"],"firstnames":["Xavier"],"suffixes":[]},{"propositions":[],"lastnames":["Gehrmann"],"firstnames":["Sebastian"],"suffixes":[]},{"propositions":[],"lastnames":["Gonzalez"],"firstnames":["Lucas"],"suffixes":[]},{"propositions":[],"lastnames":["Gur-Ari"],"firstnames":["Guy"],"suffixes":[]},{"propositions":[],"lastnames":["Hand"],"firstnames":["Steven"],"suffixes":[]},{"propositions":[],"lastnames":["Hashemi"],"firstnames":["Hadi"],"suffixes":[]},{"propositions":[],"lastnames":["Hou"],"firstnames":["Le"],"suffixes":[]},{"propositions":[],"lastnames":["Howland"],"firstnames":["Joshua"],"suffixes":[]},{"propositions":[],"lastnames":["Hu"],"firstnames":["Andrea"],"suffixes":[]},{"propositions":[],"lastnames":["Hui"],"firstnames":["Jeffrey"],"suffixes":[]},{"propositions":[],"lastnames":["Hurwitz"],"firstnames":["Jeremy"],"suffixes":[]},{"propositions":[],"lastnames":["Isard"],"firstnames":["Michael"],"suffixes":[]},{"propositions":[],"lastnames":["Ittycheriah"],"firstnames":["Abe"],"suffixes":[]},{"propositions":[],"lastnames":["Jagielski"],"firstnames":["Matthew"],"suffixes":[]},{"propositions":[],"lastnames":["Jia"],"firstnames":["Wenhao"],"suffixes":[]},{"propositions":[],"lastnames":["Kenealy"],"firstnames":["Kathleen"],"suffixes":[]},{"propositions":[],"lastnames":["Krikun"],"firstnames":["Maxim"],"suffixes":[]},{"propositions":[],"lastnames":["Kudugunta"],"firstnames":["Sneha"],"suffixes":[]},{"propositions":[],"lastnames":["Lan"],"firstnames":["Chang"],"suffixes":[]},{"propositions":[],"lastnames":["Lee"],"firstnames":["Katherine"],"suffixes":[]},{"propositions":[],"lastnames":["Lee"],"firstnames":["Benjamin"],"suffixes":[]},{"propositions":[],"lastnames":["Li"],"firstnames":["Eric"],"suffixes":[]},{"propositions":[],"lastnames":["Li"],"firstnames":["Music"],"suffixes":[]},{"propositions":[],"lastnames":["Li"],"firstnames":["Wei"],"suffixes":[]},{"propositions":[],"lastnames":["Li"],"firstnames":["YaGuang"],"suffixes":[]},{"propositions":[],"lastnames":["Li"],"firstnames":["Jian"],"suffixes":[]},{"propositions":[],"lastnames":["Lim"],"firstnames":["Hyeontaek"],"suffixes":[]},{"propositions":[],"lastnames":["Lin"],"firstnames":["Hanzhao"],"suffixes":[]},{"propositions":[],"lastnames":["Liu"],"firstnames":["Zhongtao"],"suffixes":[]},{"propositions":[],"lastnames":["Liu"],"firstnames":["Frederick"],"suffixes":[]},{"propositions":[],"lastnames":["Maggioni"],"firstnames":["Marcello"],"suffixes":[]},{"propositions":[],"lastnames":["Mahendru"],"firstnames":["Aroma"],"suffixes":[]},{"propositions":[],"lastnames":["Maynez"],"firstnames":["Joshua"],"suffixes":[]},{"propositions":[],"lastnames":["Misra"],"firstnames":["Vedant"],"suffixes":[]},{"propositions":[],"lastnames":["Moussalem"],"firstnames":["Maysam"],"suffixes":[]},{"propositions":[],"lastnames":["Nado"],"firstnames":["Zachary"],"suffixes":[]},{"propositions":[],"lastnames":["Nham"],"firstnames":["John"],"suffixes":[]},{"propositions":[],"lastnames":["Ni"],"firstnames":["Eric"],"suffixes":[]},{"propositions":[],"lastnames":["Nystrom"],"firstnames":["Andrew"],"suffixes":[]},{"propositions":[],"lastnames":["Parrish"],"firstnames":["Alicia"],"suffixes":[]},{"propositions":[],"lastnames":["Pellat"],"firstnames":["Marie"],"suffixes":[]},{"propositions":[],"lastnames":["Polacek"],"firstnames":["Martin"],"suffixes":[]},{"propositions":[],"lastnames":["Polozov"],"firstnames":["Alex"],"suffixes":[]},{"propositions":[],"lastnames":["Pope"],"firstnames":["Reiner"],"suffixes":[]},{"propositions":[],"lastnames":["Qiao"],"firstnames":["Siyuan"],"suffixes":[]},{"propositions":[],"lastnames":["Reif"],"firstnames":["Emily"],"suffixes":[]},{"propositions":[],"lastnames":["Richter"],"firstnames":["Bryan"],"suffixes":[]},{"propositions":[],"lastnames":["Riley"],"firstnames":["Parker"],"suffixes":[]},{"propositions":[],"lastnames":["Ros"],"firstnames":["Alex","Castro"],"suffixes":[]},{"propositions":[],"lastnames":["Roy"],"firstnames":["Aurko"],"suffixes":[]},{"propositions":[],"lastnames":["Saeta"],"firstnames":["Brennan"],"suffixes":[]},{"propositions":[],"lastnames":["Samuel"],"firstnames":["Rajkumar"],"suffixes":[]},{"propositions":[],"lastnames":["Shelby"],"firstnames":["Renee"],"suffixes":[]},{"propositions":[],"lastnames":["Slone"],"firstnames":["Ambrose"],"suffixes":[]},{"propositions":[],"lastnames":["Smilkov"],"firstnames":["Daniel"],"suffixes":[]},{"propositions":[],"lastnames":["So"],"firstnames":["David","R."],"suffixes":[]},{"propositions":[],"lastnames":["Sohn"],"firstnames":["Daniel"],"suffixes":[]},{"propositions":[],"lastnames":["Tokumine"],"firstnames":["Simon"],"suffixes":[]},{"propositions":[],"lastnames":["Valter"],"firstnames":["Dasha"],"suffixes":[]},{"propositions":[],"lastnames":["Vasudevan"],"firstnames":["Vijay"],"suffixes":[]},{"propositions":[],"lastnames":["Vodrahalli"],"firstnames":["Kiran"],"suffixes":[]},{"propositions":[],"lastnames":["Wang"],"firstnames":["Xuezhi"],"suffixes":[]},{"propositions":[],"lastnames":["Wang"],"firstnames":["Pidong"],"suffixes":[]},{"propositions":[],"lastnames":["Wang"],"firstnames":["Zirui"],"suffixes":[]},{"propositions":[],"lastnames":["Wang"],"firstnames":["Tao"],"suffixes":[]},{"propositions":[],"lastnames":["Wieting"],"firstnames":["John"],"suffixes":[]},{"propositions":[],"lastnames":["Wu"],"firstnames":["Yuhuai"],"suffixes":[]},{"propositions":[],"lastnames":["Xu"],"firstnames":["Kelvin"],"suffixes":[]},{"propositions":[],"lastnames":["Xu"],"firstnames":["Yunhan"],"suffixes":[]},{"propositions":[],"lastnames":["Xue"],"firstnames":["Linting"],"suffixes":[]},{"propositions":[],"lastnames":["Yin"],"firstnames":["Pengcheng"],"suffixes":[]},{"propositions":[],"lastnames":["Yu"],"firstnames":["Jiahui"],"suffixes":[]},{"propositions":[],"lastnames":["Zhang"],"firstnames":["Qiao"],"suffixes":[]},{"propositions":[],"lastnames":["Zheng"],"firstnames":["Steven"],"suffixes":[]},{"propositions":[],"lastnames":["Zheng"],"firstnames":["Ce"],"suffixes":[]},{"propositions":[],"lastnames":["Zhou"],"firstnames":["Weikang"],"suffixes":[]},{"propositions":[],"lastnames":["Zhou"],"firstnames":["Denny"],"suffixes":[]},{"propositions":[],"lastnames":["Petrov"],"firstnames":["Slav"],"suffixes":[]},{"propositions":[],"lastnames":["Wu"],"firstnames":["Yonghui"],"suffixes":[]}],"month":"September","year":"2023","note":"arXiv:2305.10403 [cs]","keywords":"Computer Science - Artificial Intelligence, Computer Science - Computation and Language","bibtex":"@misc{anil_palm_2023,\n\ttitle = {{PaLM} 2 {Technical} {Report}},\n\turl = {http://arxiv.org/abs/2305.10403},\n\tabstract = {We introduce PaLM 2, a new state-of-the-art language model that has better multilingual and reasoning capabilities and is more compute-efficient than its predecessor PaLM. PaLM 2 is a Transformer-based model trained using a mixture of objectives. Through extensive evaluations on English and multilingual language, and reasoning tasks, we demonstrate that PaLM 2 has significantly improved quality on downstream tasks across different model sizes, while simultaneously exhibiting faster and more efficient inference compared to PaLM. This improved efficiency enables broader deployment while also allowing the model to respond faster, for a more natural pace of interaction. PaLM 2 demonstrates robust reasoning capabilities exemplified by large improvements over PaLM on BIG-Bench and other reasoning tasks. PaLM 2 exhibits stable performance on a suite of responsible AI evaluations, and enables inference-time control over toxicity without additional overhead or impact on other capabilities. Overall, PaLM 2 achieves state-of-the-art performance across a diverse set of tasks and capabilities. When discussing the PaLM 2 family, it is important to distinguish between pre-trained models (of various sizes), fine-tuned variants of these models, and the user-facing products that use these models. In particular, user-facing products typically include additional pre- and post-processing steps. Additionally, the underlying models may evolve over time. Therefore, one should not expect the performance of user-facing products to exactly match the results reported in this report.},\n\turldate = {2024-11-15},\n\tpublisher = {arXiv},\n\tauthor = {Anil, Rohan and Dai, Andrew M. and Firat, Orhan and Johnson, Melvin and Lepikhin, Dmitry and Passos, Alexandre and Shakeri, Siamak and Taropa, Emanuel and Bailey, Paige and Chen, Zhifeng and Chu, Eric and Clark, Jonathan H. and Shafey, Laurent El and Huang, Yanping and Meier-Hellstern, Kathy and Mishra, Gaurav and Moreira, Erica and Omernick, Mark and Robinson, Kevin and Ruder, Sebastian and Tay, Yi and Xiao, Kefan and Xu, Yuanzhong and Zhang, Yujing and Abrego, Gustavo Hernandez and Ahn, Junwhan and Austin, Jacob and Barham, Paul and Botha, Jan and Bradbury, James and Brahma, Siddhartha and Brooks, Kevin and Catasta, Michele and Cheng, Yong and Cherry, Colin and Choquette-Choo, Christopher A. and Chowdhery, Aakanksha and Crepy, Clément and Dave, Shachi and Dehghani, Mostafa and Dev, Sunipa and Devlin, Jacob and Díaz, Mark and Du, Nan and Dyer, Ethan and Feinberg, Vlad and Feng, Fangxiaoyu and Fienber, Vlad and Freitag, Markus and Garcia, Xavier and Gehrmann, Sebastian and Gonzalez, Lucas and Gur-Ari, Guy and Hand, Steven and Hashemi, Hadi and Hou, Le and Howland, Joshua and Hu, Andrea and Hui, Jeffrey and Hurwitz, Jeremy and Isard, Michael and Ittycheriah, Abe and Jagielski, Matthew and Jia, Wenhao and Kenealy, Kathleen and Krikun, Maxim and Kudugunta, Sneha and Lan, Chang and Lee, Katherine and Lee, Benjamin and Li, Eric and Li, Music and Li, Wei and Li, YaGuang and Li, Jian and Lim, Hyeontaek and Lin, Hanzhao and Liu, Zhongtao and Liu, Frederick and Maggioni, Marcello and Mahendru, Aroma and Maynez, Joshua and Misra, Vedant and Moussalem, Maysam and Nado, Zachary and Nham, John and Ni, Eric and Nystrom, Andrew and Parrish, Alicia and Pellat, Marie and Polacek, Martin and Polozov, Alex and Pope, Reiner and Qiao, Siyuan and Reif, Emily and Richter, Bryan and Riley, Parker and Ros, Alex Castro and Roy, Aurko and Saeta, Brennan and Samuel, Rajkumar and Shelby, Renee and Slone, Ambrose and Smilkov, Daniel and So, David R. and Sohn, Daniel and Tokumine, Simon and Valter, Dasha and Vasudevan, Vijay and Vodrahalli, Kiran and Wang, Xuezhi and Wang, Pidong and Wang, Zirui and Wang, Tao and Wieting, John and Wu, Yuhuai and Xu, Kelvin and Xu, Yunhan and Xue, Linting and Yin, Pengcheng and Yu, Jiahui and Zhang, Qiao and Zheng, Steven and Zheng, Ce and Zhou, Weikang and Zhou, Denny and Petrov, Slav and Wu, Yonghui},\n\tmonth = sep,\n\tyear = {2023},\n\tnote = {arXiv:2305.10403 [cs]},\n\tkeywords = {Computer Science - Artificial Intelligence, Computer Science - Computation and Language},\n}\n\n\n\n\n\n\n\n\n\n\n\n","author_short":["Anil, R.","Dai, A. M.","Firat, O.","Johnson, M.","Lepikhin, D.","Passos, A.","Shakeri, S.","Taropa, E.","Bailey, P.","Chen, Z.","Chu, E.","Clark, J. H.","Shafey, L. E.","Huang, Y.","Meier-Hellstern, K.","Mishra, G.","Moreira, E.","Omernick, M.","Robinson, K.","Ruder, S.","Tay, Y.","Xiao, K.","Xu, Y.","Zhang, Y.","Abrego, G. H.","Ahn, J.","Austin, J.","Barham, P.","Botha, J.","Bradbury, J.","Brahma, S.","Brooks, K.","Catasta, M.","Cheng, Y.","Cherry, C.","Choquette-Choo, C. A.","Chowdhery, A.","Crepy, C.","Dave, S.","Dehghani, M.","Dev, S.","Devlin, J.","Díaz, M.","Du, N.","Dyer, E.","Feinberg, V.","Feng, F.","Fienber, V.","Freitag, M.","Garcia, X.","Gehrmann, S.","Gonzalez, L.","Gur-Ari, G.","Hand, S.","Hashemi, H.","Hou, L.","Howland, J.","Hu, A.","Hui, J.","Hurwitz, J.","Isard, M.","Ittycheriah, A.","Jagielski, M.","Jia, W.","Kenealy, K.","Krikun, M.","Kudugunta, S.","Lan, C.","Lee, K.","Lee, B.","Li, E.","Li, M.","Li, W.","Li, Y.","Li, J.","Lim, H.","Lin, H.","Liu, Z.","Liu, F.","Maggioni, M.","Mahendru, A.","Maynez, J.","Misra, V.","Moussalem, M.","Nado, Z.","Nham, J.","Ni, E.","Nystrom, A.","Parrish, A.","Pellat, M.","Polacek, M.","Polozov, A.","Pope, R.","Qiao, S.","Reif, E.","Richter, B.","Riley, P.","Ros, A. C.","Roy, A.","Saeta, B.","Samuel, R.","Shelby, R.","Slone, A.","Smilkov, D.","So, D. R.","Sohn, D.","Tokumine, S.","Valter, D.","Vasudevan, V.","Vodrahalli, K.","Wang, X.","Wang, P.","Wang, Z.","Wang, T.","Wieting, J.","Wu, Y.","Xu, K.","Xu, Y.","Xue, L.","Yin, P.","Yu, J.","Zhang, Q.","Zheng, S.","Zheng, C.","Zhou, W.","Zhou, D.","Petrov, S.","Wu, Y."],"key":"anil_palm_2023","id":"anil_palm_2023","bibbaseid":"anil-dai-firat-johnson-lepikhin-passos-shakeri-taropa-etal-palm2technicalreport-2023","role":"author","urls":{"Paper":"http://arxiv.org/abs/2305.10403"},"keyword":["Computer Science - Artificial Intelligence","Computer Science - Computation and Language"],"metadata":{"authorlinks":{}},"html":""},"bibtype":"misc","biburl":"https://bibbase.org/zotero/andreasmartin","dataSources":["YtBDXPDiQEyhyEDZC","tpWeaaCgFjPTYCjg3","jurZeGzSpYdkQ8rm4"],"keywords":["computer science - artificial intelligence","computer science - computation and language"],"search_terms":["palm","technical","report","anil","dai","firat","johnson","lepikhin","passos","shakeri","taropa","bailey","chen","chu","clark","shafey","huang","meier-hellstern","mishra","moreira","omernick","robinson","ruder","tay","xiao","xu","zhang","abrego","ahn","austin","barham","botha","bradbury","brahma","brooks","catasta","cheng","cherry","choquette-choo","chowdhery","crepy","dave","dehghani","dev","devlin","díaz","du","dyer","feinberg","feng","fienber","freitag","garcia","gehrmann","gonzalez","gur-ari","hand","hashemi","hou","howland","hu","hui","hurwitz","isard","ittycheriah","jagielski","jia","kenealy","krikun","kudugunta","lan","lee","lee","li","li","li","li","li","lim","lin","liu","liu","maggioni","mahendru","maynez","misra","moussalem","nado","nham","ni","nystrom","parrish","pellat","polacek","polozov","pope","qiao","reif","richter","riley","ros","roy","saeta","samuel","shelby","slone","smilkov","so","sohn","tokumine","valter","vasudevan","vodrahalli","wang","wang","wang","wang","wieting","wu","xu","xu","xue","yin","yu","zhang","zheng","zheng","zhou","zhou","petrov","wu"],"title":"PaLM 2 Technical Report","year":2023}