Many-to-English Machine Translation Tools, Data, and Pretrained Models. \textbfGowda, Thamme, Zhang, Z., Mattmann, C., & May, J. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: System Demonstrations, pages 306–316, Online, August, 2021. Association for Computational Linguistics. Paper doi abstract bibtex While there are more than 7000 languages in the world, most translation research efforts have targeted a few high resource languages. Commercial translation systems support only one hundred languages or fewer, and do not make these models available for transfer to low resource languages. In this work, we present useful tools for machine translation research: MTData, NLCodec and RTG. We demonstrate their usefulness by creating a multilingual neural machine translation model capable of translating from 500 source languages to English. We make this multilingual model readily downloadable and usable as a service, or as a parent model for transfer-learning to even lower-resource languages.
@inproceedings{gowda-etal-2021-many,
title = "Many-to-{E}nglish Machine Translation Tools, Data, and Pretrained Models",
author = "\textbf{Gowda, Thamme} and
Zhang, Zhao and
Mattmann, Chris and
May, Jonathan",
booktitle = "Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: System Demonstrations",
month = aug,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.acl-demo.37",
doi = "10.18653/v1/2021.acl-demo.37",
pages = "306--316",
abstract = "While there are more than 7000 languages in the world, most translation research efforts have targeted a few high resource languages. Commercial translation systems support only one hundred languages or fewer, and do not make these models available for transfer to low resource languages. In this work, we present useful tools for machine translation research: MTData, NLCodec and RTG. We demonstrate their usefulness by creating a multilingual neural machine translation model capable of translating from 500 source languages to English. We make this multilingual model readily downloadable and usable as a service, or as a parent model for transfer-learning to even lower-resource languages.",
}
Downloads: 0
{"_id":"KyehRiphHogerxPR9","bibbaseid":"textbfgowda-zhang-mattmann-may-manytoenglishmachinetranslationtoolsdataandpretrainedmodels-2021","author_short":["\\textbfGowda, Thamme","Zhang, Z.","Mattmann, C.","May, J."],"bibdata":{"bibtype":"inproceedings","type":"inproceedings","title":"Many-to-English Machine Translation Tools, Data, and Pretrained Models","author":[{"firstnames":[],"propositions":[],"lastnames":["\\textbfGowda, Thamme"],"suffixes":[]},{"propositions":[],"lastnames":["Zhang"],"firstnames":["Zhao"],"suffixes":[]},{"propositions":[],"lastnames":["Mattmann"],"firstnames":["Chris"],"suffixes":[]},{"propositions":[],"lastnames":["May"],"firstnames":["Jonathan"],"suffixes":[]}],"booktitle":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: System Demonstrations","month":"August","year":"2021","address":"Online","publisher":"Association for Computational Linguistics","url":"https://aclanthology.org/2021.acl-demo.37","doi":"10.18653/v1/2021.acl-demo.37","pages":"306–316","abstract":"While there are more than 7000 languages in the world, most translation research efforts have targeted a few high resource languages. Commercial translation systems support only one hundred languages or fewer, and do not make these models available for transfer to low resource languages. In this work, we present useful tools for machine translation research: MTData, NLCodec and RTG. We demonstrate their usefulness by creating a multilingual neural machine translation model capable of translating from 500 source languages to English. We make this multilingual model readily downloadable and usable as a service, or as a parent model for transfer-learning to even lower-resource languages.","bibtex":"@inproceedings{gowda-etal-2021-many,\n title = \"Many-to-{E}nglish Machine Translation Tools, Data, and Pretrained Models\",\n author = \"\\textbf{Gowda, Thamme} and\n Zhang, Zhao and\n Mattmann, Chris and\n May, Jonathan\",\n booktitle = \"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: System Demonstrations\",\n month = aug,\n year = \"2021\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2021.acl-demo.37\",\n doi = \"10.18653/v1/2021.acl-demo.37\",\n pages = \"306--316\",\n abstract = \"While there are more than 7000 languages in the world, most translation research efforts have targeted a few high resource languages. Commercial translation systems support only one hundred languages or fewer, and do not make these models available for transfer to low resource languages. In this work, we present useful tools for machine translation research: MTData, NLCodec and RTG. We demonstrate their usefulness by creating a multilingual neural machine translation model capable of translating from 500 source languages to English. We make this multilingual model readily downloadable and usable as a service, or as a parent model for transfer-learning to even lower-resource languages.\",\n}\n\n\n","author_short":["\\textbfGowda, Thamme","Zhang, Z.","Mattmann, C.","May, J."],"bibbaseid":"textbfgowda-zhang-mattmann-may-manytoenglishmachinetranslationtoolsdataandpretrainedmodels-2021","role":"author","urls":{"Paper":"https://aclanthology.org/2021.acl-demo.37"},"metadata":{"authorlinks":{}}},"bibtype":"inproceedings","biburl":"https://bibbase.org/f/TnumWo2CpGS8ziR7g/tg.bib","dataSources":["Yq795aWkkPdk8PFkL"],"keywords":[],"search_terms":["many","english","machine","translation","tools","data","pretrained","models","\\textbfgowda","zhang","mattmann","may"],"title":"Many-to-English Machine Translation Tools, Data, and Pretrained Models","year":2021}