Distributional Semantics for Neo-Latin. Bloem, J., Parisi, M. C., Reynaert, M., Oortwijn, Y., & Betti, A. In Proceedings of LT4HALA 2020 - 1st Workshop on Language Technologies for Historical and Ancient Languages, pages 84–93, Marseille, France, May, 2020. European Language Resources Association (ELRA). Paper abstract bibtex 4 downloads We address the problem of creating and evaluating quality Neo-Latin word embeddings for the purpose of philosophical research, adapting the Nonce2Vec tool to learn embeddings from Neo-Latin sentences. This distributional semantic modeling tool can learn from tiny data incrementally, using a larger background corpus for initialization. We conduct two evaluation tasks: definitional learning of Latin Wikipedia terms, and learning consistent embeddings from 18th century Neo-Latin sentences pertaining to the concept of mathematical method. Our results show that consistent Neo-Latin word embeddings can be learned from this type of data. While our evaluation results are promising, they do not reveal to what extent the learned models match domain expert knowledge of our Neo-Latin texts. Therefore, we propose an additional evaluation method, grounded in expert-annotated data, that would assess whether learned representations are conceptually sound in relation to the domain of study.
@inproceedings{bloem_distributional_2020,
address = {Marseille, France},
title = {Distributional {Semantics} for {Neo}-{Latin}},
isbn = {979-10-95546-53-5},
url = {https://www.aclweb.org/anthology/2020.lt4hala-1.13},
abstract = {We address the problem of creating and evaluating quality Neo-Latin word embeddings for the purpose of philosophical research, adapting the Nonce2Vec tool to learn embeddings from Neo-Latin sentences. This distributional semantic modeling tool can learn from tiny data incrementally, using a larger background corpus for initialization. We conduct two evaluation tasks: definitional learning of Latin Wikipedia terms, and learning consistent embeddings from 18th century Neo-Latin sentences pertaining to the concept of mathematical method. Our results show that consistent Neo-Latin word embeddings can be learned from this type of data. While our evaluation results are promising, they do not reveal to what extent the learned models match domain expert knowledge of our Neo-Latin texts. Therefore, we propose an additional evaluation method, grounded in expert-annotated data, that would assess whether learned representations are conceptually sound in relation to the domain of study.},
language = {English},
urldate = {2021-01-17},
booktitle = {Proceedings of {LT4HALA} 2020 - 1st {Workshop} on {Language} {Technologies} for {Historical} and {Ancient} {Languages}},
publisher = {European Language Resources Association (ELRA)},
author = {Bloem, Jelke and Parisi, Maria Chiara and Reynaert, Martin and Oortwijn, Yvette and Betti, Arianna},
month = may,
year = {2020},
pages = {84--93},
}
Downloads: 4
{"_id":"zuM8rvRXcj5kwHBJS","bibbaseid":"bloem-parisi-reynaert-oortwijn-betti-distributionalsemanticsforneolatin-2020","authorIDs":["aeqmsSj8anitxzHwC"],"author_short":["Bloem, J.","Parisi, M. C.","Reynaert, M.","Oortwijn, Y.","Betti, A."],"bibdata":{"bibtype":"inproceedings","type":"inproceedings","address":"Marseille, France","title":"Distributional Semantics for Neo-Latin","isbn":"979-10-95546-53-5","url":"https://www.aclweb.org/anthology/2020.lt4hala-1.13","abstract":"We address the problem of creating and evaluating quality Neo-Latin word embeddings for the purpose of philosophical research, adapting the Nonce2Vec tool to learn embeddings from Neo-Latin sentences. This distributional semantic modeling tool can learn from tiny data incrementally, using a larger background corpus for initialization. We conduct two evaluation tasks: definitional learning of Latin Wikipedia terms, and learning consistent embeddings from 18th century Neo-Latin sentences pertaining to the concept of mathematical method. Our results show that consistent Neo-Latin word embeddings can be learned from this type of data. While our evaluation results are promising, they do not reveal to what extent the learned models match domain expert knowledge of our Neo-Latin texts. Therefore, we propose an additional evaluation method, grounded in expert-annotated data, that would assess whether learned representations are conceptually sound in relation to the domain of study.","language":"English","urldate":"2021-01-17","booktitle":"Proceedings of LT4HALA 2020 - 1st Workshop on Language Technologies for Historical and Ancient Languages","publisher":"European Language Resources Association (ELRA)","author":[{"propositions":[],"lastnames":["Bloem"],"firstnames":["Jelke"],"suffixes":[]},{"propositions":[],"lastnames":["Parisi"],"firstnames":["Maria","Chiara"],"suffixes":[]},{"propositions":[],"lastnames":["Reynaert"],"firstnames":["Martin"],"suffixes":[]},{"propositions":[],"lastnames":["Oortwijn"],"firstnames":["Yvette"],"suffixes":[]},{"propositions":[],"lastnames":["Betti"],"firstnames":["Arianna"],"suffixes":[]}],"month":"May","year":"2020","pages":"84–93","bibtex":"@inproceedings{bloem_distributional_2020,\n\taddress = {Marseille, France},\n\ttitle = {Distributional {Semantics} for {Neo}-{Latin}},\n\tisbn = {979-10-95546-53-5},\n\turl = {https://www.aclweb.org/anthology/2020.lt4hala-1.13},\n\tabstract = {We address the problem of creating and evaluating quality Neo-Latin word embeddings for the purpose of philosophical research, adapting the Nonce2Vec tool to learn embeddings from Neo-Latin sentences. This distributional semantic modeling tool can learn from tiny data incrementally, using a larger background corpus for initialization. We conduct two evaluation tasks: definitional learning of Latin Wikipedia terms, and learning consistent embeddings from 18th century Neo-Latin sentences pertaining to the concept of mathematical method. Our results show that consistent Neo-Latin word embeddings can be learned from this type of data. While our evaluation results are promising, they do not reveal to what extent the learned models match domain expert knowledge of our Neo-Latin texts. Therefore, we propose an additional evaluation method, grounded in expert-annotated data, that would assess whether learned representations are conceptually sound in relation to the domain of study.},\n\tlanguage = {English},\n\turldate = {2021-01-17},\n\tbooktitle = {Proceedings of {LT4HALA} 2020 - 1st {Workshop} on {Language} {Technologies} for {Historical} and {Ancient} {Languages}},\n\tpublisher = {European Language Resources Association (ELRA)},\n\tauthor = {Bloem, Jelke and Parisi, Maria Chiara and Reynaert, Martin and Oortwijn, Yvette and Betti, Arianna},\n\tmonth = may,\n\tyear = {2020},\n\tpages = {84--93},\n}\n\n","author_short":["Bloem, J.","Parisi, M. C.","Reynaert, M.","Oortwijn, Y.","Betti, A."],"key":"bloem_distributional_2020","id":"bloem_distributional_2020","bibbaseid":"bloem-parisi-reynaert-oortwijn-betti-distributionalsemanticsforneolatin-2020","role":"author","urls":{"Paper":"https://www.aclweb.org/anthology/2020.lt4hala-1.13"},"metadata":{"authorlinks":{"betti, a":"https://bibbase.org/show?bib=https%3A%2F%2Fapi.zotero.org%2Fgroups%2F214700%2Fitems%3Fkey%3Dbi2Q7duoPuqjf6lgym4TgM83%26format%3Dbibtex%26limit%3D100&group0=year&css=www.axiom.humanities.uva.nl/axiom.css"}},"downloads":4},"bibtype":"inproceedings","biburl":"https://api.zotero.org/groups/2493581/items?key=esJ5YGDX6948PQKQSPOfhZpO&format=bibtex&limit=100","creationDate":"2020-04-29T13:21:12.395Z","downloads":4,"keywords":[],"search_terms":["distributional","semantics","neo","latin","bloem","parisi","reynaert","oortwijn","betti"],"title":"Distributional Semantics for Neo-Latin","year":2020,"dataSources":["ZHdsvgePffKRdgqdo","ATwFjrHEsqA4GDy8B","YqqB8NWSxm4bnL8ra","W9xJhCsMa9uhxgDWu","9kNXu23ABH7j4wxXZ","zDZS7QvC6khJT2mcu","GYdQLRi8nMpHxCYKf"]}