var bibbase_data = {"data":"\"Loading..\"\n\n
\n\n \n\n \n\n \n \n\n \n\n \n \n\n \n\n \n
\n generated by\n \n \"bibbase.org\"\n\n \n
\n \n\n
\n\n \n\n\n
\n\n Excellent! Next you can\n create a new website with this list, or\n embed it in an existing web page by copying & pasting\n any of the following snippets.\n\n
\n JavaScript\n (easiest)\n
\n \n <script src=\"https://bibbase.org/show?bib=https%3A%2F%2Fraw.githubusercontent.com%2Fdmg-illc%2Fdream%2Fmaster%2Fdream-publications.bib&theme=simple&hidemenu=true&commas=true&fullnames=1&jsonp=1&jsonp=1\"></script>\n \n
\n\n PHP\n
\n \n <?php\n $contents = file_get_contents(\"https://bibbase.org/show?bib=https%3A%2F%2Fraw.githubusercontent.com%2Fdmg-illc%2Fdream%2Fmaster%2Fdream-publications.bib&theme=simple&hidemenu=true&commas=true&fullnames=1&jsonp=1\");\n print_r($contents);\n ?>\n \n
\n\n iFrame\n (not recommended)\n
\n \n <iframe src=\"https://bibbase.org/show?bib=https%3A%2F%2Fraw.githubusercontent.com%2Fdmg-illc%2Fdream%2Fmaster%2Fdream-publications.bib&theme=simple&hidemenu=true&commas=true&fullnames=1&jsonp=1\"></iframe>\n \n
\n\n

\n For more details see the documention.\n

\n
\n
\n\n
\n\n This is a preview! To use this list on your own web site\n or create a new web site from it,\n create a free account. The file will be added\n and you will be able to edit it in the File Manager.\n We will show you instructions once you've created your account.\n
\n\n
\n\n

To the site owner:

\n\n

Action required! Mendeley is changing its\n API. In order to keep using Mendeley with BibBase past April\n 14th, you need to:\n

    \n
  1. renew the authorization for BibBase on Mendeley, and
  2. \n
  3. update the BibBase URL\n in your page the same way you did when you initially set up\n this page.\n
  4. \n
\n

\n\n

\n \n \n Fix it now\n

\n
\n\n
\n\n\n
\n \n \n
\n
\n  \n 2024\n \n \n (2)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Asking the Right Question at the Right Time: Human and Model Uncertainty Guidance To Ask Clarification Questions.\n \n \n \n \n\n\n \n Alberto Testoni, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the European Chapter of the Association for Computational Linguistics (EACL), 2024. \n To appear\n\n\n\n
\n\n\n\n \n \n \"AskingPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 6 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{testoni-fernandez-eacl-2024,\n    title = "Asking the Right Question at the Right Time: Human and Model Uncertainty Guidance To Ask Clarification Questions",\n    author = "Alberto Testoni and Raquel Fern{\\'a}ndez",\n    booktitle = "Proceedings of the European Chapter of the Association for Computational Linguistics (EACL)",\n    year = "2024",\n    url = "https://arxiv.org/abs/2402.06509",\n    abstract = "Clarification questions are an essential dialogue tool to signal misunderstanding, ambiguities, and under-specification in language use. While humans are able to resolve uncertainty by asking questions since childhood, modern dialogue systems struggle to generate effective questions. To make progress in this direction, in this work we take a collaborative dialogue task as a testbed and study how model uncertainty relates to human uncertainty -- an as yet under-explored problem. We show that model uncertainty does not mirror human clarification-seeking behavior, which suggests that using human clarification questions as supervision for deciding when to ask may not be the most effective way to resolve model uncertainty. To address this issue, we propose an approach to generating clarification questions based on model uncertainty estimation, compare it to several alternatives, and show that it leads to significant improvements in terms of task success. Our findings highlight the importance of equipping dialogue systems with the ability to assess their own uncertainty and exploit in interaction.",\n    note = "To appear"\n}\n\n
\n
\n\n\n
\n Clarification questions are an essential dialogue tool to signal misunderstanding, ambiguities, and under-specification in language use. While humans are able to resolve uncertainty by asking questions since childhood, modern dialogue systems struggle to generate effective questions. To make progress in this direction, in this work we take a collaborative dialogue task as a testbed and study how model uncertainty relates to human uncertainty – an as yet under-explored problem. We show that model uncertainty does not mirror human clarification-seeking behavior, which suggests that using human clarification questions as supervision for deciding when to ask may not be the most effective way to resolve model uncertainty. To address this issue, we propose an approach to generating clarification questions based on model uncertainty estimation, compare it to several alternatives, and show that it leads to significant improvements in terms of task success. Our findings highlight the importance of equipping dialogue systems with the ability to assess their own uncertainty and exploit in interaction.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Describing Images Fast and Slow: Quantifying and Predicting the Variation in Human Signals during Visuo-Linguistic Processes.\n \n \n \n \n\n\n \n Ece Takmaz, Sandro Pezzelle, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the European Chapter of the Association for Computational Linguistics (EACL), 2024. \n To appear\n\n\n\n
\n\n\n\n \n \n \"DescribingPaper\n  \n \n \n \"Describing github\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 7 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{takmaz-etal-eacl-2024,\n    title = "Describing Images \\textit{Fast and Slow}: Quantifying and Predicting the Variation in Human Signals during Visuo-Linguistic Processes",\n    author = "Ece Takmaz and Sandro Pezzelle and Raquel Fern{\\'a}ndez",\n    booktitle = "Proceedings of the European Chapter of the Association for Computational Linguistics (EACL)",\n    year = "2024",\n    url = "https://arxiv.org/abs/2402.01352",\n    abstract = "There is an intricate relation between the properties of an image and how humans behave while describing the image. This behavior shows ample variation, as manifested in human signals such as eye movements and when humans start to describe the image. Despite the value of such signals of visuo-linguistic variation, they are virtually disregarded in the training of current pretrained models, which motivates further investigation. Using a corpus of Dutch image descriptions with concurrently collected eye-tracking data, we explore the nature of the variation in visuo-linguistic signals, and find that they correlate with each other. Given this result, we hypothesize that variation stems partly from the properties of the images, and explore whether image representations encoded by pretrained vision encoders can capture such variation. Our results indicate that pretrained models do so to a weak-to-moderate degree, suggesting that the models lack biases about what makes a stimulus complex for humans and what leads to variations in human outputs.",\n    url_github = "https://github.com/ecekt/visuolinguistic_signal_variation",\n    note = "To appear"\n}\n\n
\n
\n\n\n
\n There is an intricate relation between the properties of an image and how humans behave while describing the image. This behavior shows ample variation, as manifested in human signals such as eye movements and when humans start to describe the image. Despite the value of such signals of visuo-linguistic variation, they are virtually disregarded in the training of current pretrained models, which motivates further investigation. Using a corpus of Dutch image descriptions with concurrently collected eye-tracking data, we explore the nature of the variation in visuo-linguistic signals, and find that they correlate with each other. Given this result, we hypothesize that variation stems partly from the properties of the images, and explore whether image representations encoded by pretrained vision encoders can capture such variation. Our results indicate that pretrained models do so to a weak-to-moderate degree, suggesting that the models lack biases about what makes a stimulus complex for humans and what leads to variations in human outputs.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2023\n \n \n (5)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Information Value: Measuring Utterance Predictability as Distance from Plausible Alternatives.\n \n \n \n \n\n\n \n Mario Giulianelli, Sarenne Wallbridge, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023. \n \n\n\n\n
\n\n\n\n \n \n \"InformationPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 14 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-etal-2023-information,\n    title = "Information Value: Measuring Utterance Predictability as Distance from Plausible Alternatives",\n    author = "Mario Giulianelli and Sarenne Wallbridge and Raquel Fern{\\'a}ndez",\n    booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP)",\n    year = "2023",\n    url ="https://aclanthology.org/2023.emnlp-main.343/",\n    abstract = "We present information value, a measure which quantifies the predictability of an utterance relative to a set of plausible alternatives. We introduce a method to obtain interpretable estimates of information value using neural text generators, and exploit their psychometric predictive power to investigate the dimensions of predictability that drive human comprehension behaviour. Information value is a stronger predictor of utterance acceptability in written and spoken dialogue than aggregates of token-level surprisal and it is complementary to surprisal for predicting eye-tracked reading times."\n}\n\n
\n
\n\n\n
\n We present information value, a measure which quantifies the predictability of an utterance relative to a set of plausible alternatives. We introduce a method to obtain interpretable estimates of information value using neural text generators, and exploit their psychometric predictive power to investigate the dimensions of predictability that drive human comprehension behaviour. Information value is a stronger predictor of utterance acceptability in written and spoken dialogue than aggregates of token-level surprisal and it is complementary to surprisal for predicting eye-tracked reading times.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n What Comes Next? Evaluating Uncertainty in Neural Text Generators Against Human Production Variability.\n \n \n \n \n\n\n \n Mario Giulianelli, Joris Baan, Wilker Aziz, Raquel Fernández, & Barbara Plank.\n\n\n \n\n\n\n In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023. \n \n\n\n\n
\n\n\n\n \n \n \"WhatPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 11 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-etal-2023-arxiv,\n    title = "What Comes Next? Evaluating Uncertainty in Neural Text Generators Against Human Production Variability",\n    author = " Giulianelli, Mario and Baan, Joris and Aziz, Wilker and Fern{\\'a}ndez, Raquel and Plank, Barbara",\n    booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP)",\n    year = "2023",\n    url = "https://aclanthology.org/2023.emnlp-main.887/",\n    abstract = "In Natural Language Generation (NLG) tasks, for any input, multiple communicative goals are plausible, and any goal can be put into words, or produced, in multiple ways. We characterise the extent to which human production varies lexically, syntactically, and semantically across four NLG tasks, connecting human production variability to aleatoric or data uncertainty. We then inspect the space of output strings shaped by a generation system{'}s predicted probability distribution and decoding algorithm to probe its uncertainty. For each test input, we measure the generator{'}s calibration to human production variability. Following this instance-level approach, we analyse NLG models and decoding strategies, demonstrating that probing a generator with multiple samples and, when possible, multiple references, provides the level of detail necessary to gain understanding of a model{'}s representation of uncertainty."\n}\n\n\n
\n
\n\n\n
\n In Natural Language Generation (NLG) tasks, for any input, multiple communicative goals are plausible, and any goal can be put into words, or produced, in multiple ways. We characterise the extent to which human production varies lexically, syntactically, and semantically across four NLG tasks, connecting human production variability to aleatoric or data uncertainty. We then inspect the space of output strings shaped by a generation system's predicted probability distribution and decoding algorithm to probe its uncertainty. For each test input, we measure the generator's calibration to human production variability. Following this instance-level approach, we analyse NLG models and decoding strategies, demonstrating that probing a generator with multiple samples and, when possible, multiple references, provides the level of detail necessary to gain understanding of a model's representation of uncertainty.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Interpretable Word Sense Representations via Definition Generation: The Case of Semantic Change Analysis.\n \n \n \n \n\n\n \n Mario Giulianelli, Iris Luden, Raquel Fernández, & Andrey Kutuzov.\n\n\n \n\n\n\n In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3130–3148, Toronto, Canada, 2023. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"InterpretablePaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 12 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-etal-2023,\n    title = "Interpretable Word Sense Representations via Definition Generation: The Case of Semantic Change Analysis",\n    author = "Giulianelli, Mario  and\n      Luden, Iris  and\n      Fern{\\'a}ndez, Raquel  and\n      Kutuzov, Andrey",\n    booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n    year = "2023",\n    address = "Toronto, Canada",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2023.acl-long.176",\n    pages = "3130--3148",\n    abstract = "We propose using automatically generated natural language definitions of contextualised word usages as interpretable word and word sense representations.Given a collection of usage examples for a target word, and the corresponding data-driven usage clusters (i.e., word senses), a definition is generated for each usage with a specialised Flan-T5 language model, and the most prototypical definition in a usage cluster is chosen as the sense label. We demonstrate how the resulting sense labels can make existing approaches to semantic change analysis more interpretable, and how they can allow users {---} historical linguists, lexicographers, or social scientists {---} to explore and intuitively explain diachronic trajectories of word meaning. Semantic change analysis is only one of many possible applications of the {`}definitions as representations{'} paradigm. Beyond being human-readable, contextualised definitions also outperform token or usage sentence embeddings in word-in-context semantic similarity judgements, making them a new promising type of lexical representation for NLP.",\n}\n\n
\n
\n\n\n
\n We propose using automatically generated natural language definitions of contextualised word usages as interpretable word and word sense representations.Given a collection of usage examples for a target word, and the corresponding data-driven usage clusters (i.e., word senses), a definition is generated for each usage with a specialised Flan-T5 language model, and the most prototypical definition in a usage cluster is chosen as the sense label. We demonstrate how the resulting sense labels can make existing approaches to semantic change analysis more interpretable, and how they can allow users — historical linguists, lexicographers, or social scientists — to explore and intuitively explain diachronic trajectories of word meaning. Semantic change analysis is only one of many possible applications of the `definitions as representations' paradigm. Beyond being human-readable, contextualised definitions also outperform token or usage sentence embeddings in word-in-context semantic similarity judgements, making them a new promising type of lexical representation for NLP.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Speaking the Language of Your Listener: Audience-Aware Adaptation via Plug-and-Play Theory of Mind.\n \n \n \n \n\n\n \n Ece Takmaz, Nicolo' Brandizzi, Mario Giulianelli, Sandro Pezzelle, & Raquel Fernández.\n\n\n \n\n\n\n In Findings of the Association for Computational Linguistics: ACL 2023, pages 4198–4217, Toronto, Canada, 2023. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"SpeakingPaper\n  \n \n \n \"Speaking github\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 10 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{takmaz-etal-2023-speaking,\n    title = "Speaking the Language of Your Listener: Audience-Aware Adaptation via Plug-and-Play Theory of Mind",\n    author = "Takmaz, Ece  and\n      Brandizzi, Nicolo{'}  and\n      Giulianelli, Mario  and\n      Pezzelle, Sandro  and\n      Fern{\\'a}ndez, Raquel",\n    booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",\n    year = "2023",\n    address = "Toronto, Canada",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2023.findings-acl.258",\n    url_github = "https://github.com/nicofirst1/speaker-adaptation",\n    pages = "4198--4217",\n    abstract = "Dialogue participants may have varying levels of knowledge about the topic under discussion. In such cases, it is essential for speakers to adapt their utterances by taking their audience into account. Yet, it is an open question how such adaptation can be modelled in computational agents. In this paper, we model a visually grounded referential game between a knowledgeable speaker and a listener with more limited visual and linguistic experience. Inspired by psycholinguistic theories, we endow our speaker with the ability to adapt its referring expressions via a simulation module that monitors the effectiveness of planned utterances from the listener{'}s perspective. We propose an adaptation mechanism building on plug-and-play approaches to controlled language generation, where utterance generation is steered on the fly by the simulator without finetuning the speaker{'}s underlying language model. Our results and analyses show that our approach is effective: the speaker{'}s utterances become closer to the listener{'}s domain of expertise, which leads to higher communicative success.",\n}\n\n
\n
\n\n\n
\n Dialogue participants may have varying levels of knowledge about the topic under discussion. In such cases, it is essential for speakers to adapt their utterances by taking their audience into account. Yet, it is an open question how such adaptation can be modelled in computational agents. In this paper, we model a visually grounded referential game between a knowledgeable speaker and a listener with more limited visual and linguistic experience. Inspired by psycholinguistic theories, we endow our speaker with the ability to adapt its referring expressions via a simulation module that monitors the effectiveness of planned utterances from the listener's perspective. We propose an adaptation mechanism building on plug-and-play approaches to controlled language generation, where utterance generation is steered on the fly by the simulator without finetuning the speaker's underlying language model. Our results and analyses show that our approach is effective: the speaker's utterances become closer to the listener's domain of expertise, which leads to higher communicative success.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Semantic Adaptation to the Interpretation of Gradable Adjectives via Active Linguistic Interaction.\n \n \n \n \n\n\n \n Sandro Pezzelle, & Raquel Fernández.\n\n\n \n\n\n\n Cognitive Science, 47(2): e13248. 2023.\n \n\n\n\n
\n\n\n\n \n \n \"SemanticPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 14 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n \n \n \n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{Pezzelle-Fernandez-2023-cogsci,\nauthor = {Pezzelle, Sandro and Fern{\\'a}ndez, Raquel},\ntitle = {Semantic Adaptation to the Interpretation of Gradable Adjectives via Active Linguistic Interaction},\njournal = {Cognitive Science},\nvolume = {47},\nnumber = {2},\npages = {e13248},\nkeywords = {Gradable adjectives, Semantic adaptation, Vagueness, Questions, Linguistic interaction, Visual grounding},\ndoi = {https://doi.org/10.1111/cogs.13248},\nurl = {https://onlinelibrary.wiley.com/doi/abs/10.1111/cogs.13248},\neprint = {https://onlinelibrary.wiley.com/doi/pdf/10.1111/cogs.13248},\nabstract = {Abstract When communicating, people adapt their linguistic representations to those of their interlocutors. Previous studies have shown that this also occurs at the semantic level for vague and context-dependent terms such as quantifiers and uncertainty expressions. However, work to date has mostly focused on passive exposure to a given speaker's interpretation, without considering the possible role of active linguistic interaction. In this study, we focus on gradable adjectives big and small and develop a novel experimental paradigm that allows participants to ask clarification questions to figure out their interlocutor's interpretation. We find that, when in doubt, speakers do resort to this strategy, despite its inherent cognitive cost, and that doing so results in higher semantic alignment measured in terms of communicative success. While not all question–answer pairs are equally informative, we show that speakers become better questioners as the interaction progresses. Yet, the higher semantic alignment observed when speakers are able to ask questions does not increase over time. This suggests that conversational interaction's key advantage may be to boost coordination without committing to long-term semantic updates. Our findings shed new light on the mechanisms used by speakers to achieve semantic alignment and on how language is shaped by communication.},\nyear = {2023}\n}\n\n
\n
\n\n\n
\n Abstract When communicating, people adapt their linguistic representations to those of their interlocutors. Previous studies have shown that this also occurs at the semantic level for vague and context-dependent terms such as quantifiers and uncertainty expressions. However, work to date has mostly focused on passive exposure to a given speaker's interpretation, without considering the possible role of active linguistic interaction. In this study, we focus on gradable adjectives big and small and develop a novel experimental paradigm that allows participants to ask clarification questions to figure out their interlocutor's interpretation. We find that, when in doubt, speakers do resort to this strategy, despite its inherent cognitive cost, and that doing so results in higher semantic alignment measured in terms of communicative success. While not all question–answer pairs are equally informative, we show that speakers become better questioners as the interaction progresses. Yet, the higher semantic alignment observed when speakers are able to ask questions does not increase over time. This suggests that conversational interaction's key advantage may be to boost coordination without committing to long-term semantic updates. Our findings shed new light on the mechanisms used by speakers to achieve semantic alignment and on how language is shaped by communication.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2022\n \n \n (10)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Controllable Text Generation for All Ages: Evaluating a Plug-and-Play Approach to Age-Adapted Dialogue.\n \n \n \n \n\n\n \n Lennert Jansen, Štěpán Lars Laichter, Arabella Sinclair, Margot Goot, Raquel Fernández, & Sandro Pezzelle.\n\n\n \n\n\n\n In Proceedings of the 2nd Workshop on Natural Language Generation, Evaluation, and Metrics (GEM), pages 172–188, Abu Dhabi, United Arab Emirates (Hybrid), 2022. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"ControllablePaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 9 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{jansen-etal-2022-controllable,\n    title = "Controllable Text Generation for All Ages: Evaluating a Plug-and-Play Approach to Age-Adapted Dialogue",\n    author = "Jansen, Lennert  and\n      Laichter, {\\v{S}}t{\\v{e}}p{\\'a}n Lars  and\n      Sinclair, Arabella  and\n      van der Goot, Margot  and\n      Fern{\\'a}ndez, Raquel  and\n      Pezzelle, Sandro",\n    booktitle = "Proceedings of the 2nd Workshop on Natural Language Generation, Evaluation, and Metrics (GEM)",\n    year = "2022",\n    address = "Abu Dhabi, United Arab Emirates (Hybrid)",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2022.gem-1.14",\n    pages = "172--188",\n    abstract = "To be trusted and perceived as natural and coherent, conversational systems must adapt to the language of their users. While personalized dialogue is a promising direction, controlling generation for fine-grained language features remains a challenge in this approach. A recent line of research showed the effectiveness of leveraging pre-trained language models toward adapting to a text{'}s topic or sentiment. In this study, we build on these approaches and focus on a higher-level dimension of language variation: speakers{'} age. We frame the task as a dialogue response generation, and test methods based on bag-of-words (BoW) and neural discriminators (Disc) to condition the output of GPT-2 and DialoGPT without altering the parameters of the language models. We show that Disc models achieve a higher degree of detectable control than BoW models based on automatic evaluation. In contrast, humans can partially detect age differences in BoW but not Disc responses. Since BoW responses are deemed better than Disc ones by humans, simple controllable methods thus appear to be a better tradeoff between adaptation and language quality. Our work confirms the challenges of adapting to higher-level dimensions of language variation. Moreover, it highlights the need to evaluate natural language generation thoroughly.",\n}\n\n
\n
\n\n\n
\n To be trusted and perceived as natural and coherent, conversational systems must adapt to the language of their users. While personalized dialogue is a promising direction, controlling generation for fine-grained language features remains a challenge in this approach. A recent line of research showed the effectiveness of leveraging pre-trained language models toward adapting to a text's topic or sentiment. In this study, we build on these approaches and focus on a higher-level dimension of language variation: speakers' age. We frame the task as a dialogue response generation, and test methods based on bag-of-words (BoW) and neural discriminators (Disc) to condition the output of GPT-2 and DialoGPT without altering the parameters of the language models. We show that Disc models achieve a higher degree of detectable control than BoW models based on automatic evaluation. In contrast, humans can partially detect age differences in BoW but not Disc responses. Since BoW responses are deemed better than Disc ones by humans, simple controllable methods thus appear to be a better tradeoff between adaptation and language quality. Our work confirms the challenges of adapting to higher-level dimensions of language variation. Moreover, it highlights the need to evaluate natural language generation thoroughly.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Stop Measuring Calibration When Humans Disagree.\n \n \n \n \n\n\n \n Joris Baan, Wilker Aziz, Barbara Plank, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 1892–1915, Abu Dhabi, United Arab Emirates, 2022. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"StopPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 12 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{baan-etal-2022-emnlp,\n    title = "Stop Measuring Calibration When Humans Disagree",\n    author = "Baan, Joris  and\n      Aziz, Wilker  and\n      Plank, Barbara  and\n      Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",\n    year = "2022",\n    address = "Abu Dhabi, United Arab Emirates",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2022.emnlp-main.124",\n    pages = "1892--1915",\n    abstract = "Calibration is a popular framework to evaluate whether a classifier knows when it does not know - i.e., its predictive probabilities are a good indication of how likely a prediction is to be correct. Correctness is commonly estimated against the human majority class. Recently, calibration to human majority has been measured on tasks where humans inherently disagree about which class applies. We show that measuring calibration to human majority given inherent disagreements is theoretically problematic, demonstrate this empirically on the ChaosNLI dataset, and derive several instance-level measures of calibration that capture key statistical properties of human judgements - including class frequency, ranking and entropy.",\n}\n\n
\n
\n\n\n
\n Calibration is a popular framework to evaluate whether a classifier knows when it does not know - i.e., its predictive probabilities are a good indication of how likely a prediction is to be correct. Correctness is commonly estimated against the human majority class. Recently, calibration to human majority has been measured on tasks where humans inherently disagree about which class applies. We show that measuring calibration to human majority given inherent disagreements is theoretically problematic, demonstrate this empirically on the ChaosNLI dataset, and derive several instance-level measures of calibration that capture key statistical properties of human judgements - including class frequency, ranking and entropy.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Towards Pragmatic Production Strategies for Natural Language Generation Tasks.\n \n \n \n \n\n\n \n Mario Giulianelli.\n\n\n \n\n\n\n In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 7978–7984, Abu Dhabi, United Arab Emirates, 2022. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"TowardsPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 7 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-2022-emnlp,\n    title = "Towards Pragmatic Production Strategies for Natural Language Generation Tasks",\n    author = "Giulianelli, Mario",\n    booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",\n    year = "2022",\n    address = "Abu Dhabi, United Arab Emirates",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2022.emnlp-main.544",\n    pages = "7978--7984",\n    abstract = "This position paper proposes a conceptual framework for the design of Natural Language Generation (NLG) systems that follow efficient and effective production strategies in order to achieve complex communicative goals. In this general framework, efficiency is characterised as the parsimonious regulation of production and comprehension costs while effectiveness is measured with respect to task-oriented and contextually grounded communicative goals. We provide concrete suggestions for the estimation of goals, costs, and utility via modern statistical methods, demonstrating applications of our framework to the classic pragmatic task of visually grounded referential games and to abstractive text summarisation, two popular generation tasks with real-world applications. In sum, we advocate for the development of NLG systems that learn to make pragmatic production decisions from experience, by reasoning about goals, costs, and utility in a human-like way.",\n}\n\n\n
\n
\n\n\n
\n This position paper proposes a conceptual framework for the design of Natural Language Generation (NLG) systems that follow efficient and effective production strategies in order to achieve complex communicative goals. In this general framework, efficiency is characterised as the parsimonious regulation of production and comprehension costs while effectiveness is measured with respect to task-oriented and contextually grounded communicative goals. We provide concrete suggestions for the estimation of goals, costs, and utility via modern statistical methods, demonstrating applications of our framework to the classic pragmatic task of visually grounded referential games and to abstractive text summarisation, two popular generation tasks with real-world applications. In sum, we advocate for the development of NLG systems that learn to make pragmatic production decisions from experience, by reasoning about goals, costs, and utility in a human-like way.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Construction Repetition Reduces Information Rate in Dialogue.\n \n \n \n \n\n\n \n Mario Giulianelli, Arabella Sinclair, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing, pages 665–682, November 2022. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"ConstructionPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 7 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-etal-2022-construction,\n    title = "Construction Repetition Reduces Information Rate in Dialogue",\n    author = "Giulianelli, Mario  and\n      Sinclair, Arabella  and\n      Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing",\n    month = nov,\n    year = "2022",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2022.aacl-main.51",\n    pages = "665--682",\n    abstract = "Speakers repeat constructions frequently in dialogue. Due to their peculiar information-theoretic properties, repetitions can be thought of as a strategy for cost-effective communication. In this study, we focus on the repetition of lexicalised constructions{---}i.e., recurring multi-word units{---}in English open-domain spoken dialogues. We hypothesise that speakers use construction repetition to mitigate information rate, leading to an overall decrease in utterance information content over the course of a dialogue. We conduct a quantitative analysis, measuring the information content of constructions and that of their containing utterances, estimating information content with an adaptive neural language model. We observe that construction usage lowers the information content of utterances. This facilitating effect (i) increases throughout dialogues, (ii) is boosted by repetition, (iii) grows as a function of repetition frequency and density, and (iv) is stronger for repetitions of referential constructions.",\n}\n\n
\n
\n\n\n
\n Speakers repeat constructions frequently in dialogue. Due to their peculiar information-theoretic properties, repetitions can be thought of as a strategy for cost-effective communication. In this study, we focus on the repetition of lexicalised constructions—i.e., recurring multi-word units—in English open-domain spoken dialogues. We hypothesise that speakers use construction repetition to mitigate information rate, leading to an overall decrease in utterance information content over the course of a dialogue. We conduct a quantitative analysis, measuring the information content of constructions and that of their containing utterances, estimating information content with an adaptive neural language model. We observe that construction usage lowers the information content of utterances. This facilitating effect (i) increases throughout dialogues, (ii) is boosted by repetition, (iii) grows as a function of repetition frequency and density, and (iv) is stronger for repetitions of referential constructions.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Alignment of code switching varies with proficiency in second language learning dialogue.\n \n \n \n \n\n\n \n Arabella J. Sinclair, & Raquel Fernández.\n\n\n \n\n\n\n System. Special Issue on Linguistic alignment in Second Language Acquisition: occurrences, learning effects, and beyond. 2022.\n \n\n\n\n
\n\n\n\n \n \n \"AlignmentPaper\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@article{sinclair-fernandez-system-2022,\ntitle = {Alignment of code switching varies with proficiency in second language learning dialogue},\njournal = {System. Special Issue on Linguistic alignment in Second Language Acquisition: occurrences, learning effects, and beyond},\nyear = {2022},\nissn = {0346-251X},\npublisher = {Elsevier},\ndoi = {https://doi.org/10.1016/j.system.2022.102952},\nurl = {https://www.sciencedirect.com/science/article/pii/S0346251X22002342},\nauthor = {Arabella J. Sinclair and Raquel Fern\\'andez},\nabstract = {Speakers in dialogue tend to adopt the language patterns of the other, aligning their language to their interlocutor. This can happen at many levels of communication, including the tendency to code switch (CS), or change to another language. Alignment has often been considered the result of an unconscious automatic process that facilitates speakers' mutual understanding. In dialogues with a second language (L2) learner, alignment is constrained by the proficiency of the learner, and additional non-automatic processes will be at play, namely the individual pedagogical goals of learner and tutor. In this study, we investigate alignment in dialogues between Spanish/Catalan learners of English and their tutors. We analyse CS incidence, whether code switching can be explained as automatic alignment between speakers, and whether this is independent of other, non-automatic factors related to speakers’ goals. We find that alignment of code switching is present, varies with learner proficiency, and that code switching can additionally be triggered by lexical overlap and turn taking asymmetry, which we attribute to conscious pedagogical choices on the part of both tutor, at lower levels, and learner, at higher levels of student proficiency.}\n}\n\n
\n
\n\n\n
\n Speakers in dialogue tend to adopt the language patterns of the other, aligning their language to their interlocutor. This can happen at many levels of communication, including the tendency to code switch (CS), or change to another language. Alignment has often been considered the result of an unconscious automatic process that facilitates speakers' mutual understanding. In dialogues with a second language (L2) learner, alignment is constrained by the proficiency of the learner, and additional non-automatic processes will be at play, namely the individual pedagogical goals of learner and tutor. In this study, we investigate alignment in dialogues between Spanish/Catalan learners of English and their tutors. We analyse CS incidence, whether code switching can be explained as automatic alignment between speakers, and whether this is independent of other, non-automatic factors related to speakers’ goals. We find that alignment of code switching is present, varies with learner proficiency, and that code switching can additionally be triggered by lexical overlap and turn taking asymmetry, which we attribute to conscious pedagogical choices on the part of both tutor, at lower levels, and learner, at higher levels of student proficiency.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Structural Persistence in Language Models: Priming as a Window into Abstract Language Representations.\n \n \n \n \n\n\n \n Arabella Sinclair, Jaap Jumelet, Willem Zuidema, & Raquel Fernández.\n\n\n \n\n\n\n Transactions of the Association for Computational Linguistics (TACL). 2022.\n \n\n\n\n
\n\n\n\n \n \n \"StructuralPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 13 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@Article{sinclair-etal-2022-tacl,\n  author = \t "Arabella Sinclair and Jaap Jumelet and Willem Zuidema and Raquel Fern\\'andez",\n  title = \t "Structural Persistence in Language Models: Priming as a Window into Abstract Language Representations",\n  journal = \t "Transactions of the Association for Computational Linguistics (TACL)",\n  year = "2022",\n  url = "https://direct.mit.edu/tacl/article/doi/10.1162/tacl_a_00504/113019/Structural-Persistence-in-Language-Models-Priming",\n  abstract = "We investigate the extent to which modern, neural language models are susceptible to structural priming,\n  the phenomenon whereby the structure of a sentence makes the same structure more probable in a follow-up  sentence.\n  We explore how priming can be used to study the potential of these models to learn abstract structural information,\n  which is a prerequisite for good performance on tasks that require natural language understanding skills. We introduce\n  a novel metric and release PRIME-LM, a large corpus where we control for various linguistic factors which interact with priming strength.\n  We find that Transformer models indeed show evidence of structural priming, but also that the generalisations they learned are to\n  some extent modulated by semantic information. Our experiments also show that the representations acquired by the models\n  may not only encode abstract sequential structure but involve certain level of hierarchical syntactic information.\n  More generally, our study shows that the priming paradigm is a useful, additional tool for gaining insights into the capacities\n  of language models and opens the door to future priming-based investigations that probe the model's internal states."\n}\n\n
\n
\n\n\n
\n We investigate the extent to which modern, neural language models are susceptible to structural priming, the phenomenon whereby the structure of a sentence makes the same structure more probable in a follow-up sentence. We explore how priming can be used to study the potential of these models to learn abstract structural information, which is a prerequisite for good performance on tasks that require natural language understanding skills. We introduce a novel metric and release PRIME-LM, a large corpus where we control for various linguistic factors which interact with priming strength. We find that Transformer models indeed show evidence of structural priming, but also that the generalisations they learned are to some extent modulated by semantic information. Our experiments also show that the representations acquired by the models may not only encode abstract sequential structure but involve certain level of hierarchical syntactic information. More generally, our study shows that the priming paradigm is a useful, additional tool for gaining insights into the capacities of language models and opens the door to future priming-based investigations that probe the model's internal states.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n AnaLog: Testing Analytical and Deductive Logic Learnability in Language Models.\n \n \n \n \n\n\n \n Samuel Ryb, Mario Giulianelli, Arabella Sinclair, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 11th Joint Conference on Lexical and Computational Semantics, pages 55–68, Seattle, Washington, July 2022. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"AnaLog:Paper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 26 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{ryb-etal-2022-analog,\n    title = "{A}na{L}og: Testing Analytical and Deductive Logic Learnability in Language Models",\n    author = "Ryb, Samuel  and\n      Giulianelli, Mario  and\n      Sinclair, Arabella  and\n      Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 11th Joint Conference on Lexical and Computational Semantics",\n    month = jul,\n    year = "2022",\n    address = "Seattle, Washington",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2022.starsem-1.5",\n    pages = "55--68",\n    abstract = "We investigate the extent to which pre-trained language models acquire analytical and deductive logical reasoning\n    capabilities as a side effect of learning word prediction. We present AnaLog, a natural language inference task designed to\n    probe models for these capabilities, controlling for different invalid heuristics the models may adopt instead of learning the\n    desired generalisations. We test four languagemodels on AnaLog, finding that they have all learned, to a different extent,\n    to encode information that is predictive of entailment beyond shallow heuristics such as lexical overlap and grammaticality.\n    We closely analyse the best performing language model and show that while it performs more consistently than other language\n    models across logical connectives and reasoning domains, it still is sensitive to lexical and syntactic variations in the\n    realisation of logical statements."\n}\n\n
\n
\n\n\n
\n We investigate the extent to which pre-trained language models acquire analytical and deductive logical reasoning capabilities as a side effect of learning word prediction. We present AnaLog, a natural language inference task designed to probe models for these capabilities, controlling for different invalid heuristics the models may adopt instead of learning the desired generalisations. We test four languagemodels on AnaLog, finding that they have all learned, to a different extent, to encode information that is predictive of entailment beyond shallow heuristics such as lexical overlap and grammaticality. We closely analyse the best performing language model and show that while it performs more consistently than other language models across logical connectives and reasoning domains, it still is sensitive to lexical and syntactic variations in the realisation of logical statements.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Less Descriptive yet Discriminative: Quantifying the Properties of Multimodal Referring Utterances via CLIP.\n \n \n \n \n\n\n \n Ece Takmaz, Sandro Pezzelle, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics, pages 36–42, Dublin, Ireland, May 2022. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"LessPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 9 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{takmaz-etal-2022-cmcl,\n    title = "Less Descriptive yet Discriminative: Quantifying the Properties of Multimodal Referring Utterances via {CLIP}",\n    author = "Takmaz, Ece  and Pezzelle, Sandro  and Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics",\n    month = may,\n    year = "2022",\n    address = "Dublin, Ireland",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2022.cmcl-1.4",\n    pages = "36--42",\n    abstract = "In this work, we use a transformer-based pre-trained multimodal model, CLIP,\n    to shed light on the mechanisms employed by human speakers when referring to visual entities.\n    In particular, we use CLIP to quantify the degree of descriptiveness (how well an utterance\n    describes an image in isolation) and discriminativeness (to what extent an utterance is effective\n    in picking out a single image among similar images) of human referring utterances within multimodal\n    dialogues. Overall, our results show that utterances become less descriptive over time while their\n    discriminativeness remains unchanged. Through analysis, we propose that this trend could be due to\n    participants relying on the previous mentions in the dialogue history, as well as being able to\n    distill the most discriminative information from the visual context. In general, our study opens\n    up the possibility of using this and similar models to quantify patterns in human data and shed\n    light on the underlying cognitive mechanisms."\n}\n\n
\n
\n\n\n
\n In this work, we use a transformer-based pre-trained multimodal model, CLIP, to shed light on the mechanisms employed by human speakers when referring to visual entities. In particular, we use CLIP to quantify the degree of descriptiveness (how well an utterance describes an image in isolation) and discriminativeness (to what extent an utterance is effective in picking out a single image among similar images) of human referring utterances within multimodal dialogues. Overall, our results show that utterances become less descriptive over time while their discriminativeness remains unchanged. Through analysis, we propose that this trend could be due to participants relying on the previous mentions in the dialogue history, as well as being able to distill the most discriminative information from the visual context. In general, our study opens up the possibility of using this and similar models to quantify patterns in human data and shed light on the underlying cognitive mechanisms.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Team DMG at CMCL 2022 Shared Task: Transformer Adapters for the Multi- and Cross-Lingual Prediction of Human Reading Behavior.\n \n \n \n \n\n\n \n Ece Takmaz.\n\n\n \n\n\n\n In Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics, pages 136–144, Dublin, Ireland, May 2022. Association for Computational Linguistics\n \n\nBest Shared Task Paper Award\n\n
\n\n\n\n \n \n \"TeamPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{takmaz-2022-team,\n    title = "Team {DMG} at {CMCL} 2022 Shared Task: Transformer Adapters for the Multi- and Cross-Lingual Prediction of Human Reading Behavior",\n    author = "Takmaz, Ece",\n    booktitle = "Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics",\n    month = may,\n    year = "2022",\n    bibbase_note = {Best Shared Task Paper Award},\n    address = "Dublin, Ireland",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2022.cmcl-1.16",\n    pages = "136--144",\n    abstract = "In this paper, we present the details of our approaches that attained the second place\n    in the shared task of the ACL 2022 Cognitive Modeling and Computational Linguistics Workshop.\n    The shared task is focused on multi- and cross-lingual prediction of eye movement features in\n    human reading behavior, which could provide valuable information regarding language processing.\n    To this end, we train {`}adapters{'} inserted into the layers of frozen transformer-based pretrained\n    language models. We find that multilingual models equipped with adapters perform well in predicting\n    eye-tracking features. Our results suggest that utilizing language- and task-specific adapters is\n    beneficial and translating test sets into similar languages that exist in the training set could\n    help with zero-shot transferability in the prediction of human reading behavior."\n}\n\n
\n
\n\n\n
\n In this paper, we present the details of our approaches that attained the second place in the shared task of the ACL 2022 Cognitive Modeling and Computational Linguistics Workshop. The shared task is focused on multi- and cross-lingual prediction of eye movement features in human reading behavior, which could provide valuable information regarding language processing. To this end, we train `adapters' inserted into the layers of frozen transformer-based pretrained language models. We find that multilingual models equipped with adapters perform well in predicting eye-tracking features. Our results suggest that utilizing language- and task-specific adapters is beneficial and translating test sets into similar languages that exist in the training set could help with zero-shot transferability in the prediction of human reading behavior.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Do Not Fire the Linguist: Grammatical Profiles Help Language Models Detect Semantic Change.\n \n \n \n \n\n\n \n Mario Giulianelli, Andrey Kutuzov, & Lidia Pivovarova.\n\n\n \n\n\n\n In Proceedings of the 3rd Workshop on Computational Approaches to Historical Language Change, pages 54–67, Dublin, Ireland, May 2022. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"DoPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-etal-2022-fire,\n    title = "Do Not Fire the Linguist: Grammatical Profiles Help Language Models Detect Semantic Change",\n    author = "Giulianelli, Mario  and Kutuzov, Andrey  and Pivovarova, Lidia",\n    booktitle = "Proceedings of the 3rd Workshop on Computational Approaches to Historical Language Change",\n    month = may,\n    year = "2022",\n    address = "Dublin, Ireland",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2022.lchange-1.6",\n    pages = "54--67",\n    abstract = "Morphological and syntactic changes in word usage {---} as captured, e.g., by\n    grammatical profiles {---} have been shown to be good predictors of a word{'}s meaning change.\n    In this work, we explore whether large pre-trained contextualised language models, a common tool\n    for lexical semantic change detection, are sensitive to such morphosyntactic changes. To this end,\n    we first compare the performance of grammatical profiles against that of a multilingual neural\n    language model (XLM-R) on 10 datasets, covering 7 languages, and then combine the two approaches\n    in ensembles to assess their complementarity. Our results show that ensembling grammatical\n    profiles with XLM-R improves semantic change detection performance for most datasets and languages.\n    This indicates that language models do not fully cover the fine-grained morphological and syntactic\n    signals that are explicitly represented in grammatical profiles. An interesting exception are the\n    test sets where the time spans under analysis are much longer than the time gap between them\n    (for example, century-long spans with a one-year gap between them). Morphosyntactic change is\n    slow so grammatical profiles do not detect in such cases. In contrast, language models, thanks\n    to their access to lexical information, are able to detect fast topical changes."\n}\n\n\n
\n
\n\n\n
\n Morphological and syntactic changes in word usage — as captured, e.g., by grammatical profiles — have been shown to be good predictors of a word's meaning change. In this work, we explore whether large pre-trained contextualised language models, a common tool for lexical semantic change detection, are sensitive to such morphosyntactic changes. To this end, we first compare the performance of grammatical profiles against that of a multilingual neural language model (XLM-R) on 10 datasets, covering 7 languages, and then combine the two approaches in ensembles to assess their complementarity. Our results show that ensembling grammatical profiles with XLM-R improves semantic change detection performance for most datasets and languages. This indicates that language models do not fully cover the fine-grained morphological and syntactic signals that are explicitly represented in grammatical profiles. An interesting exception are the test sets where the time spans under analysis are much longer than the time gap between them (for example, century-long spans with a one-year gap between them). Morphosyntactic change is slow so grammatical profiles do not detect in such cases. In contrast, language models, thanks to their access to lexical information, are able to detect fast topical changes.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2021\n \n \n (8)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Word Representation Learning in Multimodal Pre-Trained Transformers: An Intrinsic Evaluation.\n \n \n \n \n\n\n \n Sandro Pezzelle, Ece Takmaz, & Raquel Fernández.\n\n\n \n\n\n\n Transactions of the Association for Computational Linguistics (TACL). 2021.\n \n\n\n\n
\n\n\n\n \n \n \"WordPaper\n  \n \n \n \"Word github\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 52 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@Article{pezzelle-etal-2021-tacl,\n  author = \t "Sandro Pezzelle and Ece Takmaz and Raquel Fern\\'andez",\n  title = \t "Word Representation Learning in Multimodal Pre-Trained Transformers: An Intrinsic Evaluation",\n  journal = \t "Transactions of the Association for Computational Linguistics (TACL)",\n  year = "2021",\n  url = "https://direct.mit.edu/tacl/article-pdf/doi/10.1162/tacl_a_00443/1979754/tacl_a_00443.pdf",\n  url_github = "https://github.com/sandropezzelle/multimodal-evaluation",\n  abstract = "This study carries out a systematic /intrinsic/ evaluation of the semantic representations\n\tlearned by state-of-the-art pre-trained multimodal Transformers.\n\tThese representations are claimed to be task-agnostic and shown to help on many downstream language-and-vision tasks.\n\tHowever, the extent to which they align with human semantic intuitions remains unclear.\n\tWe experiment with various models and obtain /static/ word representations from the /contextualized/ ones they learn.\n\tWe then evaluate them against the semantic judgements provided by human speakers. In line with previous evidence,\n\twe observe a generalized advantage of multimodal representations over language-only ones on concrete word pairs,\n  but not on abstract ones. On the one hand, this confirms the effectiveness of these models to\n\talign language and vision, which results in better semantic representations for concepts that are /grounded/ in images.\n\tOn the other hand, models are shown to follow different representation learning patterns, which\n\tsheds some light on /how/ and /when/ they perform multimodal integration."\n}\n\n\n
\n
\n\n\n
\n This study carries out a systematic /intrinsic/ evaluation of the semantic representations learned by state-of-the-art pre-trained multimodal Transformers. These representations are claimed to be task-agnostic and shown to help on many downstream language-and-vision tasks. However, the extent to which they align with human semantic intuitions remains unclear. We experiment with various models and obtain /static/ word representations from the /contextualized/ ones they learn. We then evaluate them against the semantic judgements provided by human speakers. In line with previous evidence, we observe a generalized advantage of multimodal representations over language-only ones on concrete word pairs, but not on abstract ones. On the one hand, this confirms the effectiveness of these models to align language and vision, which results in better semantic representations for concepts that are /grounded/ in images. On the other hand, models are shown to follow different representation learning patterns, which sheds some light on /how/ and /when/ they perform multimodal integration.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Detecting Age-Related Linguistic Patterns in Dialogue: Toward Adaptive Conversational Systems.\n \n \n \n \n\n\n \n Lennert Jansen, Arabella Sinclair, Margot J. Goot, Raquel Fernández, & Sandro Pezzelle.\n\n\n \n\n\n\n In Proceedings of the Eighth Italian Conference on Computational Linguistics (CLiC-it), 2021. \n \n\n\n\n
\n\n\n\n \n \n \"DetectingPaper\n  \n \n \n \"Detecting github\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 10 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{jansen-etal-2021-detecting,\n    title = "Detecting Age-Related Linguistic Patterns in Dialogue: Toward Adaptive Conversational Systems",\n    author = "Jansen, Lennert and Sinclair, Arabella and van der Goot, Margot J. and Fern{\\'a}ndez, Raquel and Pezzelle, Sandro",\n    booktitle = "Proceedings of the Eighth Italian Conference on Computational Linguistics (CLiC-it)",\n    year = "2021",\n    url = "http://ceur-ws.org/Vol-3033/paper47.pdf",\n    url_github = "https://github.com/ lennertjansen/detecting-age-in-dialogue",\n    abstract = "This work explores an important dimension of variation in the language used by dialogue\n    participants: their age. While previous work showed differences at various linguistic levels between\n    age groups when experimenting with written discourse data (e.g., blog posts), previous work on dialogue\n    has largely been limited to acoustic information related to voice and prosody. Detecting fine-grained\n    linguistic properties of human dialogues is of crucial importance for developing AI- based conversational\n    systems which are able to adapt to their human interlocutors. We therefore investigate whether, and to\n    what extent, current text-based NLP models can detect such linguistic differences, and what the features\n    driving their predictions are. We show that models achieve a fairly good performance on age- group prediction,\n    though the task appears to be more challenging compared to discourse. Through in-depth analysis of the\n    best models’ errors and the most predictive cues, we show that, in dialogue, differences among age groups\n    mostly concern stylistic and lexical choices. We believe these findings can inform future work on developing\n    controlled generation models for adaptive conversational systems."\n}\n\n
\n
\n\n\n
\n This work explores an important dimension of variation in the language used by dialogue participants: their age. While previous work showed differences at various linguistic levels between age groups when experimenting with written discourse data (e.g., blog posts), previous work on dialogue has largely been limited to acoustic information related to voice and prosody. Detecting fine-grained linguistic properties of human dialogues is of crucial importance for developing AI- based conversational systems which are able to adapt to their human interlocutors. We therefore investigate whether, and to what extent, current text-based NLP models can detect such linguistic differences, and what the features driving their predictions are. We show that models achieve a fairly good performance on age- group prediction, though the task appears to be more challenging compared to discourse. Through in-depth analysis of the best models’ errors and the most predictive cues, we show that, in dialogue, differences among age groups mostly concern stylistic and lexical choices. We believe these findings can inform future work on developing controlled generation models for adaptive conversational systems.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Is Information Density Uniform in Task-Oriented Dialogues?.\n \n \n \n \n\n\n \n Mario Giulianelli, Arabella Sinclair, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2021. Association for Computational Linguistics.\n \n\n\n\n
\n\n\n\n \n \n \"IsPaper\n  \n \n \n \"Is github\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 20 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-etal-2021-information,\n    title = "Is Information Density Uniform in Task-Oriented Dialogues?",\n    author = "Giulianelli, Mario and Sinclair, Arabella and Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP)",\n    year = "2021",\n    publisher = "Association for Computational Linguistics.",\n    url = "https://aclanthology.org/2021.emnlp-main.652/",\n    url_github = "https://github.com/dmg-illc/uid-dialogue",\n    abstract = "The Uniform Information Density principle states that speakers plan their utterances to reduce fluctuations\n    in the density of the information transmitted. In this paper, we test whether, and within which contextual units this\n    principle holds in task-oriented dialogues. We show that there is evidence supporting the principle in written dialogues\n    where participants play a cooperative reference game as well as in spoken dialogues involving instruction giving and following.\n    Our study underlines the importance of identifying the relevant contextual components, showing that information content\n    increases particularly within topically and referentially related contextual units."\n}\n\n
\n
\n\n\n
\n The Uniform Information Density principle states that speakers plan their utterances to reduce fluctuations in the density of the information transmitted. In this paper, we test whether, and within which contextual units this principle holds in task-oriented dialogues. We show that there is evidence supporting the principle in written dialogues where participants play a cooperative reference game as well as in spoken dialogues involving instruction giving and following. Our study underlines the importance of identifying the relevant contextual components, showing that information content increases particularly within topically and referentially related contextual units.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Analysing Human Strategies of Information Transmission as a Function of Discourse Context.\n \n \n \n \n\n\n \n Mario Giulianelli, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 25th Conference on Computational Natural Language Learning (CoNLL), 2021. Association for Computational Linguistics.\n \n\n\n\n
\n\n\n\n \n \n \"AnalysingPaper\n  \n \n \n \"Analysing github\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 17 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-fernandez-2021-analysing,\n    title = "Analysing Human Strategies of Information Transmission as a Function of Discourse Context",\n    author = "Giulianelli, Mario and Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 25th Conference on Computational Natural Language Learning (CoNLL)",\n    year = "2021",\n    publisher = "Association for Computational Linguistics.",\n    url = "https://aclanthology.org/2021.conll-1.50/",\n    url_github = "https://github.com/dmg-illc/uid-dialogue",\n    abstract = "Speakers are thought to use rational information transmission strategies for efficient communication.\n    Previous work analysing these strategies in sentence production has failed to take into account how the information\n    content of sentences varies as a function of the available discourse context.\n    In this study, we estimate sentence information content within discourse context.\n    We find that speakers transmit information at a stable rate---i.e., rationally---in English newspaper articles\n    but that this rate decreases in spoken open domain and written task-oriented dialogues.\n    We also observe that speakers' choices are not oriented towards local uniformity of information,\n    which is another hypothesised rational strategy. We suggest that a more faithful model of communication\n    should explicitly include production costs and goal-oriented rewards."\n}\n\n
\n
\n\n\n
\n Speakers are thought to use rational information transmission strategies for efficient communication. Previous work analysing these strategies in sentence production has failed to take into account how the information content of sentences varies as a function of the available discourse context. In this study, we estimate sentence information content within discourse context. We find that speakers transmit information at a stable rate—i.e., rationally—in English newspaper articles but that this rate decreases in spoken open domain and written task-oriented dialogues. We also observe that speakers' choices are not oriented towards local uniformity of information, which is another hypothesised rational strategy. We suggest that a more faithful model of communication should explicitly include production costs and goal-oriented rewards.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Construction coordination in first and second language acquisition.\n \n \n \n \n\n\n \n Arabella Sinclair, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 25th Workshop on the Semantics and Pragmatics of Dialogue, Potsdam, Germany, 2021. SEMDIAL\n \n\n\n\n
\n\n\n\n \n \n \"ConstructionPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 7 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{sinclair-fernandez-2021-construction,\n    title = "Construction coordination in first and second language acquisition",\n    author = "Sinclair, Arabella  and Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 25th Workshop on the Semantics and Pragmatics of Dialogue",\n    year = "2021",\n    address = "Potsdam, Germany",\n    publisher = "SEMDIAL",\n    url = "http://semdial.org/anthology/Z21-Sinclair_semdial_0018.pdf",\n    abstract = "Repetition of linguistic forms is a pervasive coordination mechanism in interactive language use.\n    In this paper, we investigate patterns of cross-participant repetition in dialogues where participants have\n    different levels of linguistic ability. Achieving a better understanding of these patterns can not only shed\n    light on how humans coordinate in conversation, but may also contribute to developing more natural and effective\n    dialogue agents in education contexts related to language learning. Our approach is novel in several respects:\n    We focus on multi-word constructions at the lexical and morphosyntactic level, consider both first and second\n    acquisition dialogue, and contrast these setups with adult native conversation. The results of our study show\n    that language acquisition scenarios are characterised by richer inventories of shared constructions but lower\n    usage rates than fluent adult dialogues, and that shared construction use evolves as the linguistic ability\n    of the learners increases, arguably leading to a process of routinisation."\n}\n\n
\n
\n\n\n
\n Repetition of linguistic forms is a pervasive coordination mechanism in interactive language use. In this paper, we investigate patterns of cross-participant repetition in dialogues where participants have different levels of linguistic ability. Achieving a better understanding of these patterns can not only shed light on how humans coordinate in conversation, but may also contribute to developing more natural and effective dialogue agents in education contexts related to language learning. Our approach is novel in several respects: We focus on multi-word constructions at the lexical and morphosyntactic level, consider both first and second acquisition dialogue, and contrast these setups with adult native conversation. The results of our study show that language acquisition scenarios are characterised by richer inventories of shared constructions but lower usage rates than fluent adult dialogues, and that shared construction use evolves as the linguistic ability of the learners increases, arguably leading to a process of routinisation.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Grammatical Profiling for Semantic Change Detection.\n \n \n \n \n\n\n \n Mario Giulianelli, Andrey Kutuzov, & Lidia Pivovarova.\n\n\n \n\n\n\n In Proceedings of the 25th Conference on Computational Natural Language Learning (CoNLL), 2021. Association for Computational Linguistics.\n \n\n\n\n
\n\n\n\n \n \n \"GrammaticalPaper\n  \n \n \n \"Grammatical github\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 10 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-etal-2021-profiling,\n    title = "Grammatical Profiling for Semantic Change Detection",\n    author = "Giulianelli, Mario and Kutuzov, Andrey and Pivovarova, Lidia",\n    booktitle = "Proceedings of the 25th Conference on Computational Natural Language Learning (CoNLL)",\n    year = "2021",\n    publisher = "Association for Computational Linguistics.",\n    url = "https://aclanthology.org/2021.conll-1.33/",\n    url_GitHub = "https://github.com/glnmario/semchange-profiling",\n    abstract = "Semantics, morphology and syntax are strongly interdependent. However, the majority of computational methods for semantic change detection use distributional word representations which encode mostly semantics. We investigate an alternative method, grammatical profiling, based entirely on changes in the morphosyntactic behaviour of words. We demonstrate that it can be used for semantic change detection and even outperforms some distributional semantic methods. We present an in-depth qualitative and quantitative analysis of the predictions made by our grammatical profiling system, showing that they are plausible and interpretable."\n}\n\n
\n
\n\n\n
\n Semantics, morphology and syntax are strongly interdependent. However, the majority of computational methods for semantic change detection use distributional word representations which encode mostly semantics. We investigate an alternative method, grammatical profiling, based entirely on changes in the morphosyntactic behaviour of words. We demonstrate that it can be used for semantic change detection and even outperforms some distributional semantic methods. We present an in-depth qualitative and quantitative analysis of the predictions made by our grammatical profiling system, showing that they are plausible and interpretable.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Probing Cross-Modal Representations in Multi-Step Relational Reasoning.\n \n \n \n \n\n\n \n Iuliia Parfenova, Desmond Elliott, Raquel Fernández, & Sandro Pezzelle.\n\n\n \n\n\n\n In Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021), pages 152–162, Online, August 2021. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"ProbingPaper\n  \n \n \n \"Probing github\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 15 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{parfenova-etal-2021-probing,\n    title = "Probing Cross-Modal Representations in Multi-Step Relational Reasoning",\n    author = "Parfenova, Iuliia  and\n      Elliott, Desmond  and\n      Fern{\\'a}ndez, Raquel  and\n      Pezzelle, Sandro",\n    booktitle = "Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021)",\n    month = aug,\n    year = "2021",\n    address = "Online",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2021.repl4nlp-1.16.pdf",\n    doi = "10.18653/v1/2021.repl4nlp-1.16",\n    url_GitHub = "https://github.com/jig-san/multi-step-size-reasoning",\n    pages = "152--162",\n    abstract = "We investigate the representations learned by vision and language models in tasks that require relational reasoning. Focusing on the problem of assessing the relative size of objects in abstract visual contexts, we analyse both one-step and two-step reasoning. For the latter, we construct a new dataset of three-image scenes and define a task that requires reasoning at the level of the individual images and across images in a scene. We probe the learned model representations using diagnostic classifiers. Our experiments show that pretrained multimodal transformer-based architectures can perform higher-level relational reasoning, and are able to learn representations for novel tasks and data that are very different from what was seen in pretraining.",\n}\n\n
\n
\n\n\n
\n We investigate the representations learned by vision and language models in tasks that require relational reasoning. Focusing on the problem of assessing the relative size of objects in abstract visual contexts, we analyse both one-step and two-step reasoning. For the latter, we construct a new dataset of three-image scenes and define a task that requires reasoning at the level of the individual images and across images in a scene. We probe the learned model representations using diagnostic classifiers. Our experiments show that pretrained multimodal transformer-based architectures can perform higher-level relational reasoning, and are able to learn representations for novel tasks and data that are very different from what was seen in pretraining.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Semantic shift in social networks.\n \n \n \n \n\n\n \n Bill Noble, Asad Sayeed, Raquel Fernández, & Staffan Larsson.\n\n\n \n\n\n\n In Proceedings of *SEM 2021: The Tenth Joint Conference on Lexical and Computational Semantics, pages 26–37, Online, August 2021. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"SemanticPaper\n  \n \n \n \"Semantic github\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 8 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{noble-etal-2021-semantic,\n    title = "Semantic shift in social networks",\n    author = "Noble, Bill  and\n      Sayeed, Asad  and\n      Fern{\\'a}ndez, Raquel  and\n      Larsson, Staffan",\n    booktitle = "Proceedings of *SEM 2021: The Tenth Joint Conference on Lexical and Computational Semantics",\n    month = aug,\n    year = "2021",\n    address = "Online",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2021.starsem-1.3.pdf",\n    doi = "10.18653/v1/2021.starsem-1.3",\n    url_GitHub = "https://github.com/GU-CLASP/semantic-shift-in-social-networks",\n    pages = "26--37",\n    abstract = "Just as the meaning of words is tied to the communities in which they are used, so too is semantic change. But how does lexical semantic change manifest differently across different communities? In this work, we investigate the relationship between community structure and semantic change in 45 communities from the social media website Reddit. We use distributional methods to quantify lexical semantic change and induce a social network on communities, based on interactions between members. We explore the relationship between semantic change and the clustering coefficient of a community{'}s social network graph, as well as community size and stability. While none of these factors are found to be significant on their own, we report a significant effect of their three-way interaction. We also report on significant word-level effects of frequency and change in frequency, which replicate previous findings.",\n}\n\n
\n
\n\n\n
\n Just as the meaning of words is tied to the communities in which they are used, so too is semantic change. But how does lexical semantic change manifest differently across different communities? In this work, we investigate the relationship between community structure and semantic change in 45 communities from the social media website Reddit. We use distributional methods to quantify lexical semantic change and induce a social network on communities, based on interactions between members. We explore the relationship between semantic change and the clustering coefficient of a community's social network graph, as well as community size and stability. While none of these factors are found to be significant on their own, we report a significant effect of their three-way interaction. We also report on significant word-level effects of frequency and change in frequency, which replicate previous findings.\n
\n\n\n
\n\n\n\n\n\n
\n
\n\n
\n
\n  \n 2020\n \n \n (5)\n \n \n
\n
\n \n \n
\n \n\n \n \n \n \n \n \n Refer, Reuse, Reduce: Generating Subsequent References in Visual and Conversational Contexts.\n \n \n \n \n\n\n \n Ece Takmaz, Mario Giulianelli, Sandro Pezzelle, Arabella Sinclair, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 4350–4368, Online, November 2020. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"Refer,Paper\n  \n \n \n \"Refer, dataset\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 22 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{takmaz-etal-2020-refer,\n    title = "{R}efer, {R}euse, {R}educe: {G}enerating {S}ubsequent {R}eferences in {V}isual and {C}onversational {C}ontexts",\n    author = "Takmaz, Ece  and\n      Giulianelli, Mario  and\n      Pezzelle, Sandro  and\n      Sinclair, Arabella  and\n      Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)",\n    month = nov,\n    year = "2020",\n    address = "Online",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2020.emnlp-main.353.pdf",\n    url_dataset = "https://dmg-photobook.github.io",\n    doi = "10.18653/v1/2020.emnlp-main.353",\n    pages = "4350--4368",\n    abstract = "Dialogue participants often refer to entities or situations repeatedly within a conversation, which contributes to its cohesiveness. Subsequent references exploit the common ground accumulated by the interlocutors and hence have several interesting properties, namely, they tend to be shorter and reuse expressions that were effective in previous mentions. In this paper, we tackle the generation of first and subsequent references in visually grounded dialogue. We propose a generation model that produces referring utterances grounded in both the visual and the conversational context. To assess the referring effectiveness of its output, we also implement a reference resolution system. Our experiments and analyses show that the model produces better, more effective referring utterances than a model not grounded in the dialogue context, and generates subsequent references that exhibit linguistic patterns akin to humans.",\n}\n\n
\n
\n\n\n
\n Dialogue participants often refer to entities or situations repeatedly within a conversation, which contributes to its cohesiveness. Subsequent references exploit the common ground accumulated by the interlocutors and hence have several interesting properties, namely, they tend to be shorter and reuse expressions that were effective in previous mentions. In this paper, we tackle the generation of first and subsequent references in visually grounded dialogue. We propose a generation model that produces referring utterances grounded in both the visual and the conversational context. To assess the referring effectiveness of its output, we also implement a reference resolution system. Our experiments and analyses show that the model produces better, more effective referring utterances than a model not grounded in the dialogue context, and generates subsequent references that exhibit linguistic patterns akin to humans.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Generating Image Descriptions via Sequential Cross-Modal Alignment Guided by Human Gaze.\n \n \n \n \n\n\n \n Ece Takmaz, Sandro Pezzelle, Lisa Beinborn, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 4664–4677, Online, November 2020. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"GeneratingPaper\n  \n \n \n \"Generating github\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{takmaz-etal-2020-generating,\n    title = "{G}enerating {I}mage {D}escriptions via {S}equential {C}ross-{M}odal {A}lignment {G}uided by {H}uman {G}aze",\n    author = "Takmaz, Ece  and\n      Pezzelle, Sandro  and\n      Beinborn, Lisa  and\n      Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)",\n    month = nov,\n    year = "2020",\n    address = "Online",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2020.emnlp-main.377.pdf",\n    url_GitHub = "https://github.com/dmg-illc/didec-seq-gen",\n    doi = "10.18653/v1/2020.emnlp-main.353",\n    pages = "4664--4677",\n    abstract = "When speakers describe an image, they tend to look at objects before mentioning them. In this paper, we investigate such sequential cross-modal alignment by modelling the image description generation process computationally. We take as our starting point a state-of-the-art image captioning system and develop several model variants that exploit information from human gaze patterns recorded during language production. In particular, we propose the first approach to image description generation where visual processing is modelled sequentially. Our experiments and analyses confirm that better descriptions can be obtained by exploiting gaze-driven attention and shed light on human cognitive processes by comparing different ways of aligning the gaze modality with language production. We find that processing gaze data sequentially leads to descriptions that are better aligned to those produced by speakers, more diverse, and more natural{---}particularly when gaze is encoded with a dedicated recurrent component.",\n}\n\n\n
\n
\n\n\n
\n When speakers describe an image, they tend to look at objects before mentioning them. In this paper, we investigate such sequential cross-modal alignment by modelling the image description generation process computationally. We take as our starting point a state-of-the-art image captioning system and develop several model variants that exploit information from human gaze patterns recorded during language production. In particular, we propose the first approach to image description generation where visual processing is modelled sequentially. Our experiments and analyses confirm that better descriptions can be obtained by exploiting gaze-driven attention and shed light on human cognitive processes by comparing different ways of aligning the gaze modality with language production. We find that processing gaze data sequentially leads to descriptions that are better aligned to those produced by speakers, more diverse, and more natural—particularly when gaze is encoded with a dedicated recurrent component.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Analysing Lexical Semantic Change with Contextualised Word Representations.\n \n \n \n \n\n\n \n Mario Giulianelli, Marco Del Tredici, & Raquel Fernández.\n\n\n \n\n\n\n In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL), pages 3960–3973, Online, July 2020. Association for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"AnalysingPaper\n  \n \n \n \"Analysing dataset\n  \n \n\n \n \n doi\n  \n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 10 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{giulianelli-etal-2020-analysing,\n    title = "Analysing Lexical Semantic Change with Contextualised Word Representations",\n    author = "Giulianelli, Mario  and\n      Del Tredici, Marco  and\n      Fern{\\'a}ndez, Raquel",\n    booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL)",\n    month = jul,\n    year = "2020",\n    address = "Online",\n    publisher = "Association for Computational Linguistics",\n    url = "https://aclanthology.org/2020.acl-main.365.pdf",\n    url_Dataset = {https://doi.org/10.5281/zenodo.3773250},\n    doi = "10.18653/v1/2020.acl-main.365",\n    pages = "3960--3973",\n    abstract = "This paper presents the first unsupervised approach to lexical semantic change that makes use of contextualised word representations. We propose a novel method that exploits the BERT neural language model to obtain representations of word usages, clusters these representations into usage types, and measures change along time with three proposed metrics. We create a new evaluation dataset and show that the model representations and the detected semantic shifts are positively correlated with human judgements. Our extensive qualitative analysis demonstrates that our method captures a variety of synchronic and diachronic linguistic phenomena. We expect our work to inspire further research in this direction.",\n}\n\n
\n
\n\n\n
\n This paper presents the first unsupervised approach to lexical semantic change that makes use of contextualised word representations. We propose a novel method that exploits the BERT neural language model to obtain representations of word usages, clusters these representations into usage types, and measures change along time with three proposed metrics. We create a new evaluation dataset and show that the model representations and the detected semantic shifts are positively correlated with human judgements. Our extensive qualitative analysis demonstrates that our method captures a variety of synchronic and diachronic linguistic phenomena. We expect our work to inspire further research in this direction.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n UiO-UvA at SemEval-2020 Task 1: Contextualised Embeddings for Lexical Semantic Change Detection.\n \n \n \n \n\n\n \n Andrey Kutuzov, & Mario Giulianelli.\n\n\n \n\n\n\n In Proceedings of the Fourteenth Workshop on Semantic Evaluation, pages 126–134, Barcelona (online), December 2020. International Committee for Computational Linguistics\n \n\n\n\n
\n\n\n\n \n \n \"UiO-UvAPaper\n  \n \n \n \"UiO-UvA github\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n  \n \n abstract \n \n\n \n  \n \n 5 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{kutuzov-giulianelli-2020-change,\n    title = "{U}i{O}-{U}v{A} at {S}em{E}val-2020 Task 1: Contextualised Embeddings for Lexical Semantic Change Detection",\n    author = "Kutuzov, Andrey  and\n      Giulianelli, Mario",\n    booktitle = "Proceedings of the Fourteenth Workshop on Semantic Evaluation",\n    month = dec,\n    year = "2020",\n    address = "Barcelona (online)",\n    publisher = "International Committee for Computational Linguistics",\n    url = "https://aclanthology.org/2020.semeval-1.14",\n    url_GitHub = "https://github.com/akutuzov/semeval2020",\n    pages = "126--134",\n    abstract = "We apply contextualised word embeddings to lexical semantic change detection in the SemEval-2020 Shared Task 1. This paper focuses on Subtask 2, ranking words by the degree of their semantic drift over time. We analyse the performance of two contextualising architectures (BERT and ELMo) and three change detection algorithms. We find that the most effective algorithms rely on the cosine similarity between averaged token embeddings and the pairwise distances between token embeddings. They outperform strong baselines by a large margin (in the post-evaluation phase, we have the best Subtask 2 submission for SemEval-2020 Task 1), but interestingly, the choice of a particular algorithm depends on the distribution of gold scores in the test set.",\n}\n\n\n
\n
\n\n\n
\n We apply contextualised word embeddings to lexical semantic change detection in the SemEval-2020 Shared Task 1. This paper focuses on Subtask 2, ranking words by the degree of their semantic drift over time. We analyse the performance of two contextualising architectures (BERT and ELMo) and three change detection algorithms. We find that the most effective algorithms rely on the cosine similarity between averaged token embeddings and the pairwise distances between token embeddings. They outperform strong baselines by a large margin (in the post-evaluation phase, we have the best Subtask 2 submission for SemEval-2020 Task 1), but interestingly, the choice of a particular algorithm depends on the distribution of gold scores in the test set.\n
\n\n\n
\n\n\n
\n \n\n \n \n \n \n \n \n Grounded and Ungrounded Referring Expressions in Human Dialogues: Language Mirrors Different Grounding Conditions.\n \n \n \n \n\n\n \n Eleonora Gualdoni, Raffaella Bernardi, Raquel Fernández, & Sandro Pezzelle.\n\n\n \n\n\n\n In Proceedings of the 7th Italian Conference on Computational Linguistics (CLiC-it), 2020. \n \n\n\n\n
\n\n\n\n \n \n \"GroundedPaper\n  \n \n\n \n\n \n link\n  \n \n\n bibtex\n \n\n \n\n \n  \n \n 8 downloads\n \n \n\n \n \n \n \n \n \n \n\n  \n \n \n\n\n\n
\n
@inproceedings{gualdoni-etal:2020:clicit,\n  author={Eleonora Gualdoni and Raffaella Bernardi and Raquel Fern\\'andez and Sandro Pezzelle},\n  title={Grounded and Ungrounded Referring Expressions in Human Dialogues: Language Mirrors Different Grounding Conditions},\n  booktitle={Proceedings of the 7th Italian Conference on Computational Linguistics (CLiC-it)},\n  year={2020},\n  url={http://ceur-ws.org/Vol-2769/paper_38.pdf}\n}\n
\n
\n\n\n\n
\n\n\n\n\n\n
\n
\n\n\n\n\n
\n\n\n \n\n \n \n \n \n\n
\n"}; document.write(bibbase_data.data);