\n \n \n
\n
\n\n \n \n \n \n \n \n When and Where Did it Happen? An Encoder-Decoder Model to Identify Scenario Context.\n \n \n \n \n\n\n \n Enrique Noriega-Atala; Robert Vacareanu; Salena Torres Ashton; Adarsh Pyarelal; Clayton T Morrison; and Mihai Surdeanu.\n\n\n \n\n\n\n In Yaser Al-Onaizan; Mohit Bansal; and Yun-Nung Chen., editor(s),
Findings of the Association for Computational Linguistics: EMNLP 2024, pages 3821–3829, Miami, Florida, USA, November 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{noriega-atala-etal-2024-happen,\n title = "When and Where Did it Happen? An Encoder-Decoder Model to Identify Scenario Context",\n author = "Noriega-Atala, Enrique and\n Vacareanu, Robert and\n Ashton, Salena Torres and\n Pyarelal, Adarsh and\n Morrison, Clayton T and\n Surdeanu, Mihai",\n editor = "Al-Onaizan, Yaser and\n Bansal, Mohit and\n Chen, Yun-Nung",\n booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",\n month = nov,\n year = "2024",\n address = "Miami, Florida, USA",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.findings-emnlp.219/",\n doi = "10.18653/v1/2024.findings-emnlp.219",\n pages = "3821--3829",\n abstract = "We introduce a neural architecture finetuned for the task of scenario context generation: The relevant location and time of an event or entity mentioned in text. Contextualizing information extraction helps to scope the validity of automated finings when aggregating them as knowledge graphs. Our approach uses a high-quality curated dataset of time and location annotations in a corpus of epidemiology papers to train an encoder-decoder architecture. We also explored the use of data augmentation techniques during training. Our findings suggest that a relatively small fine-tuned encoder-decoder model performs better than out-of-the-box LLMs and semantic role labeling parsers to accurate predict the relevant scenario information of a particular entity or event."\n}\n\n
\n
\n\n\n
\n We introduce a neural architecture finetuned for the task of scenario context generation: The relevant location and time of an event or entity mentioned in text. Contextualizing information extraction helps to scope the validity of automated finings when aggregating them as knowledge graphs. Our approach uses a high-quality curated dataset of time and location annotations in a corpus of epidemiology papers to train an encoder-decoder architecture. We also explored the use of data augmentation techniques during training. Our findings suggest that a relatively small fine-tuned encoder-decoder model performs better than out-of-the-box LLMs and semantic role labeling parsers to accurate predict the relevant scenario information of a particular entity or event.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Finding a Wolf in Sheep's Clothing: Combating Adversarial Text-To-Image Prompts with Text Summarization.\n \n \n \n \n\n\n \n Portia Cooper; Harshita Narnoli; and Mihai Surdeanu.\n\n\n \n\n\n\n 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@misc{cooper2024findingwolfsheepsclothing,\n title={Finding a Wolf in Sheep's Clothing: Combating Adversarial Text-To-Image Prompts with Text Summarization}, \n author={Portia Cooper and Harshita Narnoli and Mihai Surdeanu},\n year={2024},\n eprint={2412.12212},\n archivePrefix={arXiv},\n primaryClass={cs.CR},\n url={https://arxiv.org/abs/2412.12212}, \n}\n\n
\n
\n\n\n\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Learning to Generate Rules for Realistic Few-Shot Relation Classification: An Encoder-Decoder Approach.\n \n \n \n\n\n \n Mayank Singh; and Eduardo Blanco.\n\n\n \n\n\n\n In
Findings of the Association for Computational Linguistics: EMNLP 2024, Miami, USA and virtual meeting, November 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{singh-2024-learning,\n title = "Learning to Generate Rules for Realistic Few-Shot Relation Classification: An Encoder-Decoder Approach",\n author = "Singh, Mayank and Blanco, Eduardo",\n booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",\n month = nov,\n year = "2024",\n address = "Miami, USA and virtual meeting",\n publisher = "Association for Computational Linguistics",\n abstract = "We propose a neuro-symbolic approach for realistic few-shot relation classification via rules. Instead of building neural models to predict relations, we design them to output straightforward rules that can be used to extract relations. The rules are generated using custom T5-style Encoder-Decoder Language Models. Crucially, our rules are fully interpretable and pliable (i.e., humans can easily modify them to boost performance). Through a combination of rules generated by these models along with a very effective, novel baseline, we demonstrate a few-shot relation-classification performance that is comparable to or stronger than the state of the art on the Few-Shot TACRED and NYT29 benchmarks while increasing interpretability and maintaining pliability.",\n}\n\n
\n
\n\n\n
\n We propose a neuro-symbolic approach for realistic few-shot relation classification via rules. Instead of building neural models to predict relations, we design them to output straightforward rules that can be used to extract relations. The rules are generated using custom T5-style Encoder-Decoder Language Models. Crucially, our rules are fully interpretable and pliable (i.e., humans can easily modify them to boost performance). Through a combination of rules generated by these models along with a very effective, novel baseline, we demonstrate a few-shot relation-classification performance that is comparable to or stronger than the state of the art on the Few-Shot TACRED and NYT29 benchmarks while increasing interpretability and maintaining pliability.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Paraphrasing in Affirmative Terms Improves Negation Understanding.\n \n \n \n \n\n\n \n MohammadHossein Rezaei; and Eduardo Blanco.\n\n\n \n\n\n\n In
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Bangkok, Thailand, August 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{rezaei-blanco-2024-paraphrasing,\n author = {Rezaei, MohammadHossein and Blanco, Eduardo},\n title = {Paraphrasing in Affirmative Terms Improves Negation Understanding},\n booktitle = {Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)},\n month = {August},\n year = {2024},\n address = {Bangkok, Thailand},\n publisher = {Association for Computational Linguistics},\n pages = {},\n abstract = {Negation is a common linguistic phenomenon. Yet language models face challenges with negation in many natural language understanding tasks such as question answering and natural language inference. In this paper, we experiment with seamless strategies that incorporate affirmative interpretations (i.e., paraphrases without negation) to make models more robust against negation. Crucially, our affirmative interpretations are obtained automatically. We show improvements with CondaQA, a large corpus requiring reasoning with negation, and five natural language understanding tasks.},\n url = {https://arxiv.org/pdf/2406.07492}\n}\n\n
\n
\n\n\n
\n Negation is a common linguistic phenomenon. Yet language models face challenges with negation in many natural language understanding tasks such as question answering and natural language inference. In this paper, we experiment with seamless strategies that incorporate affirmative interpretations (i.e., paraphrases without negation) to make models more robust against negation. Crucially, our affirmative interpretations are obtained automatically. We show improvements with CondaQA, a large corpus requiring reasoning with negation, and five natural language understanding tasks.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n MARiA at SemEval 2024 Task-6: Hallucination Detection Through LLMs, MNLI, and Cosine similarity.\n \n \n \n \n\n\n \n Reza Sanayei; Abhyuday Singh; Mohammadhossein Rezaei; and Steven Bethard.\n\n\n \n\n\n\n In Atul Kr. Ojha; A. Seza Doğruöz; Harish Tayyar Madabushi; Giovanni Da San Martino; Sara Rosenthal; and Aiala Rosá., editor(s),
Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024), pages 1584–1588, Mexico City, Mexico, June 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 2 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{sanayei-etal-2024-maria,\n title = "{MAR}i{A} at {S}em{E}val 2024 Task-6: Hallucination Detection Through {LLM}s, {MNLI}, and Cosine similarity",\n author = "Sanayei, Reza and\n Singh, Abhyuday and\n Rezaei, Mohammadhossein and\n Bethard, Steven",\n editor = {Ojha, Atul Kr. and\n Do{\\u{g}}ru{\\"o}z, A. Seza and\n Tayyar Madabushi, Harish and\n Da San Martino, Giovanni and\n Rosenthal, Sara and\n Ros{\\'a}, Aiala},\n booktitle = "Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.semeval-1.225",\n pages = "1584--1588",\n abstract = "The advent of large language models (LLMs) has revolutionized Natural Language Generation (NLG), offering unmatched text generation capabilities. However, this progress introduces significant challenges, notably hallucinations{---}semantically incorrect yet fluent outputs. This phenomenon undermines content reliability, as traditional detection systems focus more on fluency than accuracy, posing a risk of misinformation spread.Our study addresses these issues by proposing a unified strategy for detecting hallucinations in neural model-generated text, focusing on the SHROOM task in SemEval 2024. We employ diverse methodologies to identify output divergence from the source content. We utilized Sentence Transformers to measure cosine similarity between source-hypothesis and source-target embeddings, experimented with omitting source content in the cosine similarity computations, and Leveragied LLMs{'} In-Context Learning with detailed task prompts as our methodologies. The varying performance of our different approaches across the subtasks underscores the complexity of Natural Language Understanding tasks, highlighting the importance of addressing the nuances of semantic correctness in the era of advanced language models.",\n}\n\n
\n
\n\n\n
\n The advent of large language models (LLMs) has revolutionized Natural Language Generation (NLG), offering unmatched text generation capabilities. However, this progress introduces significant challenges, notably hallucinations—semantically incorrect yet fluent outputs. This phenomenon undermines content reliability, as traditional detection systems focus more on fluency than accuracy, posing a risk of misinformation spread.Our study addresses these issues by proposing a unified strategy for detecting hallucinations in neural model-generated text, focusing on the SHROOM task in SemEval 2024. We employ diverse methodologies to identify output divergence from the source content. We utilized Sentence Transformers to measure cosine similarity between source-hypothesis and source-target embeddings, experimented with omitting source content in the cosine similarity computations, and Leveragied LLMs' In-Context Learning with detailed task prompts as our methodologies. The varying performance of our different approaches across the subtasks underscores the complexity of Natural Language Understanding tasks, highlighting the importance of addressing the nuances of semantic correctness in the era of advanced language models.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n CLULab-UofA at SemEval-2024 Task 8: Detecting Machine-Generated Text Using Triplet-Loss-Trained Text Similarity and Text Classification.\n \n \n \n \n\n\n \n Mohammadhossein Rezaei; Yeaeun Kwon; Reza Sanayei; Abhyuday Singh; and Steven Bethard.\n\n\n \n\n\n\n In Atul Kr. Ojha; A. Seza Doğruöz; Harish Tayyar Madabushi; Giovanni Da San Martino; Sara Rosenthal; and Aiala Rosá., editor(s),
Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024), pages 1498–1504, Mexico City, Mexico, June 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 3 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{rezaei-etal-2024-clulab,\n title = "{CLUL}ab-{U}of{A} at {S}em{E}val-2024 Task 8: Detecting Machine-Generated Text Using Triplet-Loss-Trained Text Similarity and Text Classification",\n author = "Rezaei, Mohammadhossein and\n Kwon, Yeaeun and\n Sanayei, Reza and\n Singh, Abhyuday and\n Bethard, Steven",\n editor = {Ojha, Atul Kr. and\n Do{\\u{g}}ru{\\"o}z, A. Seza and\n Tayyar Madabushi, Harish and\n Da San Martino, Giovanni and\n Rosenthal, Sara and\n Ros{\\'a}, Aiala},\n booktitle = "Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.semeval-1.215",\n pages = "1498--1504",\n abstract = "Detecting machine-generated text is a critical task in the era of large language models. In this paper, we present our systems for SemEval-2024 Task 8, which focuses on multi-class classification to discern between human-written and maching-generated texts by five state-of-the-art large language models. We propose three different systems: unsupervised text similarity, triplet-loss-trained text similarity, and text classification. We show that the triplet-loss trained text similarity system outperforms the other systems, achieving 80{\\%} accuracy on the test set and surpassing the baseline model for this subtask. Additionally, our text classification system, which takes into account sentence paraphrases generated by the candidate models, also outperforms the unsupervised text similarity system, achieving 74{\\%} accuracy.",\n}\n\n
\n
\n\n\n
\n Detecting machine-generated text is a critical task in the era of large language models. In this paper, we present our systems for SemEval-2024 Task 8, which focuses on multi-class classification to discern between human-written and maching-generated texts by five state-of-the-art large language models. We propose three different systems: unsupervised text similarity, triplet-loss-trained text similarity, and text classification. We show that the triplet-loss trained text similarity system outperforms the other systems, achieving 80% accuracy on the test set and surpassing the baseline model for this subtask. Additionally, our text classification system, which takes into account sentence paraphrases generated by the candidate models, also outperforms the unsupervised text similarity system, achieving 74% accuracy.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Retrieval Augmented Generation of Subjective Explanations for Socioeconomic Scenarios.\n \n \n \n \n\n\n \n Razvan-Gabriel Dumitru; Maria Alexeeva; Keith Alcock; Nargiza Ludgate; Cheonkam Jeong; Zara Fatima Abdurahaman; Prateek Puri; Brian Kirchhoff; Santadarshan Sadhu; and Mihai Surdeanu.\n\n\n \n\n\n\n In
Sixth Workshop on NLP and Computational Social Science (at NAACL) 2024, 2024. \n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 5 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{\n dumitru2024retrieval,\n title={Retrieval Augmented Generation of Subjective Explanations for Socioeconomic Scenarios},\n author={Dumitru, Razvan-Gabriel and Alexeeva, Maria and Alcock, Keith and Ludgate, Nargiza and Jeong, Cheonkam and Abdurahaman, Zara Fatima and Puri, Prateek and Kirchhoff, Brian and Sadhu, Santadarshan and Surdeanu, Mihai},\n booktitle={Sixth Workshop on NLP and Computational Social Science (at NAACL) 2024},\n year={2024},\n url={http://clulab.org/papers/naacl-css2024-rag.pdf}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n General Purpose Verification for Chain of Thought Prompting.\n \n \n \n \n\n\n \n Robert Vacareanu; Anurag Pratik; Evangelia Spiliopoulou; Zheng Qi; Giovanni Paolini; Neha Anna John; Jie Ma; Yassine Benajiba; and Miguel Ballesteros.\n\n\n \n\n\n\n
ArXiv, abs/2405.00204. 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{Vacareanu2024GeneralVerificationLLM,\n title={General Purpose Verification for Chain of Thought Prompting},\n author={Robert Vacareanu and Anurag Pratik and Evangelia Spiliopoulou and Zheng Qi and Giovanni Paolini and Neha Anna John and Jie Ma and Yassine Benajiba and Miguel Ballesteros},\n journal={ArXiv},\n year={2024},\n volume={abs/2405.00204},\n url={https://arxiv.org/pdf/2405.00204.pdf}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n From Words to Numbers: Your Large Language Model Is Secretly A Capable Regressor When Given In-Context Examples.\n \n \n \n \n\n\n \n Robert Vacareanu; Vlad-Andrei Negru; Vasile Suciu; and Mihai Surdeanu.\n\n\n \n\n\n\n
ArXiv, abs/2404.07544. 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 7 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{Vacareanu2024LLMsRegression,\n title={From Words to Numbers: Your Large Language Model Is Secretly A Capable Regressor When Given In-Context Examples},\n author={Robert Vacareanu and Vlad-Andrei Negru and Vasile Suciu and Mihai Surdeanu},\n journal={ArXiv},\n year={2024},\n volume={abs/2404.07544},\n url={https://arxiv.org/pdf/2404.07544.pdf}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Best of Both Worlds: A Pliable and Generalizable Neuro-Symbolic Approach for Relation Classification.\n \n \n \n \n\n\n \n Robert Vacareanu; Fahmida Alam; Md Asiful Islam; Haris Riaz; and Mihai Surdeanu.\n\n\n \n\n\n\n In
Findings of the Association for Computational Linguistics: NAACL 2024, Mexico City, Mexico, June 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 25 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{vacareanu2024softrules,\n title = "Best of Both Worlds: A Pliable and Generalizable Neuro-Symbolic Approach for Relation Classification",\n author = "Robert Vacareanu and Fahmida Alam and Md Asiful Islam and Haris Riaz and Mihai Surdeanu",\n booktitle = "Findings of the Association for Computational Linguistics: NAACL 2024",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://arxiv.org/pdf/2403.03305.pdf",\n abstract = "This paper introduces a novel neuro-symbolic architecture for relation classification (RC) that combines rule-based methods with contemporary deep learning techniques. This approach capitalizes on the strengths of both paradigms: the adaptability of rule-based systems and the generalization power of neural networks. Our architecture consists of two components: a declarative rule-based model for transparent classification and a neural component to enhance rule generalizability through semantic text matching. Notably, our semantic matcher is trained in an unsupervised domain-agnostic way, solely with synthetic data. Further, these components are loosely coupled, allowing for rule modifications without retraining the semantic matcher. In our evaluation, we focused on two few-shot relation classification datasets: Few-Shot TACRED and a Few-Shot version of NYT29. We show that our proposed method outperforms previous state-of-the-art models in three out of four settings, despite not seeing any human-annotated training data. Further, we show that our approach remains modular and pliable, i.e., the corresponding rules can be locally modified to improve the overall model. Human interventions to the rules for the TACRED relation \\texttt{org:parents} boost the performance on that relation by as much as 26\\% relative improvement, without negatively impacting the other relations, and without retraining the semantic matching component.",\n}\n\n
\n
\n\n\n
\n This paper introduces a novel neuro-symbolic architecture for relation classification (RC) that combines rule-based methods with contemporary deep learning techniques. This approach capitalizes on the strengths of both paradigms: the adaptability of rule-based systems and the generalization power of neural networks. Our architecture consists of two components: a declarative rule-based model for transparent classification and a neural component to enhance rule generalizability through semantic text matching. Notably, our semantic matcher is trained in an unsupervised domain-agnostic way, solely with synthetic data. Further, these components are loosely coupled, allowing for rule modifications without retraining the semantic matcher. In our evaluation, we focused on two few-shot relation classification datasets: Few-Shot TACRED and a Few-Shot version of NYT29. We show that our proposed method outperforms previous state-of-the-art models in three out of four settings, despite not seeing any human-annotated training data. Further, we show that our approach remains modular and pliable, i.e., the corresponding rules can be locally modified to improve the overall model. Human interventions to the rules for the TACRED relation \\textttorg:parents boost the performance on that relation by as much as 26% relative improvement, without negatively impacting the other relations, and without retraining the semantic matching component.\n
\n\n\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n A Weak Supervision Approach for Few-Shot Aspect Based Sentiment Analysis.\n \n \n \n \n\n\n \n Robert Vacareanu; Siddharth Varia; Kishaloy Halder; Shuai Wang; Giovanni Paolini; Neha Anna John; Miguel Ballesteros; and Smaranda Muresan.\n\n\n \n\n\n\n In Yvette Graham; and Matthew Purver., editor(s),
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2734–2752, St. Julian's, Malta, March 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 2 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{vacareanu-etal-2024-weak,\n title = "A Weak Supervision Approach for Few-Shot Aspect Based Sentiment Analysis",\n author = "Vacareanu, Robert and\n Varia, Siddharth and\n Halder, Kishaloy and\n Wang, Shuai and\n Paolini, Giovanni and\n Anna John, Neha and\n Ballesteros, Miguel and\n Muresan, Smaranda",\n editor = "Graham, Yvette and\n Purver, Matthew",\n booktitle = "Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = mar,\n year = "2024",\n address = "St. Julian{'}s, Malta",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.eacl-long.167",\n pages = "2734--2752",\n abstract = "We explore how weak supervision on abundant unlabeled data can be leveraged to improve few-shot performance in aspect-based sentiment analysis (ABSA) tasks. We propose a pipeline approach to construct a noisy ABSA dataset, and we use it to adapt a pre-trained sequence-to-sequence model to the ABSA tasks. We test the resulting model on three widely used ABSA datasets, before and after fine-tuning. Our proposed method preserves the full fine-tuning performance while showing significant improvements (15.84 absolute F1) in the few-shot learning scenario for the harder tasks. In zero-shot (i.e., without fine-tuning), our method outperforms the previous state of the art on the aspect extraction sentiment classification (AESC) task and is, additionally, capable of performing the harder aspect sentiment triplet extraction (ASTE) task.",\n}\n\n
\n
\n\n\n
\n We explore how weak supervision on abundant unlabeled data can be leveraged to improve few-shot performance in aspect-based sentiment analysis (ABSA) tasks. We propose a pipeline approach to construct a noisy ABSA dataset, and we use it to adapt a pre-trained sequence-to-sequence model to the ABSA tasks. We test the resulting model on three widely used ABSA datasets, before and after fine-tuning. Our proposed method preserves the full fine-tuning performance while showing significant improvements (15.84 absolute F1) in the few-shot learning scenario for the harder tasks. In zero-shot (i.e., without fine-tuning), our method outperforms the previous state of the art on the aspect extraction sentiment classification (AESC) task and is, additionally, capable of performing the harder aspect sentiment triplet extraction (ASTE) task.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Interpreting Answers to Yes-No Questions in Dialogues from Multiple Domains.\n \n \n \n\n\n \n Zijie Wang; Farzana Rashid; and Eduardo Blanco.\n\n\n \n\n\n\n In
Findings of the Association for Computational Linguistics: NAACL 2024, Mexico City, Mexico, June 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{wang2024naaclfindings,\n title = "Interpreting Answers to Yes-No Questions in Dialogues from Multiple Domains",\n author = "Wang, Zijie and Rashid, Farzana and Blanco, Eduardo",\n booktitle = "Findings of the Association for Computational Linguistics: NAACL 2024",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics"\n}\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Time Travel in LLMs: Tracing Data Contamination in Large Language Models.\n \n \n \n \n\n\n \n Shahriar Golchin; and Mihai Surdeanu.\n\n\n \n\n\n\n In
Proceedings of the Twelfth International Conference on Learning Representations (ICLR), 2024. \n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 6 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{golchin2024time,\n\ttitle={Time Travel in {LLM}s: Tracing Data Contamination in Large Language Models},\n\tauthor={Shahriar Golchin and Mihai Surdeanu},\n\tbooktitle={Proceedings of the Twelfth International Conference on Learning Representations (ICLR)},\n\tyear={2024},\n\turl={https://openreview.net/forum?id=2Rwq6c3tvr}\n}\n\n
\n
\n\n\n\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n ELLEN: Extremely Lightly Supervised Learning For Efficient Named Entity Recognition.\n \n \n \n \n\n\n \n Haris Riaz; Razvan-Gabriel Dumitru; and Mihai Surdeanu.\n\n\n \n\n\n\n In
Proceedings of the Joint International Conference on Computational Linguistics, Language Resources and Evaluation, Torino, Italy, May 2024. European Language Resources Association\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{riaz2024ellen,\n title = "ELLEN: Extremely Lightly Supervised Learning For Efficient Named Entity Recognition",\n author = "Haris Riaz and Razvan-Gabriel Dumitru and Mihai Surdeanu",\n booktitle = "Proceedings of the Joint International Conference on Computational Linguistics, Language Resources and Evaluation",\n month = may,\n year = "2024",\n address = "Torino, Italy",\n publisher = "European Language Resources Association",\n url = "https://arxiv.org/pdf/2403.17385.pdf",\n abstract = "In this work, we revisit the problem of semi-supervised named entity recognition (NER) focusing on extremely light supervision, consisting of a lexicon containing only 10 examples per class. We introduce ELLEN, a simple, fully modular, neuro-symbolic method that blends fine-tuned language models with linguistic rules. These rules include insights such as ''One Sense Per Discourse'', using a Masked Language Model as an unsupervised NER, leveraging part-of-speech tags to identify and eliminate unlabeled entities as false negatives, and other intuitions about classifier confidence scores in local and global context. ELLEN achieves very strong performance on the CoNLL-2003 dataset when using the minimal supervision from the lexicon above. It also outperforms most existing (and considerably more complex) semi-supervised NER methods under the same supervision settings commonly used in the literature (i.e., 5% of the training data). Further, we evaluate our CoNLL-2003 model in a zero-shot scenario on WNUT-17 where we find that it outperforms GPT-3.5 and achieves comparable performance to GPT-4. In a zero-shot setting, ELLEN also achieves over 75% of the performance of a strong, fully supervised model trained on gold data. Our code is available at: https://github.com/hriaz17/ELLEN",\n}\n\n
\n
\n\n\n
\n In this work, we revisit the problem of semi-supervised named entity recognition (NER) focusing on extremely light supervision, consisting of a lexicon containing only 10 examples per class. We introduce ELLEN, a simple, fully modular, neuro-symbolic method that blends fine-tuned language models with linguistic rules. These rules include insights such as ''One Sense Per Discourse'', using a Masked Language Model as an unsupervised NER, leveraging part-of-speech tags to identify and eliminate unlabeled entities as false negatives, and other intuitions about classifier confidence scores in local and global context. ELLEN achieves very strong performance on the CoNLL-2003 dataset when using the minimal supervision from the lexicon above. It also outperforms most existing (and considerably more complex) semi-supervised NER methods under the same supervision settings commonly used in the literature (i.e., 5% of the training data). Further, we evaluate our CoNLL-2003 model in a zero-shot scenario on WNUT-17 where we find that it outperforms GPT-3.5 and achieves comparable performance to GPT-4. In a zero-shot setting, ELLEN also achieves over 75% of the performance of a strong, fully supervised model trained on gold data. Our code is available at: https://github.com/hriaz17/ELLEN\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n On Learning Bipolar Gradual Argumentation Semantics with Neural Networks.\n \n \n \n \n\n\n \n Caren Al Anaissy; Sandeep Suntwal; Mihai Surdeanu; and Srdjan Vesic.\n\n\n \n\n\n\n In
Proceedings of the 16th International Conference on Agents and Artificial Intelligence (ICAART), 2024. \n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 11 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{anaissy-icaart2024,\n title = "On Learning Bipolar Gradual Argumentation Semantics with Neural Networks",\n author = "Caren Al Anaissy and Sandeep Suntwal and Mihai Surdeanu and Srdjan Vesic",\n booktitle = "Proceedings of the 16th International Conference on Agents and Artificial Intelligence (ICAART)",\n year = "2024",\n url = "https://clulab.org/papers/icaart2024.pdf",\n abstract = "Computational argumentation has evolved as a key area in artificial intelligence, used to analyze aspects of thinking, making decisions, and conversing. As a result, it is currently employed in a variety of real-world contexts, from legal reasoning to intelligence analysis. An argumentation framework is modelled as a graph where the nodes represent arguments and the edges of the graph represent relations (i.e., supports, attacks) between nodes. In this work, we investigate the ability of neural network methods to learn a gradual bipolar argumentation semantics, which allows for both supports and attacks. We begin by calculating the acceptability degrees for graph nodes. These scores are generated using Quantitative Argumentation Debate (QuAD) argumentation semantics. We apply this approach to two benchmark datasets: Twelve Angry Men and Debate- pedia. Using this data, we train and evaluate the performance of three benchmark architectures: Multilayer Perceptron (MLP), Graph Convolution Network (GCN), and Graph Attention Network (GAT) to learn the acceptability degree scores produced by the QuAD semantics. Our results show that these neural network methods can learn bipolar gradual argumentation semantics. The models trained on GCN architecture perform better than the other two architectures underscoring the importance of modelling argumentation graphs explicitly."\n}\n\n
\n
\n\n\n
\n Computational argumentation has evolved as a key area in artificial intelligence, used to analyze aspects of thinking, making decisions, and conversing. As a result, it is currently employed in a variety of real-world contexts, from legal reasoning to intelligence analysis. An argumentation framework is modelled as a graph where the nodes represent arguments and the edges of the graph represent relations (i.e., supports, attacks) between nodes. In this work, we investigate the ability of neural network methods to learn a gradual bipolar argumentation semantics, which allows for both supports and attacks. We begin by calculating the acceptability degrees for graph nodes. These scores are generated using Quantitative Argumentation Debate (QuAD) argumentation semantics. We apply this approach to two benchmark datasets: Twelve Angry Men and Debate- pedia. Using this data, we train and evaluate the performance of three benchmark architectures: Multilayer Perceptron (MLP), Graph Convolution Network (GCN), and Graph Attention Network (GAT) to learn the acceptability degree scores produced by the QuAD semantics. Our results show that these neural network methods can learn bipolar gradual argumentation semantics. The models trained on GCN architecture perform better than the other two architectures underscoring the importance of modelling argumentation graphs explicitly.\n
\n\n\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n hinoki at SemEval-2024 Task 7: Numeral-Aware Headline Generation (English).\n \n \n \n \n\n\n \n Hinoki Crum; and Steven Bethard.\n\n\n \n\n\n\n In Atul Kr. Ojha; A. Seza Doğruöz; Harish Tayyar Madabushi; Giovanni Da San Martino; Sara Rosenthal; and Aiala Rosá., editor(s),
Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024), pages 34–39, Mexico City, Mexico, June 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n\n\n\n
\n
@inproceedings{crum-bethard-2024-hinoki,\n title = "hinoki at {S}em{E}val-2024 Task 7: Numeral-Aware Headline Generation ({E}nglish)",\n author = "Crum, Hinoki and\n Bethard, Steven",\n editor = {Ojha, Atul Kr. and\n Do{\\u{g}}ru{\\"o}z, A. Seza and\n Tayyar Madabushi, Harish and\n Da San Martino, Giovanni and\n Rosenthal, Sara and\n Ros{\\'a}, Aiala},\n booktitle = "Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.semeval-1.6",\n doi = "10.18653/v1/2024.semeval-1.6",\n pages = "34--39",\n keywords = {shared task paper},\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Semi-Structured Chain-of-Thought: Integrating Multiple Sources of Knowledge for Improved Language Model Reasoning.\n \n \n \n \n\n\n \n Xin Su; Tiep Le; Steven Bethard; and Phillip Howard.\n\n\n \n\n\n\n In
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 8597–8613, Mexico City, Mexico, June 2024. Association for Computational Linguistics\n
[Acceptance rate 23%]\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 3 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n\n\n\n
\n
@inproceedings{su-etal-2024-semi,\n title = "Semi-Structured Chain-of-Thought: Integrating Multiple Sources of Knowledge for Improved Language Model Reasoning",\n author = "Su, Xin and\n Le, Tiep and\n Bethard, Steven and\n Howard, Phillip",\n booktitle = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.naacl-long.475",\n doi = "10.18653/v1/2024.naacl-long.475",\n pages = "8597--8613",\n keywords = {question answering},\n note = {[Acceptance rate 23\\%]},\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Improving Toponym Resolution by Predicting Attributes to Constrain Geographical Ontology Entries.\n \n \n \n \n\n\n \n Zeyu Zhang; Egoitz Laparra; and Steven Bethard.\n\n\n \n\n\n\n In
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers), pages 35–44, Mexico City, Mexico, June 2024. Association for Computational Linguistics\n
[Acceptance rate 23%]\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{zhang-etal-2024-improving-toponym,\n title = "Improving Toponym Resolution by Predicting Attributes to Constrain Geographical Ontology Entries",\n author = "Zhang, Zeyu and\n Laparra, Egoitz and\n Bethard, Steven",\n booktitle = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers)",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.naacl-short.3",\n doi = "10.18653/v1/2024.naacl-short.3",\n pages = "35--44",\n keywords = {geospatial normalization, information extraction},\n note = {[Acceptance rate 23\\%]},\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Proceedings of the 6th Clinical Natural Language Processing Workshop.\n \n \n \n \n\n\n \n Tristan Naumann; Asma Ben Abacha; Steven Bethard; Kirk Roberts; and Danielle Bitterman.,\n editors.\n \n\n\n \n\n\n\n Association for Computational Linguistics. Mexico City, Mexico, June 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n\n\n\n
\n
@proceedings{clinicalnlp-2024-clinical,\n title = "Proceedings of the 6th Clinical Natural Language Processing Workshop",\n editor = "Naumann, Tristan and\n Ben Abacha, Asma and\n Bethard, Steven and\n Roberts, Kirk and\n Bitterman, Danielle",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.clinicalnlp-1.0",\n keywords = "health",\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Findings of the Association for Computational Linguistics: NAACL 2024.\n \n \n \n \n\n\n \n Kevin Duh; Helena Gomez; and Steven Bethard.,\n editors.\n \n\n\n \n\n\n\n Association for Computational Linguistics. Mexico City, Mexico, June 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@proceedings{findings-2024-findings-association,\n title = "Findings of the Association for Computational Linguistics: NAACL 2024",\n editor = "Duh, Kevin and\n Gomez, Helena and\n Bethard, Steven",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.findings-naacl.0",\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers).\n \n \n \n \n\n\n \n Kevin Duh; Helena Gomez; and Steven Bethard.,\n editors.\n \n\n\n \n\n\n\n Association for Computational Linguistics. Mexico City, Mexico, June 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@proceedings{naacl-2024-long,\n title = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)",\n editor = "Duh, Kevin and\n Gomez, Helena and\n Bethard, Steven",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.naacl-long.0",\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers).\n \n \n \n \n\n\n \n Kevin Duh; Helena Gomez; and Steven Bethard.,\n editors.\n \n\n\n \n\n\n\n Association for Computational Linguistics. Mexico City, Mexico, June 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@proceedings{naacl-2024-short,\n title = "Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers)",\n editor = "Duh, Kevin and\n Gomez, Helena and\n Bethard, Steven",\n month = jun,\n year = "2024",\n address = "Mexico City, Mexico",\n publisher = "Association for Computational Linguistics",\n url = "https://aclanthology.org/2024.naacl-short.0",\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Machine Learning and Deep Learning Algorithms.\n \n \n \n \n\n\n \n Steven Bethard.\n\n\n \n\n\n\n In Hua Xu; and Dina Demner Fushman., editor(s),
Natural Language Processing in Biomedicine: A Practical Guide, pages 43–76. Springer International Publishing, Cham, 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 4 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n\n\n\n
\n
@incollection{Bethard2024,\nauthor="Bethard, Steven",\neditor="Xu, Hua\nand Demner Fushman, Dina",\ntitle="Machine Learning and Deep Learning Algorithms",\nbookTitle="Natural Language Processing in Biomedicine: A Practical Guide",\nyear="2024",\npublisher="Springer International Publishing",\naddress="Cham",\npages="43--76",\nisbn="978-3-031-55865-8",\ndoi="10.1007/978-3-031-55865-8_3",\nurl="https://doi.org/10.1007/978-3-031-55865-8_3",\nkeywords={machine learning},\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n A survey on geocoding: algorithms and datasets for toponym resolution.\n \n \n \n \n\n\n \n Zeyu Zhang; and Steven Bethard.\n\n\n \n\n\n\n
Language Resources and Evaluation. June 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 2 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@article{zhang_survey_2024,\n\ttitle = {A survey on geocoding: algorithms and datasets for toponym resolution},\n\tissn = {1574-0218},\n\tshorttitle = {A survey on geocoding},\n\turl = {https://doi.org/10.1007/s10579-024-09730-2},\n\tdoi = {10.1007/s10579-024-09730-2},\n\tlanguage = {en},\n\turldate = {2024-06-12},\n\tjournal = {Language Resources and Evaluation},\n\tauthor = {Zhang, Zeyu and Bethard, Steven},\n\tmonth = jun,\n\tyear = {2024},\n\tkeywords = {Geocoding, Geographical entity normalization, Toponym resolution},\n}\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Examining the Dynamics of Uncivil Discourse Between Sub-National Political Officials and the Public on Twitter.\n \n \n \n \n\n\n \n Juliana L. Barbati; Stephen A. Rains; Kate Kenski; Yotam Shmargad; Steven Bethard; and Kevin Coe.\n\n\n \n\n\n\n
Mass Communication and Society, 0(0): 1-20. February 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n\n\n\n
\n
@article{barbati-etal-2024-examining,\nauthor = {Juliana L. Barbati and Stephen A. Rains and Kate Kenski and Yotam Shmargad and Steven Bethard and Kevin Coe},\ntitle = {Examining the Dynamics of Uncivil Discourse Between Sub-National Political Officials and the Public on Twitter},\njournal = {Mass Communication and Society},\nvolume = {0},\nnumber = {0},\npages = {1-20},\nmonth = feb,\nyear = {2024},\npublisher = {Routledge},\nurl = {https://doi.org/10.1080/15205436.2024.2313095},\nkeywords = {social media, civility},\n}\n
\n
\n\n\n\n
\n\n\n\n\n\n