\n \n \n
\n
\n\n \n \n \n \n \n Invariance & Causal Representation Learning: Prospects and Limitations.\n \n \n \n\n\n \n Bing, S.; Hochsprung, T.; Wahl, J.; Ninad, U.; and Runge, J.\n\n\n \n\n\n\n
Transactions of Machine Learning Research (TMLR). ISSN: 2835-8856., (https://openreview.net/forum?id=lpOC6s4B). 2024.\n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{bing_invariance_2024,\n\ttitle = {Invariance \\& {Causal} {Representation} {Learning}: {Prospects} and {Limitations}},\n\tnumber = {https://openreview.net/forum?id=lpOC6s4B},\n\tjournal = {Transactions of Machine Learning Research (TMLR). ISSN: 2835-8856.},\n\tauthor = {Bing, Simon and Hochsprung, Tom and Wahl, Jonas and Ninad, Urmi and Runge, Jakob},\n\tyear = {2024},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Identifying Linearly-Mixed Causal Representations from Multi-Node Interventions.\n \n \n \n\n\n \n Bing, S.; Ninad, U.; Wahl, J.; and Runge, J.\n\n\n \n\n\n\n In
Causal Learning and Reasoning, pages 843–867, 2024. PMLR\n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{bing_identifying_2024,\n\ttitle = {Identifying {Linearly}-{Mixed} {Causal} {Representations} from {Multi}-{Node} {Interventions}},\n\tbooktitle = {Causal {Learning} and {Reasoning}},\n\tpublisher = {PMLR},\n\tauthor = {Bing, Simon and Ninad, Urmi and Wahl, Jonas and Runge, Jakob},\n\tyear = {2024},\n\tpages = {843--867},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Foundations of causal discovery on groups of variables.\n \n \n \n\n\n \n Wahl, J.; Ninad, U.; and Runge, J.\n\n\n \n\n\n\n
Journal of Causal Inference, 12(1): 20230041. 2024.\n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{wahl_foundations_2024,\n\ttitle = {Foundations of causal discovery on groups of variables},\n\tvolume = {12},\n\tnumber = {1},\n\tjournal = {Journal of Causal Inference},\n\tauthor = {Wahl, Jonas and Ninad, Urmi and Runge, Jakob},\n\tyear = {2024},\n\tpages = {20230041},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Causal Inference for Spatial Data Analytics (Dagstuhl Seminar 24202).\n \n \n \n \n\n\n \n Tomko, M.; Xin, Y.; and Wahl, J.\n\n\n \n\n\n\n
Dagstuhl Reports, 14(5): 25–57. 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{tomko_causal_2024,\n\ttitle = {Causal {Inference} for {Spatial} {Data} {Analytics} ({Dagstuhl} {Seminar} 24202)},\n\tvolume = {14},\n\tissn = {2192-5283},\n\turl = {https://drops.dagstuhl.de/entities/document/10.4230/DagRep.14.5.25},\n\tdoi = {10.4230/DagRep.14.5.25},\n\tnumber = {5},\n\tjournal = {Dagstuhl Reports},\n\tauthor = {Tomko, Martin and Xin, Yanan and Wahl, Jonas},\n\teditor = {Tomko, Martin and Xin, Yanan and Wahl, Jonas},\n\tyear = {2024},\n\tpages = {25--57},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n MMAR: Multilingual and Multimodal Anaphora Resolution in Instructional Videos.\n \n \n \n \n\n\n \n Oguz, C.; Denis, P.; Ostermann, S.; Vincent, E.; Skachkova, N.; and Genabith, J. V.\n\n\n \n\n\n\n In Al-Onaizan, Y.; Bansal, M.; and Chen, Y., editor(s),
Findings of the Association for Computational Linguistics: EMNLP 2024, pages 1618–1633, Miami, Florida, USA, November 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{oguz_mmar_2024,\n\taddress = {Miami, Florida, USA},\n\ttitle = {{MMAR}: {Multilingual} and {Multimodal} {Anaphora} {Resolution} in {Instructional} {Videos}},\n\tshorttitle = {{MMAR}},\n\turl = {https://aclanthology.org/2024.findings-emnlp.88/},\n\tdoi = {10.18653/v1/2024.findings-emnlp.88},\n\tabstract = {Multilingual anaphora resolution identifies referring expressions and implicit arguments in texts and links to antecedents that cover several languages. In the most challenging setting, cross-lingual anaphora resolution, training data, and test data are in different languages. As knowledge needs to be transferred across languages, this task is challenging, both in the multilingual and cross-lingual setting. We hypothesize that one way to alleviate some of the difficulty of the task is to include multimodal information in the form of images (i.e. frames extracted from instructional videos). Such visual inputs are by nature language agnostic, therefore cross- and multilingual anaphora resolution should benefit from visual information. In this paper, we provide the first multilingual and multimodal dataset annotated with anaphoric relations and present experimental results for end-to-end multimodal and multilingual anaphora resolution. Given gold mentions, multimodal features improve anaphora resolution results by {\\textbackslash}textbackslashtextasciitilde10 \\% for unseen languages.},\n\turldate = {2025-05-22},\n\tbooktitle = {Findings of the {Association} for {Computational} {Linguistics}: {EMNLP} 2024},\n\tpublisher = {Association for Computational Linguistics},\n\tauthor = {Oguz, Cennet and Denis, Pascal and Ostermann, Simon and Vincent, Emmanuel and Skachkova, Natalia and Genabith, Josef Van},\n\teditor = {Al-Onaizan, Yaser and Bansal, Mohit and Chen, Yun-Nung},\n\tmonth = nov,\n\tyear = {2024},\n\tpages = {1618--1633},\n}\n\n\n\n
\n
\n\n\n
\n Multilingual anaphora resolution identifies referring expressions and implicit arguments in texts and links to antecedents that cover several languages. In the most challenging setting, cross-lingual anaphora resolution, training data, and test data are in different languages. As knowledge needs to be transferred across languages, this task is challenging, both in the multilingual and cross-lingual setting. We hypothesize that one way to alleviate some of the difficulty of the task is to include multimodal information in the form of images (i.e. frames extracted from instructional videos). Such visual inputs are by nature language agnostic, therefore cross- and multilingual anaphora resolution should benefit from visual information. In this paper, we provide the first multilingual and multimodal dataset annotated with anaphoric relations and present experimental results for end-to-end multimodal and multilingual anaphora resolution. Given gold mentions, multimodal features improve anaphora resolution results by \\textbackslashtextasciitilde10 % for unseen languages.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n A Comparison of Different Tokenization Methods for the Georgian Language.\n \n \n \n \n\n\n \n Mikaberidze, B.; Saghinadze, T.; Mikaberidze, G.; Kalandadze, R.; Pkhakadze, K.; van Genabith, J.; Ostermann, S.; van der Plas, L.; and Müller, P.\n\n\n \n\n\n\n In Abbas, M.; and Freihat, A. A., editor(s),
Proceedings of the 7th International Conference on Natural Language and Speech Processing (ICNLSP 2024), pages 199–208, Trento, October 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{mikaberidze_comparison_2024,\n\taddress = {Trento},\n\ttitle = {A {Comparison} of {Different} {Tokenization} {Methods} for the {Georgian} {Language}},\n\turl = {https://aclanthology.org/2024.icnlsp-1.22/},\n\turldate = {2025-05-22},\n\tbooktitle = {Proceedings of the 7th {International} {Conference} on {Natural} {Language} and {Speech} {Processing} ({ICNLSP} 2024)},\n\tpublisher = {Association for Computational Linguistics},\n\tauthor = {Mikaberidze, Beso and Saghinadze, Temo and Mikaberidze, Guram and Kalandadze, Raphael and Pkhakadze, Konstantine and van Genabith, Josef and Ostermann, Simon and van der Plas, Lonneke and Müller, Philipp},\n\teditor = {Abbas, Mourad and Freihat, Abed Alhakim},\n\tmonth = oct,\n\tyear = {2024},\n\tpages = {199--208},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Adapting Multilingual LLMs to Low-Resource Languages with Knowledge Graphs via Adapters.\n \n \n \n \n\n\n \n Gurgurov, D.; Hartmann, M.; and Ostermann, S.\n\n\n \n\n\n\n In Biswas, R.; Kaffee, L.; Agarwal, O.; Minervini, P.; Singh, S.; and de Melo, G., editor(s),
Proceedings of the 1st Workshop on Knowledge Graphs and Large Language Models (KaLLM 2024), pages 63–74, Bangkok, Thailand, August 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{gurgurov_adapting_2024,\n\taddress = {Bangkok, Thailand},\n\ttitle = {Adapting {Multilingual} {LLMs} to {Low}-{Resource} {Languages} with {Knowledge} {Graphs} via {Adapters}},\n\turl = {https://aclanthology.org/2024.kallm-1.7/},\n\tdoi = {10.18653/v1/2024.kallm-1.7},\n\tabstract = {This paper explores the integration of graph knowledge from linguistic ontologies into multilingual Large Language Models (LLMs) using adapters to improve performance for low-resource languages (LRLs) in sentiment analysis (SA) and named entity recognition (NER). Building upon successful parameter-efficient fine-tuning techniques, such as K-ADAPTER and MAD-X, we propose a similar approach for incorporating knowledge from multilingual graphs, connecting concepts in various languages with each other through linguistic relationships, into multilingual LLMs for LRLs. Specifically, we focus on eight LRLs — Maltese, Bulgarian, Indonesian, Nepali, Javanese, Uyghur, Tibetan, and Sinhala — and employ language-specific adapters fine-tuned on data extracted from the language-specific section of ConceptNet, aiming to enable knowledge transfer across the languages covered by the knowledge graph. We compare various fine-tuning objectives, including standard Masked Language Modeling (MLM), MLM with full-word masking, and MLM with targeted masking, to analyze their effectiveness in learning and integrating the extracted graph data. Through empirical evaluation on language-specific tasks, we assess how structured graph knowledge affects the performance of multilingual LLMs for LRLs in SA and NER, providing insights into the potential benefits of adapting language models for low-resource scenarios.},\n\turldate = {2025-05-22},\n\tbooktitle = {Proceedings of the 1st {Workshop} on {Knowledge} {Graphs} and {Large} {Language} {Models} ({KaLLM} 2024)},\n\tpublisher = {Association for Computational Linguistics},\n\tauthor = {Gurgurov, Daniil and Hartmann, Mareike and Ostermann, Simon},\n\teditor = {Biswas, Russa and Kaffee, Lucie-Aimée and Agarwal, Oshin and Minervini, Pasquale and Singh, Sameer and de Melo, Gerard},\n\tmonth = aug,\n\tyear = {2024},\n\tpages = {63--74},\n}\n\n\n\n
\n
\n\n\n
\n This paper explores the integration of graph knowledge from linguistic ontologies into multilingual Large Language Models (LLMs) using adapters to improve performance for low-resource languages (LRLs) in sentiment analysis (SA) and named entity recognition (NER). Building upon successful parameter-efficient fine-tuning techniques, such as K-ADAPTER and MAD-X, we propose a similar approach for incorporating knowledge from multilingual graphs, connecting concepts in various languages with each other through linguistic relationships, into multilingual LLMs for LRLs. Specifically, we focus on eight LRLs — Maltese, Bulgarian, Indonesian, Nepali, Javanese, Uyghur, Tibetan, and Sinhala — and employ language-specific adapters fine-tuned on data extracted from the language-specific section of ConceptNet, aiming to enable knowledge transfer across the languages covered by the knowledge graph. We compare various fine-tuning objectives, including standard Masked Language Modeling (MLM), MLM with full-word masking, and MLM with targeted masking, to analyze their effectiveness in learning and integrating the extracted graph data. Through empirical evaluation on language-specific tasks, we assess how structured graph knowledge affects the performance of multilingual LLMs for LRLs in SA and NER, providing insights into the potential benefits of adapting language models for low-resource scenarios.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n HybridBERT - Making BERT Pretraining More Efficient Through Hybrid Mixture of Attention Mechanisms.\n \n \n \n \n\n\n \n Srinivasagan, G.; and Ostermann, S.\n\n\n \n\n\n\n In Cao, Y. (.; Papadimitriou, I.; Ovalle, A.; Zampieri, M.; Ferraro, F.; and Swayamdipta, S., editor(s),
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 4: Student Research Workshop), pages 285–291, Mexico City, Mexico, June 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{srinivasagan_hybridbert_2024,\n\taddress = {Mexico City, Mexico},\n\ttitle = {{HybridBERT} - {Making} {BERT} {Pretraining} {More} {Efficient} {Through} {Hybrid} {Mixture} of {Attention} {Mechanisms}},\n\turl = {https://aclanthology.org/2024.naacl-srw.30/},\n\tdoi = {10.18653/v1/2024.naacl-srw.30},\n\tabstract = {Pretrained transformer-based language models have produced state-of-the-art performance in most natural language understanding tasks. These models undergo two stages of training: pretraining on a huge corpus of data and fine-tuning on a specific downstream task. The pretraining phase is extremely compute-intensive and requires several high-performance computing devices like GPUs and several days or even months of training, but it is crucial for the model to capture global knowledge and also has a significant impact on the fine-tuning task. This is a major roadblock for researchers without access to sophisticated computing resources. To overcome this challenge, we propose two novel hybrid architectures called HybridBERT (HBERT), which combine self-attention and additive attention mechanisms together with sub-layer normalization. We introduce a computing budget to the pretraining phase, limiting the training time and usage to a single GPU. We show that HBERT attains twice the pretraining accuracy of a vanilla-BERT baseline. We also evaluate our proposed models on two downstream tasks, where we outperform BERT-base while accelerating inference. Moreover, we study the effect of weight initialization with a limited pretraining budget. The code and models are publicly available at: www.github.com/gokulsg/HBERT/.},\n\turldate = {2025-05-22},\n\tbooktitle = {Proceedings of the 2024 {Conference} of the {North} {American} {Chapter} of the {Association} for {Computational} {Linguistics}: {Human} {Language} {Technologies} ({Volume} 4: {Student} {Research} {Workshop})},\n\tpublisher = {Association for Computational Linguistics},\n\tauthor = {Srinivasagan, Gokul and Ostermann, Simon},\n\teditor = {Cao, Yang (Trista) and Papadimitriou, Isabel and Ovalle, Anaelia and Zampieri, Marcos and Ferraro, Francis and Swayamdipta, Swabha},\n\tmonth = jun,\n\tyear = {2024},\n\tpages = {285--291},\n}\n\n\n\n
\n
\n\n\n
\n Pretrained transformer-based language models have produced state-of-the-art performance in most natural language understanding tasks. These models undergo two stages of training: pretraining on a huge corpus of data and fine-tuning on a specific downstream task. The pretraining phase is extremely compute-intensive and requires several high-performance computing devices like GPUs and several days or even months of training, but it is crucial for the model to capture global knowledge and also has a significant impact on the fine-tuning task. This is a major roadblock for researchers without access to sophisticated computing resources. To overcome this challenge, we propose two novel hybrid architectures called HybridBERT (HBERT), which combine self-attention and additive attention mechanisms together with sub-layer normalization. We introduce a computing budget to the pretraining phase, limiting the training time and usage to a single GPU. We show that HBERT attains twice the pretraining accuracy of a vanilla-BERT baseline. We also evaluate our proposed models on two downstream tasks, where we outperform BERT-base while accelerating inference. Moreover, we study the effect of weight initialization with a limited pretraining budget. The code and models are publicly available at: www.github.com/gokulsg/HBERT/.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n CoXQL: A Dataset for Parsing Explanation Requests in Conversational XAI Systems.\n \n \n \n \n\n\n \n Wang, Q.; Anikina, T.; Feldhus, N.; Ostermann, S.; and Möller, S.\n\n\n \n\n\n\n In Al-Onaizan, Y.; Bansal, M.; and Chen, Y., editor(s),
Findings of the Association for Computational Linguistics: EMNLP 2024, pages 1410–1422, Miami, Florida, USA, November 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{wang_coxql_2024,\n\taddress = {Miami, Florida, USA},\n\ttitle = {{CoXQL}: {A} {Dataset} for {Parsing} {Explanation} {Requests} in {Conversational} {XAI} {Systems}},\n\tshorttitle = {{CoXQL}},\n\turl = {https://aclanthology.org/2024.findings-emnlp.76/},\n\tdoi = {10.18653/v1/2024.findings-emnlp.76},\n\tabstract = {Conversational explainable artificial intelligence (ConvXAI) systems based on large language models (LLMs) have garnered significant interest from the research community in natural language processing (NLP) and human-computer interaction (HCI). Such systems can provide answers to user questions about explanations in dialogues, have the potential to enhance users' comprehension and offer more information about the decision-making and generation processes of LLMs. Currently available ConvXAI systems are based on intent recognition rather than free chat, as this has been found to be more precise and reliable in identifying users' intentions. However, the recognition of intents still presents a challenge in the case of ConvXAI, since little training data exist and the domain is highly specific, as there is a broad range of XAI methods to map requests onto. In order to bridge this gap, we present CoXQL, the first dataset in the NLP domain for user intent recognition in ConvXAI, covering 31 intents, seven of which require filling multiple slots. Subsequently, we enhance an existing parsing approach by incorporating template validations, and conduct an evaluation of several LLMs on CoXQL using different parsing strategies. We conclude that the improved parsing approach (MP+) surpasses the performance of previous approaches. We also discover that intents with multiple slots remain highly challenging for LLMs.},\n\turldate = {2025-05-22},\n\tbooktitle = {Findings of the {Association} for {Computational} {Linguistics}: {EMNLP} 2024},\n\tpublisher = {Association for Computational Linguistics},\n\tauthor = {Wang, Qianli and Anikina, Tatiana and Feldhus, Nils and Ostermann, Simon and Möller, Sebastian},\n\teditor = {Al-Onaizan, Yaser and Bansal, Mohit and Chen, Yun-Nung},\n\tmonth = nov,\n\tyear = {2024},\n\tpages = {1410--1422},\n}\n\n\n\n
\n
\n\n\n
\n Conversational explainable artificial intelligence (ConvXAI) systems based on large language models (LLMs) have garnered significant interest from the research community in natural language processing (NLP) and human-computer interaction (HCI). Such systems can provide answers to user questions about explanations in dialogues, have the potential to enhance users' comprehension and offer more information about the decision-making and generation processes of LLMs. Currently available ConvXAI systems are based on intent recognition rather than free chat, as this has been found to be more precise and reliable in identifying users' intentions. However, the recognition of intents still presents a challenge in the case of ConvXAI, since little training data exist and the domain is highly specific, as there is a broad range of XAI methods to map requests onto. In order to bridge this gap, we present CoXQL, the first dataset in the NLP domain for user intent recognition in ConvXAI, covering 31 intents, seven of which require filling multiple slots. Subsequently, we enhance an existing parsing approach by incorporating template validations, and conduct an evaluation of several LLMs on CoXQL using different parsing strategies. We conclude that the improved parsing approach (MP+) surpasses the performance of previous approaches. We also discover that intents with multiple slots remain highly challenging for LLMs.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Common European Language Data Space.\n \n \n \n \n\n\n \n Rehm, G.; Piperidis, S.; Choukri, K.; Vasiļjevs, A.; Marheinecke, K.; Arranz, V.; Bērziņš, A.; Deligiannis, M.; Galanis, D.; Giagkou, M.; Gkirtzou, K.; Gkoumas, D.; Grützner-Zahn, A.; Kolovou, A.; Labropoulou, P.; Lagzdiņš, A.; Leitner, E.; Mapelli, V.; Mazo, H.; Ostermann, S.; Racioppa, S.; Rigault, M.; and Voukoutis, L.\n\n\n \n\n\n\n In Calzolari, N.; Kan, M.; Hoste, V.; Lenci, A.; Sakti, S.; and Xue, N., editor(s),
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pages 3579–3586, Torino, Italia, May 2024. ELRA and ICCL\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{rehm_common_2024,\n\taddress = {Torino, Italia},\n\ttitle = {Common {European} {Language} {Data} {Space}},\n\turl = {https://aclanthology.org/2024.lrec-main.317/},\n\tabstract = {The Common European Language Data Space (LDS) is an integral part of the EU data strategy, which aims at developing a single market for data. Its decentralised technical infrastructure and governance scheme are currently being developed by the LDS project, which also has dedicated tasks for proof-of-concept prototypes, handling legal aspects, raising awareness and promoting the LDS through events and social media channels. The LDS is part of a broader vision for establishing all necessary components to develop European large language models.},\n\turldate = {2025-05-22},\n\tbooktitle = {Proceedings of the 2024 {Joint} {International} {Conference} on {Computational} {Linguistics}, {Language} {Resources} and {Evaluation} ({LREC}-{COLING} 2024)},\n\tpublisher = {ELRA and ICCL},\n\tauthor = {Rehm, Georg and Piperidis, Stelios and Choukri, Khalid and Vasiļjevs, Andrejs and Marheinecke, Katrin and Arranz, Victoria and Bērziņš, Aivars and Deligiannis, Miltos and Galanis, Dimitris and Giagkou, Maria and Gkirtzou, Katerina and Gkoumas, Dimitris and Grützner-Zahn, Annika and Kolovou, Athanasia and Labropoulou, Penny and Lagzdiņš, Andis and Leitner, Elena and Mapelli, Valérie and Mazo, Hélène and Ostermann, Simon and Racioppa, Stefania and Rigault, Mickaël and Voukoutis, Leon},\n\teditor = {Calzolari, Nicoletta and Kan, Min-Yen and Hoste, Veronique and Lenci, Alessandro and Sakti, Sakriani and Xue, Nianwen},\n\tmonth = may,\n\tyear = {2024},\n\tpages = {3579--3586},\n}\n\n\n\n
\n
\n\n\n
\n The Common European Language Data Space (LDS) is an integral part of the EU data strategy, which aims at developing a single market for data. Its decentralised technical infrastructure and governance scheme are currently being developed by the LDS project, which also has dedicated tasks for proof-of-concept prototypes, handling legal aspects, raising awareness and promoting the LDS through events and social media channels. The LDS is part of a broader vision for establishing all necessary components to develop European large language models.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Effective Human Oversight of AI-Based Systems: A Signal Detection Perspective on the Detection of Inaccurate and Unfair Outputs.\n \n \n \n\n\n \n Langer, M.; Baum, K.; and Schlicker, N.\n\n\n \n\n\n\n
Minds and Machines, 35(1): 1–30. 2024.\n
\n\n
\n\n
\n\n
\n\n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{langer_effective_2024,\n\ttitle = {Effective {Human} {Oversight} of {AI}-{Based} {Systems}: {A} {Signal} {Detection} {Perspective} on the {Detection} of {Inaccurate} and {Unfair} {Outputs}},\n\tvolume = {35},\n\tdoi = {10.1007/s11023-024-09701-0},\n\tnumber = {1},\n\tjournal = {Minds and Machines},\n\tauthor = {Langer, Markus and Baum, Kevin and Schlicker, Nadine},\n\tyear = {2024},\n\tpages = {1--30},\n}\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Acting for the Right Reasons: Creating Reason-Sensitive Artificial Moral Agents.\n \n \n \n \n\n\n \n Baum, K.; Dargasz, L.; Jahn, F.; Gros, T. P.; and Wolf, V.\n\n\n \n\n\n\n In 2024. \n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n \n \n 1 download\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{baum_acting_2024,\n\ttitle = {Acting for the {Right} {Reasons}: {Creating} {Reason}-{Sensitive} {Artificial} {Moral} {Agents}},\n\turl = {https://openreview.net/forum?id=Mgyr7EY7X9&referrer=%5Bthe%20profile%20of%20Kevin%20Baum%5D(%2Fprofile%3Fid%3D~Kevin_Baum1)},\n\tdoi = {https://doi.org/10.48550/arXiv.2409.15014},\n\tauthor = {Baum, Kevin and Dargasz, Lisa and Jahn, Felix and Gros, Timo P. and Wolf, Verena},\n\tyear = {2024},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Divergent Token Metrics: Measuring degradation to prune away LLM components – and optimize quantization.\n \n \n \n \n\n\n \n Deiseroth, B.; Meuer, M.; Gritsch, N.; Eichenberg, C.; Schramowski, P.; Aßenmacher, M.; and Kersting, K.\n\n\n \n\n\n\n In
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics, 2024. arXiv\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@inproceedings{deiseroth_divergent_2024,\n\ttitle = {Divergent {Token} {Metrics}: {Measuring} degradation to prune away {LLM} components -- and optimize quantization},\n\tcopyright = {Creative Commons Attribution 4.0 International},\n\tshorttitle = {Divergent {Token} {Metrics}},\n\turl = {https://arxiv.org/abs/2311.01544},\n\tdoi = {10.48550/ARXIV.2311.01544},\n\tabstract = {Large Language Models (LLMs) have reshaped natural language processing with their impressive capabilities. However, their ever-increasing size has raised concerns about their effective deployment and the need for LLM compression. This study introduces the Divergent Token Metrics (DTMs), a novel approach to assessing compressed LLMs, addressing the limitations of traditional perplexity or accuracy measures that fail to accurately reflect text generation quality. DTMs measure token divergences that allow deeper insights into the subtleties of model compression, in particular, when evaluating components' impacts individually. Utilizing the First Divergent Token Metric (FDTM) in model sparsification reveals that 25\\% of all attention components can be pruned beyond 90\\% on the Llama-2 model family, still keeping SOTA performance. For quantization, FDTM suggests that more than 80\\% of parameters can be naively transformed to int8 without special outlier management. These evaluations indicate the necessity of choosing appropriate compressions for parameters individually -- and that FDTM can identify those -- while standard metrics result in deteriorated outcomes.},\n\turldate = {2025-05-21},\n\tbooktitle = {Proceedings of the 2024 {Conference} of the {North} {American} {Chapter} of the {Association} for {Computational} {Linguistics}},\n\tpublisher = {arXiv},\n\tauthor = {Deiseroth, Björn and Meuer, Max and Gritsch, Nikolas and Eichenberg, Constantin and Schramowski, Patrick and Aßenmacher, Matthias and Kersting, Kristian},\n\tyear = {2024},\n\tkeywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, Machine Learning (cs.LG)},\n}\n\n\n\n
\n
\n\n\n
\n Large Language Models (LLMs) have reshaped natural language processing with their impressive capabilities. However, their ever-increasing size has raised concerns about their effective deployment and the need for LLM compression. This study introduces the Divergent Token Metrics (DTMs), a novel approach to assessing compressed LLMs, addressing the limitations of traditional perplexity or accuracy measures that fail to accurately reflect text generation quality. DTMs measure token divergences that allow deeper insights into the subtleties of model compression, in particular, when evaluating components' impacts individually. Utilizing the First Divergent Token Metric (FDTM) in model sparsification reveals that 25% of all attention components can be pruned beyond 90% on the Llama-2 model family, still keeping SOTA performance. For quantization, FDTM suggests that more than 80% of parameters can be naively transformed to int8 without special outlier management. These evaluations indicate the necessity of choosing appropriate compressions for parameters individually – and that FDTM can identify those – while standard metrics result in deteriorated outcomes.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n LLMs Lost in Translation: M-ALERT uncovers Cross-Linguistic Safety Gaps.\n \n \n \n \n\n\n \n Friedrich, F.; Tedeschi, S.; Schramowski, P.; Brack, M.; Navigli, R.; Nguyen, H.; Li, B.; and Kersting, K.\n\n\n \n\n\n\n 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n\n\n\n
\n
@misc{friedrich_llms_2024,\n\ttitle = {{LLMs} {Lost} in {Translation}: {M}-{ALERT} uncovers {Cross}-{Linguistic} {Safety} {Gaps}},\n\tcopyright = {Creative Commons Attribution 4.0 International},\n\tshorttitle = {{LLMs} {Lost} in {Translation}},\n\turl = {https://arxiv.org/abs/2412.15035},\n\tdoi = {10.48550/ARXIV.2412.15035},\n\tabstract = {Building safe Large Language Models (LLMs) across multiple languages is essential in ensuring both safe access and linguistic diversity. To this end, we introduce M-ALERT, a multilingual benchmark that evaluates the safety of LLMs in five languages: English, French, German, Italian, and Spanish. M-ALERT includes 15k high-quality prompts per language, totaling 75k, following the detailed ALERT taxonomy. Our extensive experiments on 10 state-of-the-art LLMs highlight the importance of language-specific safety analysis, revealing that models often exhibit significant inconsistencies in safety across languages and categories. For instance, Llama3.2 shows high unsafety in the category crime\\_tax for Italian but remains safe in other languages. Similar differences can be observed across all models. In contrast, certain categories, such as substance\\_cannabis and crime\\_propaganda, consistently trigger unsafe responses across models and languages. These findings underscore the need for robust multilingual safety practices in LLMs to ensure safe and responsible usage across diverse user communities.},\n\turldate = {2025-05-13},\n\tpublisher = {arXiv},\n\tauthor = {Friedrich, Felix and Tedeschi, Simone and Schramowski, Patrick and Brack, Manuel and Navigli, Roberto and Nguyen, Huu and Li, Bo and Kersting, Kristian},\n\tyear = {2024},\n\tkeywords = {Computation and Language (cs.CL), FOS: Computer and information sciences},\n}\n\n\n\n
\n
\n\n\n
\n Building safe Large Language Models (LLMs) across multiple languages is essential in ensuring both safe access and linguistic diversity. To this end, we introduce M-ALERT, a multilingual benchmark that evaluates the safety of LLMs in five languages: English, French, German, Italian, and Spanish. M-ALERT includes 15k high-quality prompts per language, totaling 75k, following the detailed ALERT taxonomy. Our extensive experiments on 10 state-of-the-art LLMs highlight the importance of language-specific safety analysis, revealing that models often exhibit significant inconsistencies in safety across languages and categories. For instance, Llama3.2 shows high unsafety in the category crime_tax for Italian but remains safe in other languages. Similar differences can be observed across all models. In contrast, certain categories, such as substance_cannabis and crime_propaganda, consistently trigger unsafe responses across models and languages. These findings underscore the need for robust multilingual safety practices in LLMs to ensure safe and responsible usage across diverse user communities.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n DeiSAM: Segment Anything with Deictic Prompting.\n \n \n \n\n\n \n Shindo, H.; Brack, M.; Sudhakaran, G.; Dhami, D. S.; Schramowski, P.; and Kersting, K.\n\n\n \n\n\n\n In volume 37, pages 52266–52295, 2024. \n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{shindo_deisam_2024,\n\ttitle = {{DeiSAM}: {Segment} {Anything} with {Deictic} {Prompting}},\n\tvolume = {37},\n\tauthor = {Shindo, Hikaru and Brack, Manuel and Sudhakaran, Gopika and Dhami, Devendra Singh and Schramowski, Patrick and Kersting, Kristian},\n\tyear = {2024},\n\tpages = {52266--52295},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n SCAR: Sparse Conditioned Autoencoders for Concept Detection and Steering in LLMs.\n \n \n \n \n\n\n \n Härle, R.; Friedrich, F.; Brack, M.; Deiseroth, B.; Schramowski, P.; and Kersting, K.\n\n\n \n\n\n\n 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n\n\n\n
\n
@misc{harle_scar_2024,\n\ttitle = {{SCAR}: {Sparse} {Conditioned} {Autoencoders} for {Concept} {Detection} and {Steering} in {LLMs}},\n\tcopyright = {Creative Commons Attribution Non Commercial Share Alike 4.0 International},\n\tshorttitle = {{SCAR}},\n\turl = {https://arxiv.org/abs/2411.07122},\n\tdoi = {10.48550/ARXIV.2411.07122},\n\tabstract = {Large Language Models (LLMs) have demonstrated remarkable capabilities in generating human-like text, but their output may not be aligned with the user or even produce harmful content. This paper presents a novel approach to detect and steer concepts such as toxicity before generation. We introduce the Sparse Conditioned Autoencoder (SCAR), a single trained module that extends the otherwise untouched LLM. SCAR ensures full steerability, towards and away from concepts (e.g., toxic content), without compromising the quality of the model's text generation on standard evaluation benchmarks. We demonstrate the effective application of our approach through a variety of concepts, including toxicity, safety, and writing style alignment. As such, this work establishes a robust framework for controlling LLM generations, ensuring their ethical and safe deployment in real-world applications.},\n\turldate = {2025-05-13},\n\tpublisher = {arXiv},\n\tauthor = {Härle, Ruben and Friedrich, Felix and Brack, Manuel and Deiseroth, Björn and Schramowski, Patrick and Kersting, Kristian},\n\tyear = {2024},\n\tkeywords = {Computation and Language (cs.CL), FOS: Computer and information sciences},\n}\n\n\n\n
\n
\n\n\n
\n Large Language Models (LLMs) have demonstrated remarkable capabilities in generating human-like text, but their output may not be aligned with the user or even produce harmful content. This paper presents a novel approach to detect and steer concepts such as toxicity before generation. We introduce the Sparse Conditioned Autoencoder (SCAR), a single trained module that extends the otherwise untouched LLM. SCAR ensures full steerability, towards and away from concepts (e.g., toxic content), without compromising the quality of the model's text generation on standard evaluation benchmarks. We demonstrate the effective application of our approach through a variety of concepts, including toxicity, safety, and writing style alignment. As such, this work establishes a robust framework for controlling LLM generations, ensuring their ethical and safe deployment in real-world applications.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Occiglot at WMT24: European Open-source Large Language Models Evaluated on Translation.\n \n \n \n \n\n\n \n Avramidis, E.; Grützner-Zahn, A.; Brack, M.; Schramowski, P.; Ortiz Suarez, P.; Ostendorff, M.; Barth, F.; Manakhimova, S.; Macketanz, V.; Rehm, G.; and Kersting, K.\n\n\n \n\n\n\n In
Proceedings of the Ninth Conference on Machine Translation, pages 292–298, Miami, Florida, USA, 2024. Association for Computational Linguistics\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{avramidis_occiglot_2024,\n\taddress = {Miami, Florida, USA},\n\ttitle = {Occiglot at {WMT24}: {European} {Open}-source {Large} {Language} {Models} {Evaluated} on {Translation}},\n\tshorttitle = {Occiglot at {WMT24}},\n\turl = {https://aclanthology.org/2024.wmt-1.23},\n\tdoi = {10.18653/v1/2024.wmt-1.23},\n\tlanguage = {en},\n\turldate = {2025-05-13},\n\tbooktitle = {Proceedings of the {Ninth} {Conference} on {Machine} {Translation}},\n\tpublisher = {Association for Computational Linguistics},\n\tauthor = {Avramidis, Eleftherios and Grützner-Zahn, Annika and Brack, Manuel and Schramowski, Patrick and Ortiz Suarez, Pedro and Ostendorff, Malte and Barth, Fabio and Manakhimova, Shushen and Macketanz, Vivien and Rehm, Georg and Kersting, Kristian},\n\tyear = {2024},\n\tpages = {292--298},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n\n\n \n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Auditing and instructing text-to-image generation models on fairness.\n \n \n \n \n\n\n \n Friedrich, F.; Brack, M.; Struppek, L.; Hintersdorf, D.; Schramowski, P.; Luccioni, S.; and Kersting, K.\n\n\n \n\n\n\n
AI and Ethics. August 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{friedrich_auditing_2024,\n\ttitle = {Auditing and instructing text-to-image generation models on fairness},\n\tissn = {2730-5953, 2730-5961},\n\turl = {https://link.springer.com/10.1007/s43681-024-00531-5},\n\tdoi = {10.1007/s43681-024-00531-5},\n\tabstract = {Abstract \n \n Generative AI models have recently achieved astonishing results in quality and are consequently employed in a fast-growing number of applications. However, since they are highly data-driven, relying on billion-sized datasets randomly scraped from the internet, they also suffer from degenerated and biased human behavior, as we demonstrate. In fact, they may even reinforce such biases. To not only uncover but also combat these undesired effects, we present a novel strategy, called \n Fair Diffusion \n , to attenuate biases during the deployment of generative text-to-image models. Specifically, we demonstrate shifting a bias in any direction based on human instructions yielding arbitrary proportions for, e.g., identity groups. As our empirical evaluation demonstrates, this introduced control enables instructing generative image models on fairness, requiring no data filtering nor additional training.},\n\tlanguage = {en},\n\turldate = {2025-05-13},\n\tjournal = {AI and Ethics},\n\tauthor = {Friedrich, Felix and Brack, Manuel and Struppek, Lukas and Hintersdorf, Dominik and Schramowski, Patrick and Luccioni, Sasha and Kersting, Kristian},\n\tmonth = aug,\n\tyear = {2024},\n}\n\n\n\n
\n
\n\n\n
\n Abstract Generative AI models have recently achieved astonishing results in quality and are consequently employed in a fast-growing number of applications. However, since they are highly data-driven, relying on billion-sized datasets randomly scraped from the internet, they also suffer from degenerated and biased human behavior, as we demonstrate. In fact, they may even reinforce such biases. To not only uncover but also combat these undesired effects, we present a novel strategy, called Fair Diffusion , to attenuate biases during the deployment of generative text-to-image models. Specifically, we demonstrate shifting a bias in any direction based on human instructions yielding arbitrary proportions for, e.g., identity groups. As our empirical evaluation demonstrates, this introduced control enables instructing generative image models on fairness, requiring no data filtering nor additional training.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Does CLIP Know My Face?.\n \n \n \n \n\n\n \n Hintersdorf, D.; Struppek, L.; Brack, M.; Friedrich, F.; Schramowski, P.; and Kersting, K.\n\n\n \n\n\n\n
Journal of Artificial Intelligence Research, 80: 1033–1062. July 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{hintersdorf_does_2024,\n\ttitle = {Does {CLIP} {Know} {My} {Face}?},\n\tvolume = {80},\n\tissn = {1076-9757},\n\turl = {https://www.jair.org/index.php/jair/article/view/15461},\n\tdoi = {10.1613/jair.1.15461},\n\tabstract = {With the rise of deep learning in various applications, privacy concerns around the protection of training data have become a critical area of research. Whereas prior studies have focused on privacy risks in single-modal models, we introduce a novel method to assess privacy for multi-modal models, specifically vision-language models like CLIP. The proposed Identity Inference Attack (IDIA) reveals whether an individual was included in the training data by querying the model with images of the same person. Letting the model choose from a wide variety of possible text labels, the model reveals whether it recognizes the person and, therefore, was used for training. Our large-scale experiments on CLIP demonstrate that individuals used for training can be identified with very high accuracy. We confirm that the model has learned to associate names with depicted individuals, implying the existence of sensitive information that can be extracted by adversaries. Our results highlight the need for stronger privacy protection in large-scale models and suggest that IDIAs can be used to prove the unauthorized use of data for training and to enforce privacy laws. \nThis article appears in the AI \\& Society track.},\n\turldate = {2025-05-13},\n\tjournal = {Journal of Artificial Intelligence Research},\n\tauthor = {Hintersdorf, Dominik and Struppek, Lukas and Brack, Manuel and Friedrich, Felix and Schramowski, Patrick and Kersting, Kristian},\n\tmonth = jul,\n\tyear = {2024},\n\tpages = {1033--1062},\n}\n\n\n\n
\n
\n\n\n
\n With the rise of deep learning in various applications, privacy concerns around the protection of training data have become a critical area of research. Whereas prior studies have focused on privacy risks in single-modal models, we introduce a novel method to assess privacy for multi-modal models, specifically vision-language models like CLIP. The proposed Identity Inference Attack (IDIA) reveals whether an individual was included in the training data by querying the model with images of the same person. Letting the model choose from a wide variety of possible text labels, the model reveals whether it recognizes the person and, therefore, was used for training. Our large-scale experiments on CLIP demonstrate that individuals used for training can be identified with very high accuracy. We confirm that the model has learned to associate names with depicted individuals, implying the existence of sensitive information that can be extracted by adversaries. Our results highlight the need for stronger privacy protection in large-scale models and suggest that IDIAs can be used to prove the unauthorized use of data for training and to enforce privacy laws. This article appears in the AI & Society track.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Introducing v0.5 of the AI Safety Benchmark from MLCommons.\n \n \n \n \n\n\n \n Vidgen, B.; Agrawal, A.; Ahmed, A. M.; Akinwande, V.; Al-Nuaimi, N.; Alfaraj, N.; Alhajjar, E.; Aroyo, L.; Bavalatti, T.; Bartolo, M.; Blili-Hamelin, B.; Bollacker, K.; Bomassani, R.; Boston, M. F.; Campos, S.; Chakra, K.; Chen, C.; Coleman, C.; Coudert, Z. D.; Derczynski, L.; Dutta, D.; Eisenberg, I.; Ezick, J.; Frase, H.; Fuller, B.; Gandikota, R.; Gangavarapu, A.; Gangavarapu, A.; Gealy, J.; Ghosh, R.; Goel, J.; Gohar, U.; Goswami, S.; Hale, S. A.; Hutiri, W.; Imperial, J. M.; Jandial, S.; Judd, N.; Juefei-Xu, F.; Khomh, F.; Kailkhura, B.; Kirk, H. R.; Klyman, K.; Knotz, C.; Kuchnik, M.; Kumar, S. H.; Kumar, S.; Lengerich, C.; Li, B.; Liao, Z.; Long, E. P.; Lu, V.; Luger, S.; Mai, Y.; Mammen, P. M.; Manyeki, K.; McGregor, S.; Mehta, V.; Mohammed, S.; Moss, E.; Nachman, L.; Naganna, D. J.; Nikanjam, A.; Nushi, B.; Oala, L.; Orr, I.; Parrish, A.; Patlak, C.; Pietri, W.; Poursabzi-Sangdeh, F.; Presani, E.; Puletti, F.; Röttger, P.; Sahay, S.; Santos, T.; Scherrer, N.; Sebag, A. S.; Schramowski, P.; Shahbazi, A.; Sharma, V.; Shen, X.; Sistla, V.; Tang, L.; Testuggine, D.; Thangarasa, V.; Watkins, E. A.; Weiss, R.; Welty, C.; Wilbers, T.; Williams, A.; Wu, C.; Yadav, P.; Yang, X.; Zeng, Y.; Zhang, W.; Zhdanov, F.; Zhu, J.; Liang, P.; Mattson, P.; and Vanschoren, J.\n\n\n \n\n\n\n 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@misc{vidgen_introducing_2024,\n\ttitle = {Introducing v0.5 of the {AI} {Safety} {Benchmark} from {MLCommons}},\n\tcopyright = {Creative Commons Attribution 4.0 International},\n\turl = {https://arxiv.org/abs/2404.12241},\n\tdoi = {10.48550/ARXIV.2404.12241},\n\tabstract = {This paper introduces v0.5 of the AI Safety Benchmark, which has been created by the MLCommons AI Safety Working Group. The AI Safety Benchmark has been designed to assess the safety risks of AI systems that use chat-tuned language models. We introduce a principled approach to specifying and constructing the benchmark, which for v0.5 covers only a single use case (an adult chatting to a general-purpose assistant in English), and a limited set of personas (i.e., typical users, malicious users, and vulnerable users). We created a new taxonomy of 13 hazard categories, of which 7 have tests in the v0.5 benchmark. We plan to release version 1.0 of the AI Safety Benchmark by the end of 2024. The v1.0 benchmark will provide meaningful insights into the safety of AI systems. However, the v0.5 benchmark should not be used to assess the safety of AI systems. We have sought to fully document the limitations, flaws, and challenges of v0.5. This release of v0.5 of the AI Safety Benchmark includes (1) a principled approach to specifying and constructing the benchmark, which comprises use cases, types of systems under test (SUTs), language and context, personas, tests, and test items; (2) a taxonomy of 13 hazard categories with definitions and subcategories; (3) tests for seven of the hazard categories, each comprising a unique set of test items, i.e., prompts. There are 43,090 test items in total, which we created with templates; (4) a grading system for AI systems against the benchmark; (5) an openly available platform, and downloadable tool, called ModelBench that can be used to evaluate the safety of AI systems on the benchmark; (6) an example evaluation report which benchmarks the performance of over a dozen openly available chat-tuned language models; (7) a test specification for the benchmark.},\n\turldate = {2025-05-13},\n\tpublisher = {arXiv},\n\tauthor = {Vidgen, Bertie and Agrawal, Adarsh and Ahmed, Ahmed M. and Akinwande, Victor and Al-Nuaimi, Namir and Alfaraj, Najla and Alhajjar, Elie and Aroyo, Lora and Bavalatti, Trupti and Bartolo, Max and Blili-Hamelin, Borhane and Bollacker, Kurt and Bomassani, Rishi and Boston, Marisa Ferrara and Campos, Siméon and Chakra, Kal and Chen, Canyu and Coleman, Cody and Coudert, Zacharie Delpierre and Derczynski, Leon and Dutta, Debojyoti and Eisenberg, Ian and Ezick, James and Frase, Heather and Fuller, Brian and Gandikota, Ram and Gangavarapu, Agasthya and Gangavarapu, Ananya and Gealy, James and Ghosh, Rajat and Goel, James and Gohar, Usman and Goswami, Sujata and Hale, Scott A. and Hutiri, Wiebke and Imperial, Joseph Marvin and Jandial, Surgan and Judd, Nick and Juefei-Xu, Felix and Khomh, Foutse and Kailkhura, Bhavya and Kirk, Hannah Rose and Klyman, Kevin and Knotz, Chris and Kuchnik, Michael and Kumar, Shachi H. and Kumar, Srijan and Lengerich, Chris and Li, Bo and Liao, Zeyi and Long, Eileen Peters and Lu, Victor and Luger, Sarah and Mai, Yifan and Mammen, Priyanka Mary and Manyeki, Kelvin and McGregor, Sean and Mehta, Virendra and Mohammed, Shafee and Moss, Emanuel and Nachman, Lama and Naganna, Dinesh Jinenhally and Nikanjam, Amin and Nushi, Besmira and Oala, Luis and Orr, Iftach and Parrish, Alicia and Patlak, Cigdem and Pietri, William and Poursabzi-Sangdeh, Forough and Presani, Eleonora and Puletti, Fabrizio and Röttger, Paul and Sahay, Saurav and Santos, Tim and Scherrer, Nino and Sebag, Alice Schoenauer and Schramowski, Patrick and Shahbazi, Abolfazl and Sharma, Vin and Shen, Xudong and Sistla, Vamsi and Tang, Leonard and Testuggine, Davide and Thangarasa, Vithursan and Watkins, Elizabeth Anne and Weiss, Rebecca and Welty, Chris and Wilbers, Tyler and Williams, Adina and Wu, Carole-Jean and Yadav, Poonam and Yang, Xianjun and Zeng, Yi and Zhang, Wenhui and Zhdanov, Fedor and Zhu, Jiacheng and Liang, Percy and Mattson, Peter and Vanschoren, Joaquin},\n\tyear = {2024},\n\tkeywords = {Artificial Intelligence (cs.AI), Computation and Language (cs.CL), FOS: Computer and information sciences},\n}\n\n\n\n
\n
\n\n\n
\n This paper introduces v0.5 of the AI Safety Benchmark, which has been created by the MLCommons AI Safety Working Group. The AI Safety Benchmark has been designed to assess the safety risks of AI systems that use chat-tuned language models. We introduce a principled approach to specifying and constructing the benchmark, which for v0.5 covers only a single use case (an adult chatting to a general-purpose assistant in English), and a limited set of personas (i.e., typical users, malicious users, and vulnerable users). We created a new taxonomy of 13 hazard categories, of which 7 have tests in the v0.5 benchmark. We plan to release version 1.0 of the AI Safety Benchmark by the end of 2024. The v1.0 benchmark will provide meaningful insights into the safety of AI systems. However, the v0.5 benchmark should not be used to assess the safety of AI systems. We have sought to fully document the limitations, flaws, and challenges of v0.5. This release of v0.5 of the AI Safety Benchmark includes (1) a principled approach to specifying and constructing the benchmark, which comprises use cases, types of systems under test (SUTs), language and context, personas, tests, and test items; (2) a taxonomy of 13 hazard categories with definitions and subcategories; (3) tests for seven of the hazard categories, each comprising a unique set of test items, i.e., prompts. There are 43,090 test items in total, which we created with templates; (4) a grading system for AI systems against the benchmark; (5) an openly available platform, and downloadable tool, called ModelBench that can be used to evaluate the safety of AI systems on the benchmark; (6) an example evaluation report which benchmarks the performance of over a dozen openly available chat-tuned language models; (7) a test specification for the benchmark.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Ledits++: Limitless image editing using text-to-image models.\n \n \n \n\n\n \n Brack, M.; Friedrich, F.; Kornmeier, K.; Tsaban, L.; Schramowski, P.; Kersting, K.; and Passos, A.\n\n\n \n\n\n\n In
Proceedings of the IEEE/CVF, pages 8861–8870, 2024. \n
\n\n
\n\n
\n\n
\n\n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{brack_ledits_2024,\n\ttitle = {Ledits++: {Limitless} image editing using text-to-image models},\n\tbooktitle = {Proceedings of the {IEEE}/{CVF}},\n\tauthor = {Brack, Manuel and Friedrich, Felix and Kornmeier, Katharina and Tsaban, Linoy and Schramowski, Patrick and Kersting, Kristian and Passos, Apolinário},\n\tyear = {2024},\n\tpages = {8861--8870},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Motion Primitives as the Action Space of Deep Q-Learning for Planning in Autonomous Driving.\n \n \n \n \n\n\n \n Schneider, T.; Pedrosa, M. V. A.; Gros, T. P.; Wolf, V.; and Flaßkamp, K.\n\n\n \n\n\n\n
IEEE Transactions on Intelligent Transportation Systems, 25(11): 17852–17864. November 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{schneider_motion_2024,\n\ttitle = {Motion {Primitives} as the {Action} {Space} of {Deep} {Q}-{Learning} for {Planning} in {Autonomous} {Driving}},\n\tvolume = {25},\n\tcopyright = {https://ieeexplore.ieee.org/Xplorehelp/downloads/license-information/IEEE.html},\n\tissn = {1524-9050, 1558-0016},\n\turl = {https://ieeexplore.ieee.org/document/10693315/},\n\tdoi = {10.1109/TITS.2024.3436530},\n\tnumber = {11},\n\turldate = {2025-05-13},\n\tjournal = {IEEE Transactions on Intelligent Transportation Systems},\n\tauthor = {Schneider, Tristan and Pedrosa, Matheus V. A. and Gros, Timo P. and Wolf, Verena and Flaßkamp, Kathrin},\n\tmonth = nov,\n\tyear = {2024},\n\tpages = {17852--17864},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n On the Quest for Effectiveness in Human Oversight: Interdisciplinary Perspectives.\n \n \n \n \n\n\n \n Sterz, S.; Baum, K.; Biewer, S.; Hermanns, H.; Lauber-Rönsberg, A.; Meinel, P.; and Langer, M.\n\n\n \n\n\n\n In
The 2024 ACM Conference on Fairness, Accountability, and Transparency, pages 2495–2507, Rio de Janeiro Brazil, June 2024. ACM\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{sterz_quest_2024,\n\taddress = {Rio de Janeiro Brazil},\n\ttitle = {On the {Quest} for {Effectiveness} in {Human} {Oversight}: {Interdisciplinary} {Perspectives}},\n\tisbn = {9798400704505},\n\tshorttitle = {On the {Quest} for {Effectiveness} in {Human} {Oversight}},\n\turl = {https://dl.acm.org/doi/10.1145/3630106.3659051},\n\tdoi = {10.1145/3630106.3659051},\n\tlanguage = {en},\n\turldate = {2025-05-13},\n\tbooktitle = {The 2024 {ACM} {Conference} on {Fairness}, {Accountability}, and {Transparency}},\n\tpublisher = {ACM},\n\tauthor = {Sterz, Sarah and Baum, Kevin and Biewer, Sebastian and Hermanns, Holger and Lauber-Rönsberg, Anne and Meinel, Philip and Langer, Markus},\n\tmonth = jun,\n\tyear = {2024},\n\tpages = {2495--2507},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Software doping analysis for human oversight.\n \n \n \n \n\n\n \n Biewer, S.; Baum, K.; Sterz, S.; Hermanns, H.; Hetmank, S.; Langer, M.; Lauber-Rönsberg, A.; and Lehr, F.\n\n\n \n\n\n\n
Formal Methods in System Design. April 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n \n \n 2 downloads\n \n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{biewer_software_2024,\n\ttitle = {Software doping analysis for human oversight},\n\tissn = {0925-9856, 1572-8102},\n\turl = {https://link.springer.com/10.1007/s10703-024-00445-2},\n\tdoi = {10.1007/s10703-024-00445-2},\n\tabstract = {Abstract \n \n This article introduces a framework that is meant to assist in mitigating societal risks that software can pose. Concretely, this encompasses facets of software doping as well as unfairness and discrimination in high-risk decision-making systems. The term \n software doping \n refers to software that contains surreptitiously added functionality that is against the interest of the user. A prominent example of software doping are the tampered emission cleaning systems that were found in millions of cars around the world when the diesel emissions scandal surfaced. The first part of this article combines the formal foundations of software doping analysis with established probabilistic falsification techniques to arrive at a black-box analysis technique for identifying undesired effects of software. We apply this technique to emission cleaning systems in diesel cars but also to high-risk systems that evaluate humans in a possibly unfair or discriminating way. We demonstrate how our approach can assist humans-in-the-loop to make better informed and more responsible decisions. This is to promote effective human oversight, which will be a central requirement enforced by the European Union’s upcoming AI Act. We complement our technical contribution with a juridically, philosophically, and psychologically informed perspective on the potential problems caused by such systems.},\n\tlanguage = {en},\n\turldate = {2025-05-13},\n\tjournal = {Formal Methods in System Design},\n\tauthor = {Biewer, Sebastian and Baum, Kevin and Sterz, Sarah and Hermanns, Holger and Hetmank, Sven and Langer, Markus and Lauber-Rönsberg, Anne and Lehr, Franz},\n\tmonth = apr,\n\tyear = {2024},\n}\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
\n
\n\n\n
\n Abstract This article introduces a framework that is meant to assist in mitigating societal risks that software can pose. Concretely, this encompasses facets of software doping as well as unfairness and discrimination in high-risk decision-making systems. The term software doping refers to software that contains surreptitiously added functionality that is against the interest of the user. A prominent example of software doping are the tampered emission cleaning systems that were found in millions of cars around the world when the diesel emissions scandal surfaced. The first part of this article combines the formal foundations of software doping analysis with established probabilistic falsification techniques to arrive at a black-box analysis technique for identifying undesired effects of software. We apply this technique to emission cleaning systems in diesel cars but also to high-risk systems that evaluate humans in a possibly unfair or discriminating way. We demonstrate how our approach can assist humans-in-the-loop to make better informed and more responsible decisions. This is to promote effective human oversight, which will be a central requirement enforced by the European Union’s upcoming AI Act. We complement our technical contribution with a juridically, philosophically, and psychologically informed perspective on the potential problems caused by such systems.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n ALERT: A Comprehensive Benchmark for Assessing Large Language Models' Safety through Red Teaming.\n \n \n \n \n\n\n \n Tedeschi, S.; Friedrich, F.; Schramowski, P.; Kersting, K.; Navigli, R.; Nguyen, H.; and Li, B.\n\n\n \n\n\n\n June 2024.\n
arXiv:2404.08676 [cs]\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n \n \n \n \n \n \n\n\n\n
\n
@misc{tedeschi_alert_2024,\n\ttitle = {{ALERT}: {A} {Comprehensive} {Benchmark} for {Assessing} {Large} {Language} {Models}' {Safety} through {Red} {Teaming}},\n\tshorttitle = {{ALERT}},\n\turl = {http://arxiv.org/abs/2404.08676},\n\tdoi = {10.48550/arXiv.2404.08676},\n\tabstract = {When building Large Language Models (LLMs), it is paramount to bear safety in mind and protect them with guardrails. Indeed, LLMs should never generate content promoting or normalizing harmful, illegal, or unethical behavior that may contribute to harm to individuals or society. This principle applies to both normal and adversarial use. In response, we introduce ALERT, a large-scale benchmark to assess safety based on a novel fine-grained risk taxonomy. It is designed to evaluate the safety of LLMs through red teaming methodologies and consists of more than 45k instructions categorized using our novel taxonomy. By subjecting LLMs to adversarial testing scenarios, ALERT aims to identify vulnerabilities, inform improvements, and enhance the overall safety of the language models. Furthermore, the fine-grained taxonomy enables researchers to perform an in-depth evaluation that also helps one to assess the alignment with various policies. In our experiments, we extensively evaluate 10 popular open- and closed-source LLMs and demonstrate that many of them still struggle to attain reasonable levels of safety.},\n\turldate = {2024-12-04},\n\tpublisher = {arXiv},\n\tauthor = {Tedeschi, Simone and Friedrich, Felix and Schramowski, Patrick and Kersting, Kristian and Navigli, Roberto and Nguyen, Huu and Li, Bo},\n\tmonth = jun,\n\tyear = {2024},\n\tnote = {arXiv:2404.08676 [cs]},\n\tkeywords = {Computer Science - Computation and Language, Computer Science - Computers and Society, Computer Science - Machine Learning},\n}\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
\n
\n\n\n
\n When building Large Language Models (LLMs), it is paramount to bear safety in mind and protect them with guardrails. Indeed, LLMs should never generate content promoting or normalizing harmful, illegal, or unethical behavior that may contribute to harm to individuals or society. This principle applies to both normal and adversarial use. In response, we introduce ALERT, a large-scale benchmark to assess safety based on a novel fine-grained risk taxonomy. It is designed to evaluate the safety of LLMs through red teaming methodologies and consists of more than 45k instructions categorized using our novel taxonomy. By subjecting LLMs to adversarial testing scenarios, ALERT aims to identify vulnerabilities, inform improvements, and enhance the overall safety of the language models. Furthermore, the fine-grained taxonomy enables researchers to perform an in-depth evaluation that also helps one to assess the alignment with various policies. In our experiments, we extensively evaluate 10 popular open- and closed-source LLMs and demonstrate that many of them still struggle to attain reasonable levels of safety.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Modular Design Patterns for Generative Neuro-Symbolic Systems.\n \n \n \n \n\n\n \n Boer, M. H. T. d.; Smit, Q. S.; Bekkum, M. v.; Meyer-Vitali, A.; and Schmid, T.\n\n\n \n\n\n\n In Sartini, B.; Raad, J.; Lisena, P.; Peñuela, A. M.; Beetz, M.; Blin, I.; Cimiano, P.; Berardinis, J. d.; Gottschalk, S.; Ilievski, F.; Jain, N.; Kim, J.; Kümpel, M.; Motta, E.; Tiddi, I.; and Töberg, J., editor(s),
Joint Proceedings of the ESWC 2024 Workshops and Tutorials, volume 3749, of
CEUR Workshop Proceedings, Hersonissos, Greece, May 2024. CEUR\n
ISSN: 1613-0073\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{boer_modular_2024,\n\taddress = {Hersonissos, Greece},\n\tseries = {{CEUR} {Workshop} {Proceedings}},\n\ttitle = {Modular {Design} {Patterns} for {Generative} {Neuro}-{Symbolic} {Systems}},\n\tvolume = {3749},\n\turl = {https://ceur-ws.org/Vol-3749/#genesy-03},\n\tlanguage = {en},\n\turldate = {2024-09-11},\n\tbooktitle = {Joint {Proceedings} of the {ESWC} 2024 {Workshops} and {Tutorials}},\n\tpublisher = {CEUR},\n\tauthor = {Boer, Maaike H. T. de and Smit, Quirine S. and Bekkum, Michael van and Meyer-Vitali, André and Schmid, Thomas},\n\teditor = {Sartini, Bruno and Raad, Joe and Lisena, Pasquale and Peñuela, Albert Meroño and Beetz, Michael and Blin, Inès and Cimiano, Philipp and Berardinis, Jacopo de and Gottschalk, Simon and Ilievski, Filip and Jain, Nitisha and Kim, Jongmo and Kümpel, Michaela and Motta, Enrico and Tiddi, Ilaria and Töberg, Jan-Philipp},\n\tmonth = may,\n\tyear = {2024},\n\tnote = {ISSN: 1613-0073},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Trusted AI – Vertrauenswürdigkeit und große Sprachmodelle.\n \n \n \n \n\n\n \n Meyer-Vitali, A.; and Ostermann, S.\n\n\n \n\n\n\n
dfki ai next, 2024(2): 6–7. September 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{meyer-vitali_trusted_2024,\n\ttitle = {Trusted {AI} – {Vertrauenswürdigkeit} und große {Sprachmodelle}},\n\tvolume = {2024},\n\turl = {https://www.dfki.de/web/news-media/news/dfki-ai-next},\n\tlanguage = {German},\n\tnumber = {2},\n\tjournal = {dfki ai next},\n\tauthor = {Meyer-Vitali, André and Ostermann, Simon},\n\tmonth = sep,\n\tyear = {2024},\n\tpages = {6--7},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n AI Engineering for Trust by Design.\n \n \n \n \n\n\n \n Meyer-Vitali, A.\n\n\n \n\n\n\n In
Proceedings of the 12th International Conference on Model-Based Software and Systems Engineering (MODELSWARD 2024), pages 357–364, Rome, Italy, February 2024. SCITEPRESS – Science and Technology Publications, Lda\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{meyer-vitali_ai_2024,\n\taddress = {Rome, Italy},\n\ttitle = {{AI} {Engineering} for {Trust} by {Design}},\n\tcopyright = {All rights reserved},\n\tisbn = {978-989-758-682-8},\n\turl = {https://www.scitepress.org/PublicationsDetail.aspx?ID=/skO/EwOJr4=&t=1},\n\tdoi = {10.5220/0012622400003645},\n\tabstract = {Digital Library},\n\turldate = {2024-02-29},\n\tbooktitle = {Proceedings of the 12th {International} {Conference} on {Model}-{Based} {Software} and {Systems} {Engineering} ({MODELSWARD} 2024)},\n\tpublisher = {SCITEPRESS – Science and Technology Publications, Lda},\n\tauthor = {Meyer-Vitali, André},\n\tmonth = feb,\n\tyear = {2024},\n\tpages = {357--364},\n}\n\n\n\n
\n
\n\n\n
\n Digital Library\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n Human-AI Engineering for Adults.\n \n \n \n \n\n\n \n Meyer-Vitali, A.; and Mulder, W.\n\n\n \n\n\n\n In
HHAI 2024: Hybrid Human AI Systems for the Social Good, pages 228–240. IOS Press, 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@incollection{meyer-vitali_human-ai_2024,\n\ttitle = {Human-{AI} {Engineering} for {Adults}},\n\tcopyright = {All rights reserved},\n\turl = {https://ebooks.iospress.nl/doi/10.3233/FAIA240197},\n\turldate = {2024-06-08},\n\tbooktitle = {{HHAI} 2024: {Hybrid} {Human} {AI} {Systems} for the {Social} {Good}},\n\tpublisher = {IOS Press},\n\tauthor = {Meyer-Vitali, Andr\\&\\#233 and Mulder, Wico},\n\tyear = {2024},\n\tdoi = {10.3233/FAIA240197},\n\tpages = {228--240},\n}\n\n\n\n
\n
\n\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n Engineering Principles for Building Trusted Human-AI Systems.\n \n \n \n\n\n \n Meyer-Vitali, A.; and Mulder, W.\n\n\n \n\n\n\n In Arai, K., editor(s),
Intelligent Systems and Applications, pages 468–485, Cham, 2024. Springer Nature Switzerland\n
\n\n
\n\n
\n\n
\n\n \n\n \n \n doi\n \n \n\n \n link\n \n \n\n bibtex\n \n\n \n \n \n abstract \n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@inproceedings{meyer-vitali_engineering_2024,\n\taddress = {Cham},\n\ttitle = {Engineering {Principles} for {Building} {Trusted} {Human}-{AI} {Systems}},\n\tcopyright = {All rights reserved},\n\tisbn = {978-3-031-66428-1},\n\tdoi = {10.1007/978-3-031-66428-1_30},\n\tabstract = {In the process engineering reliable and trustworthy AI systems there is significant wisdom to be gained from traditional engineering domains. Extending on earlier work our attention is on topics that stress the principles of building human-AI systems. We plea for a reinforced attention for engineering methods and processes in order to urge the essence for improved scientific progress and industrial AI applications where one can stand on the shoulders of giants. On the one hand, we see their complexity increase on an individual level, as well as on their connected dependency levels, whilst on the other hand, we see a growing lack of experience on the level of their design and engineering. The complexity of current AI models often limits our understanding. The methods and processes to ensure safety, reliability, and transparency are insufficient. This poses serious risks at the level of trustworthiness, particularly when it comes to critical applications with significant social, economic or even physical impact. Future AI systems must adhere to stringent requirements, as mandated, for instance, by the European AI Act, ensuring meticulous design, validation, and certification based on clearly defined criteria.},\n\tlanguage = {en},\n\tbooktitle = {Intelligent {Systems} and {Applications}},\n\tpublisher = {Springer Nature Switzerland},\n\tauthor = {Meyer-Vitali, André and Mulder, Wico},\n\teditor = {Arai, Kohei},\n\tyear = {2024},\n\tpages = {468--485},\n}\n\n\n\n
\n
\n\n\n
\n In the process engineering reliable and trustworthy AI systems there is significant wisdom to be gained from traditional engineering domains. Extending on earlier work our attention is on topics that stress the principles of building human-AI systems. We plea for a reinforced attention for engineering methods and processes in order to urge the essence for improved scientific progress and industrial AI applications where one can stand on the shoulders of giants. On the one hand, we see their complexity increase on an individual level, as well as on their connected dependency levels, whilst on the other hand, we see a growing lack of experience on the level of their design and engineering. The complexity of current AI models often limits our understanding. The methods and processes to ensure safety, reliability, and transparency are insufficient. This poses serious risks at the level of trustworthiness, particularly when it comes to critical applications with significant social, economic or even physical impact. Future AI systems must adhere to stringent requirements, as mandated, for instance, by the European AI Act, ensuring meticulous design, validation, and certification based on clearly defined criteria.\n
\n\n\n
\n\n\n
\n
\n\n \n \n \n \n \n \n AI Engineering for Trust by Design.\n \n \n \n \n\n\n \n Meyer-Vitali, A.\n\n\n \n\n\n\n
Survey of Tools for Software Engineering, 24(1): 20–22. 2024.\n
\n\n
\n\n
\n\n
\n\n \n \n
Paper\n \n \n\n \n\n \n link\n \n \n\n bibtex\n \n\n \n\n \n\n \n \n \n \n \n \n \n\n \n \n \n\n\n\n
\n
@article{meyer-vitali_ai_2024,\n\ttitle = {{AI} {Engineering} for {Trust} by {Design}},\n\tvolume = {24},\n\turl = {https://www.software-innovations.eu/publikationen/},\n\tnumber = {1},\n\tjournal = {Survey of Tools for Software Engineering},\n\tauthor = {Meyer-Vitali, André},\n\tyear = {2024},\n\tpages = {20--22},\n}\n\n\n\n\n
\n
\n\n\n\n
\n\n\n\n\n\n