Case-based Data Quality Management for IoT Logs: A Case Study Focusing on Detection of Data Quality Issues.
Schultheis, A.; Bertrand, Y.; Grüger, J.; Malburg, L.; Bergmann, R.; and Serral Asensio, E.
IoT, 6(4). 2025.
doi
link
bibtex
abstract
@article{SchultheisBGMBSA2025,
author = {Schultheis, Alexander and Bertrand, Yannis and Grüger, Joscha and Malburg, Lukas and Bergmann, Ralph and {Serral Asensio}, Estefanía},
title = {{Case-based Data Quality Management for IoT Logs: A Case Study Focusing on Detection of Data Quality Issues}},
journal = {IoT},
year = {2025},
volume = {6},
number = {4},
article-number = {63},
doi = {10.3390/iot6040063},
abstract = {Smart manufacturing applications increasingly rely on time-series data from Industrial IoT sensors, yet these data streams often contain data quality issues (DQIs) that affect analysis and disrupt production. While traditional Machine Learning methods are difficult to apply due to the small amount of data available, the knowledge-based approach of Case-Based Reasoning (CBR) offers a way to reuse previously gained experience. We introduce the first end-to-end Case-Based Reasoning (CBR) framework that both detects and remedies DQIs in near real time, even when only a handful of annotated fault instances are available. Our solution encodes expert experience in the four CBR knowledge containers: (i) a vocabulary that represents sensor streams and their context in the DataStream format; (ii) a case base populated with fault-annotated event logs; (iii) tailored similarity measures—including a weighted Dynamic Time Warping variant and structure-aware list mapping—that isolate the signatures of missing-value, missing-sensor, and time-shift errors; and (iv) lightweight adaptation rules that recommend concrete repair actions or, where appropriate, invoke automated imputation and alignment routines. A case study is used to examine and present the suitability of the approach for a specific application domain. Although the case study demonstrates only limited capabilities in identifying Data Quality Issues (DQIs), we aim to support transparent evaluation and future research by publishing (1) a prototype of the Case-Based Reasoning (CBR) system and (2) a publicly accessible, meticulously annotated sensor-log benchmark. Together, these resources provide a reproducible baseline and a modular foundation for advancing similarity metrics, expanding the DQI taxonomy, and enabling knowledge-intensive reasoning in IoT data quality management.},
keywords = {Time Series Data, Industrial Internet of Things, Data Quality Issues, Temporal Case-Based Reasoning}
}
Smart manufacturing applications increasingly rely on time-series data from Industrial IoT sensors, yet these data streams often contain data quality issues (DQIs) that affect analysis and disrupt production. While traditional Machine Learning methods are difficult to apply due to the small amount of data available, the knowledge-based approach of Case-Based Reasoning (CBR) offers a way to reuse previously gained experience. We introduce the first end-to-end Case-Based Reasoning (CBR) framework that both detects and remedies DQIs in near real time, even when only a handful of annotated fault instances are available. Our solution encodes expert experience in the four CBR knowledge containers: (i) a vocabulary that represents sensor streams and their context in the DataStream format; (ii) a case base populated with fault-annotated event logs; (iii) tailored similarity measures—including a weighted Dynamic Time Warping variant and structure-aware list mapping—that isolate the signatures of missing-value, missing-sensor, and time-shift errors; and (iv) lightweight adaptation rules that recommend concrete repair actions or, where appropriate, invoke automated imputation and alignment routines. A case study is used to examine and present the suitability of the approach for a specific application domain. Although the case study demonstrates only limited capabilities in identifying Data Quality Issues (DQIs), we aim to support transparent evaluation and future research by publishing (1) a prototype of the Case-Based Reasoning (CBR) system and (2) a publicly accessible, meticulously annotated sensor-log benchmark. Together, these resources provide a reproducible baseline and a modular foundation for advancing similarity metrics, expanding the DQI taxonomy, and enabling knowledge-intensive reasoning in IoT data quality management.
Adaptive Workflow Management for Cyber-Physical Systems - A Hybrid Approach Combining Experience-Based Learning and Automated Planning for Industry 4.0.
Malburg, L.
Ph.D. Thesis, Universität Trier, 2025.
link
bibtex
abstract
@phdthesis{Malburg_PhDThesis_2025,
abstract = {The manufacturing industry is confronted with challenges caused by shorter innovation and product lifecycles, increasing complexity of production due to individualization and customization, and production disturbances of various kinds.
To address these challenges, more autonomous and intelligent production systems are required, enabling the manufacturing industry to react more flexibly to these changing conditions.
One goal of the Fourth Industrial Revolution, also known as Industry 4.0 (I4.0), is to enhance the digitalization and connectivity of high-level information systems with low-level production processes.
This connectivity is enabled by the fact that an increasing number of devices and machines are connected to the Internet, known as the Internet of Things (IoT).
According to this, IoT is a key technology for realizing Cyber-Physical Systems (CPSs), linking the physical world, i.e., devices and machines, with the digital world.
In the manufacturing industry, Cyber-Physical Production Systems (CPPSs) as a special kind of CPSs are applied to optimize production processes and to increase efficiency and flexibility in manufacturing.
Although CPPSs are already used in industry and there is a growing amount of research in this field, a bidirectional coupling between high-level information systems and low-level production processes is often still missing.
Consequently, the IoT sensor data generated during production cannot be fully utilized to monitor low-level production processes, which in turn makes it more difficult for high-level information systems to detect emerging situations at an early stage and to perform real-time adaptations.
The approaches for enabling flexibility and adaptability of processes in CPPSs are still in their infancy and other available approaches do not adequately consider the specific characteristics of CPPSs such as unanticipated behavior or physical errors.
Consequently, current production processes are often still too rigid, which makes it difficult to react to changing environmental conditions or disturbances in real-time and in an automated way.
This thesis aims to contribute to the aforementioned problems by developing an approach for adaptive workflow management for CPPSs by using Artificial Intelligence (AI) techniques.
For this purpose, a systems architecture is developed that enables the bidirectional coupling between high-level information systems and low-level processes.
Based on this coupling, Business Process Management (BPM) methods are used to model production processes on an abstract level and in a more structured way, allowing the flexible composition and execution of low-level production processes.
Moreover, methods from BPM enable the utilization of IoT sensor data for process monitoring and analysis, building the basis for detecting emerging situations and disturbances in production processes earlier.
The presented adaptive workflow management approach enables resolving such situations by ad hoc adaptations of production processes in near real-time.
This is achieved by combining search-intensive AI planning and knowledge-intensive Case-Based Reasoning (CBR).
One main advantage of this combined approach is that already experienced problem situations in the form of performed ad hoc adaptations (called cases) can be reused to solve new upcoming similar problems.
In addition, the search-intensive AI planning solves problems that have not yet been experienced as cases.
In this way, the combined approach helps to limit both the high knowledge acquisition and modeling efforts and the high computational complexity compared to using search-intensive AI planning solely.
At the same time, it also increases problem-solving competence in situations where CBR alone cannot yet solve these problems.
Finally, the approach enables incremental self-learning and, thus, improves the problem-solving competence continuously.
In addition, the use of CBR enables the integration of domain experts and their expertise into the combined approach.
The experimental evaluations indicate the suitability of the proposed approach for performing ad hoc adaptations in CPPSs.
For this purpose, a learning factory has been used as a test bed to simulate production processes of a real manufacturing environment.
Using such learning factories, the characteristics of real-world production environments are considered, and at the same time, such learning factories allow a more cost-effective and easier realization of I4.0 research.
For evaluation, the production processes have been modeled and executed in the learning factory.
In addition, failure situations have been generated to test whether these failures can be detected with enhanced monitoring techniques and to adapt production processes using the proposed adaptive workflow management approach.},
year = {2025},
title = {{Adaptive Workflow Management for Cyber-Physical Systems - A Hybrid Approach Combining Experience-Based Learning and Automated Planning for Industry 4.0}},
school = {Universität Trier},
author = {Lukas Malburg},
keywords = {Adaptive Workflow Management, Cyber-Physical Production Systems, Internet of Things, Case-Based Reasoning, Automated Planning}
}
The manufacturing industry is confronted with challenges caused by shorter innovation and product lifecycles, increasing complexity of production due to individualization and customization, and production disturbances of various kinds. To address these challenges, more autonomous and intelligent production systems are required, enabling the manufacturing industry to react more flexibly to these changing conditions. One goal of the Fourth Industrial Revolution, also known as Industry 4.0 (I4.0), is to enhance the digitalization and connectivity of high-level information systems with low-level production processes. This connectivity is enabled by the fact that an increasing number of devices and machines are connected to the Internet, known as the Internet of Things (IoT). According to this, IoT is a key technology for realizing Cyber-Physical Systems (CPSs), linking the physical world, i.e., devices and machines, with the digital world. In the manufacturing industry, Cyber-Physical Production Systems (CPPSs) as a special kind of CPSs are applied to optimize production processes and to increase efficiency and flexibility in manufacturing. Although CPPSs are already used in industry and there is a growing amount of research in this field, a bidirectional coupling between high-level information systems and low-level production processes is often still missing. Consequently, the IoT sensor data generated during production cannot be fully utilized to monitor low-level production processes, which in turn makes it more difficult for high-level information systems to detect emerging situations at an early stage and to perform real-time adaptations. The approaches for enabling flexibility and adaptability of processes in CPPSs are still in their infancy and other available approaches do not adequately consider the specific characteristics of CPPSs such as unanticipated behavior or physical errors. Consequently, current production processes are often still too rigid, which makes it difficult to react to changing environmental conditions or disturbances in real-time and in an automated way. This thesis aims to contribute to the aforementioned problems by developing an approach for adaptive workflow management for CPPSs by using Artificial Intelligence (AI) techniques. For this purpose, a systems architecture is developed that enables the bidirectional coupling between high-level information systems and low-level processes. Based on this coupling, Business Process Management (BPM) methods are used to model production processes on an abstract level and in a more structured way, allowing the flexible composition and execution of low-level production processes. Moreover, methods from BPM enable the utilization of IoT sensor data for process monitoring and analysis, building the basis for detecting emerging situations and disturbances in production processes earlier. The presented adaptive workflow management approach enables resolving such situations by ad hoc adaptations of production processes in near real-time. This is achieved by combining search-intensive AI planning and knowledge-intensive Case-Based Reasoning (CBR). One main advantage of this combined approach is that already experienced problem situations in the form of performed ad hoc adaptations (called cases) can be reused to solve new upcoming similar problems. In addition, the search-intensive AI planning solves problems that have not yet been experienced as cases. In this way, the combined approach helps to limit both the high knowledge acquisition and modeling efforts and the high computational complexity compared to using search-intensive AI planning solely. At the same time, it also increases problem-solving competence in situations where CBR alone cannot yet solve these problems. Finally, the approach enables incremental self-learning and, thus, improves the problem-solving competence continuously. In addition, the use of CBR enables the integration of domain experts and their expertise into the combined approach. The experimental evaluations indicate the suitability of the proposed approach for performing ad hoc adaptations in CPPSs. For this purpose, a learning factory has been used as a test bed to simulate production processes of a real manufacturing environment. Using such learning factories, the characteristics of real-world production environments are considered, and at the same time, such learning factories allow a more cost-effective and easier realization of I4.0 research. For evaluation, the production processes have been modeled and executed in the learning factory. In addition, failure situations have been generated to test whether these failures can be detected with enhanced monitoring techniques and to adapt production processes using the proposed adaptive workflow management approach.
Case-Based Reasoning Meets Large Language Models: A Research Manifesto For Open Challenges and Research Directions.
Bach, K.; Bergmann, R.; Brand, F.; Caro-Martínez, Marta; Eisenstadt, V.; W. Floyd, M.; Jayawardena, L.; Leake, D.; Lenz, M.; Malburg, L.; H. Ménager, D.; Minor, M.; Schack, B.; Watson, I.; Wilkerson, K.; and Wiratunga, N.
3 2025.
working paper or preprint
link
bibtex
abstract
@misc{BachEtAl_CBRMeetsLLMs_2025,
abstract = {In recent years, the surge of Generative Artificial Intelligence (GenAI), particularly Large Language Models (LLMs), has led to a significant increase in the use of hybrid systems, which combine the strengths of different Artificial Intelligence (AI) paradigms to achieve better performance and efficiency. Although LLMs demonstrate remarkable effectiveness across numerous tasks due to their flexibility and general knowledge, they often face challenges related to accuracy, explainability, and their limited memory. CaseBased Reasoning (CBR), on the other hand, excels by recalling past experiences and using them to solve new problems, making it particularly well suited for tasks that require contextual understanding and decisionmaking. However, CBR systems suffer from issues such as the acquisition of various kinds of knowledge and the application of methods during the 4R cycle. In this paper, we identify several challenges plaguing LLMs and CBR systems and propose opportunities to combine the strengths of both methodologies to address these challenges. In addition, we outline future research directions for the community to explore.},
month = {3},
year = {2025},
title = {{Case-Based Reasoning Meets Large Language Models: A Research Manifesto For Open Challenges and Research Directions}},
note = {working paper or preprint},
author = {Bach, Kerstin and Bergmann, Ralph and Brand, Florian and Caro-Mart{\'i}nez, Marta and Eisenstadt, Viktor and W. Floyd, Michael and Jayawardena, Lasal and Leake, David and Lenz, Mirko and Malburg, Lukas and H. M{\'e}nager, David and Minor, Mirjam and Schack, Brian and Watson, Ian and Wilkerson, Kaitlynne and Wiratunga, Nirmalie},
keywords = {Large Language Models, Case-Based Reasoning, Large Reasoning Models, Retrieval Augmented Generation, Challenges and Opportunities}
}
In recent years, the surge of Generative Artificial Intelligence (GenAI), particularly Large Language Models (LLMs), has led to a significant increase in the use of hybrid systems, which combine the strengths of different Artificial Intelligence (AI) paradigms to achieve better performance and efficiency. Although LLMs demonstrate remarkable effectiveness across numerous tasks due to their flexibility and general knowledge, they often face challenges related to accuracy, explainability, and their limited memory. CaseBased Reasoning (CBR), on the other hand, excels by recalling past experiences and using them to solve new problems, making it particularly well suited for tasks that require contextual understanding and decisionmaking. However, CBR systems suffer from issues such as the acquisition of various kinds of knowledge and the application of methods during the 4R cycle. In this paper, we identify several challenges plaguing LLMs and CBR systems and propose opportunities to combine the strengths of both methodologies to address these challenges. In addition, we outline future research directions for the community to explore.
ReadBench: Measuring the Dense Text Visual Reading Ability of Vision-Language Models.
Clavié, B.; and Brand, F.
2025.
Paper
link
bibtex
@misc{clavie2025readbenchmeasuringdensetext,
title={ReadBench: Measuring the Dense Text Visual Reading Ability of Vision-Language Models},
author={Benjamin Clavié and Florian Brand},
year={2025},
eprint={2505.19091},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://www.wi2.uni-trier.de/shared/publications/clavie2025readbenchmeasuringdensetext.pdf},
}
Towards Fairness in Synthetic Healthcare Data: A Framework for the Evaluation of Synthetization Algorithms.
Warnecke, Y.; Kuhn, M.; Diederichs, F.; Brix, T. J; Clever, L.; Bergmann, R.; Heider, D.; and Storck, M.
Studies in Health Technology and Informatics, 331: 25–34. 2025.
doi
link
bibtex
abstract
@article{Warnecke2025FairnessSynthetic,
title = {Towards Fairness in Synthetic Healthcare Data: A Framework for the Evaluation of Synthetization Algorithms},
author = {Warnecke, Yannik and Kuhn, Martin and Diederichs, Felix and Brix, Tobias J and Clever, Lena and Bergmann, Ralph and Heider, Dominik and Storck, Michael},
journal = {Studies in Health Technology and Informatics},
volume = {331},
pages = {25--34},
year = {2025},
doi = {10.3233/SHTI251376},
issn = {1879-8365},
publisher = {IOS Press},
address = {Netherlands},
abstract = {Synthetic data generation is a rapidly evolving field, with significant potential for improving data privacy. However, evaluating the performance of synthetic data generation methods, especially the tradeoff between fairness and utility of the generated data, remains a challenge. In this work, we present our comprehensive framework, which evaluates fair synthetic data generation methods, benchmarking them against state-of-the-art synthesizers. The proposed framework consists of selection, evaluation, and application components that assess fairness, utility, and resemblance in real-world scenarios. The framework was applied to state-of-the-art data synthesizers, including TabFairGAN, DECAF, TVAE, and CTGAN, using a publicly available medical dataset. The results reveal the strengths and limitations of each synthesizer, including their bias mitigation strategies and trade-offs between fairness and utility, thereby showing the framework's effectiveness.},
pmid = {40899524},
keywords = {Algorithms, Artificial Intelligence, Data Quality, Delivery of Health Care, Health Equity, Medical Informatics}
}
Synthetic data generation is a rapidly evolving field, with significant potential for improving data privacy. However, evaluating the performance of synthetic data generation methods, especially the tradeoff between fairness and utility of the generated data, remains a challenge. In this work, we present our comprehensive framework, which evaluates fair synthetic data generation methods, benchmarking them against state-of-the-art synthesizers. The proposed framework consists of selection, evaluation, and application components that assess fairness, utility, and resemblance in real-world scenarios. The framework was applied to state-of-the-art data synthesizers, including TabFairGAN, DECAF, TVAE, and CTGAN, using a publicly available medical dataset. The results reveal the strengths and limitations of each synthesizer, including their bias mitigation strategies and trade-offs between fairness and utility, thereby showing the framework's effectiveness.
Digital Twins of Business Processes: A Research Manifesto.
Fornari, F.; Compagnucci, I.; Callisto De Donato, M.; Bertrand, Y.; Beyel, H. H.; Carrión, E.; Franceschetti, M.; Groher, W.; Grüger, J.; Kilic, E.; Koschmider, A.; Leotta, F.; Li, C.; Lugaresi, G.; Malburg, L.; Mangler, J.; Mecella, M.; Pastor, O.; Riss, U.; Seiger, R.; Serral, E.; Torres, V.; and Valderas, P.
Internet of Things, 30: 101477. 2025.
Paper
doi
link
bibtex
abstract
@article{FORNARI2025101477,
title = {{Digital Twins of Business Processes: A Research Manifesto}},
journal = {Internet of Things},
volume = {30},
pages = {101477},
year = {2025},
issn = {2542-6605},
doi = {https://doi.org/10.1016/j.iot.2024.101477},
url={https://www.wi2.uni-trier.de/shared/publications/Fornari2025DigitalTwinsOfBP.pdf},
author = {Fabrizio Fornari and Ivan Compagnucci and Massimo {Callisto De Donato} and Yannis Bertrand and Harry H. Beyel and Emilio Carrión and Marco Franceschetti and Wolfgang Groher and Joscha Grüger and Emre Kilic and Agnes Koschmider and Francesco Leotta and Chiao-Yun Li and Giovani Lugaresi and Lukas Malburg and Juergen Mangler and Massimo Mecella and Oscar Pastor and Uwe Riss and Ronny Seiger and Estefania Serral and Victoria Torres and Pedro Valderas},
keywords = {Digital Twin, Business process, Internet of Things},
abstract = {Modern organizations necessitate continuous business processes improvement to maintain efficiency, adaptability, and competitiveness. In the last few years, the Internet of Things, via the deployment of sensors and actuators, has heavily been adopted in organizational and industrial settings to monitor and automatize physical processes influencing and enhancing how people and organizations work. Such advancements are now pushed forward by the rise of the Digital Twin paradigm applied to organizational processes. Advanced ways of managing and maintaining business processes come within reach as there is a Digital Twin of a business process - a virtual replica with real-time capabilities of a real process occurring in an organization. Combining business process models with real-time data and simulation capabilities promises to provide a new way to guide day-to-day organization activities. However, integrating Digital Twins and business processes is a non-trivial task, presenting numerous challenges and ambiguities. This manifesto paper aims to contribute to the current state of the art by clarifying the relationship between business processes and Digital Twins, identifying ongoing research and open challenges, thereby shedding light on and driving future exploration of this innovative interplay.}
}
Modern organizations necessitate continuous business processes improvement to maintain efficiency, adaptability, and competitiveness. In the last few years, the Internet of Things, via the deployment of sensors and actuators, has heavily been adopted in organizational and industrial settings to monitor and automatize physical processes influencing and enhancing how people and organizations work. Such advancements are now pushed forward by the rise of the Digital Twin paradigm applied to organizational processes. Advanced ways of managing and maintaining business processes come within reach as there is a Digital Twin of a business process - a virtual replica with real-time capabilities of a real process occurring in an organization. Combining business process models with real-time data and simulation capabilities promises to provide a new way to guide day-to-day organization activities. However, integrating Digital Twins and business processes is a non-trivial task, presenting numerous challenges and ambiguities. This manifesto paper aims to contribute to the current state of the art by clarifying the relationship between business processes and Digital Twins, identifying ongoing research and open challenges, thereby shedding light on and driving future exploration of this innovative interplay.
Probabilistic Programming for Trace Generation (and Beyond).
Kuhn, M.; Grüger, J.; Matheja, C.; and Rivkin, A.
In
International Workshop on Algorithms & Theories for the Analysis of Event Data (ATAED’25), Paris, France, June 2025. CEUR-WS.org
Workshop held as part of a satellite event 46th International Conference on Application and Theory of Petri Nets and Concurrency (Petri Nets 2025)
Paper
link
bibtex
@inproceedings{kuhn2025probabilistic_beyond,
author = {Martin Kuhn and Joscha Grüger and Christoph Matheja and Andrey Rivkin},
title = {Probabilistic Programming for Trace Generation (and Beyond)},
booktitle = {International Workshop on Algorithms \& Theories for the Analysis of Event Data (ATAED’25)},
year = {2025},
address = {Paris, France},
month = {June},
publisher = {CEUR-WS.org},
note = {Workshop held as part of a satellite event 46th International Conference on Application and Theory of Petri Nets and Concurrency (Petri Nets 2025)},
url = {https://ceur-ws.org/Vol-3998/paper10.pdf}
}
Research Paper: Enhancing Healthcare Decision-Making with Analogy-Based Reasoning.
Grüger, J.; Kuhn, M.; Amri, K.; and Bergmann, R.
In Delgado, A.; and Slaats, T., editor(s),
Process Mining Workshops, pages 447–459, Cham, 2025. Springer Nature Switzerland
link
bibtex
abstract
@InProceedings{healthcare_decision_making_2025,
author="Gr{\"u}ger, Joscha
and Kuhn, Martin
and Amri, Karim
and Bergmann, Ralph",
editor="Delgado, Andrea
and Slaats, Tijs",
title="Research Paper: Enhancing Healthcare Decision-Making with Analogy-Based Reasoning",
booktitle="Process Mining Workshops",
year="2025",
publisher="Springer Nature Switzerland",
address="Cham",
pages="447--459",
abstract="Analogy-based reasoning is often employed in the treatment of hospitalized patients, especially when clinical guidelines or robust evidence bases are unavailable. This approach is based on the assumption that similar patients respond similarly to comparable treatments. Traditionally, this reasoning has relied on the memory and experience of physicians. However, the complexity of managing patient data---such as treatment sequences and responses---presents significant challenges without technological support. In particular, the procedural perspective of comparing patients is especially demanding. To address these challenges, we introduce the MAPI framework, an innovative approach for analogy-based, process-oriented search within patient data. This framework systematically manages treatment data, defines precise similarity measures, and retrieves comparable patient cases using case-based reasoning (CBR). By integrating analogy-based reasoning, MAPI enhances decision-making and improves the explainability of treatment choices, offering a more reliable and transparent tool for clinical practice.",
isbn="978-3-031-82225-4"
}
Analogy-based reasoning is often employed in the treatment of hospitalized patients, especially when clinical guidelines or robust evidence bases are unavailable. This approach is based on the assumption that similar patients respond similarly to comparable treatments. Traditionally, this reasoning has relied on the memory and experience of physicians. However, the complexity of managing patient data—such as treatment sequences and responses—presents significant challenges without technological support. In particular, the procedural perspective of comparing patients is especially demanding. To address these challenges, we introduce the MAPI framework, an innovative approach for analogy-based, process-oriented search within patient data. This framework systematically manages treatment data, defines precise similarity measures, and retrieves comparable patient cases using case-based reasoning (CBR). By integrating analogy-based reasoning, MAPI enhances decision-making and improves the explainability of treatment choices, offering a more reliable and transparent tool for clinical practice.
Clinical Decision Support for Skin Tumor Treatment: A Case-Based Reasoning Approach.
Kuhn, M.; Warnecke, Y.; Preciado-Marquez, D.; Grüger, J.; Bley, L. I.; Storck, M.; Weishaupt, C.; Bergmann, R.; and Braun, S. A.
In Bichindaritz, I.; and López, B., editor(s),
Case-Based Reasoning Research and Development, pages 407–422, Cham, 2025. Springer Nature Switzerland
link
bibtex
abstract
@InProceedings{CBR_CDSS_2025,
author="Kuhn, Martin
and Warnecke, Yannik
and Preciado-Marquez, Daniel
and Gr{\"u}ger, Joscha
and Bley, Laura Isabell
and Storck, Michael
and Weishaupt, Carsten
and Bergmann, Ralph
and Braun, Stephan Alexander",
editor="Bichindaritz, Isabelle
and L{\'o}pez, Beatriz",
title="Clinical Decision Support for Skin Tumor Treatment: A Case-Based Reasoning Approach",
booktitle="Case-Based Reasoning Research and Development",
year="2025",
publisher="Springer Nature Switzerland",
address="Cham",
pages="407--422",
abstract="Cancer treatment planning is a complex and individualized process due to the variability of patient-specific factors, tumor characteristics, and evolving medical standards. Predicting the next step in diagnosis or therapy remains a significant challenge, due to the high variability and limited availability of structured medical data. Clinical Decision Support Systems (CDSS) offer a promising solution, with Case-Based Reasoning (CBR) standing out for its ability to provide interpretable and transparent recommendations. Unlike black-box machine learning models, CBR leverages past cases to generate predictions by analogy, aligning with the way clinicians naturally reason. In this work, we propose a CBR-based CDSS for skin cancer treatment that integrates medical taxonomies and patient-specific clinical features to predict the next treatment step. By focusing on both technical performance and real-world application in a medical setting, this study provides insights for the deployment of CBR-based systems in medical practice.",
isbn="978-3-031-96559-3"
}
Cancer treatment planning is a complex and individualized process due to the variability of patient-specific factors, tumor characteristics, and evolving medical standards. Predicting the next step in diagnosis or therapy remains a significant challenge, due to the high variability and limited availability of structured medical data. Clinical Decision Support Systems (CDSS) offer a promising solution, with Case-Based Reasoning (CBR) standing out for its ability to provide interpretable and transparent recommendations. Unlike black-box machine learning models, CBR leverages past cases to generate predictions by analogy, aligning with the way clinicians naturally reason. In this work, we propose a CBR-based CDSS for skin cancer treatment that integrates medical taxonomies and patient-specific clinical features to predict the next treatment step. By focusing on both technical performance and real-world application in a medical setting, this study provides insights for the deployment of CBR-based systems in medical practice.
Bewertung von Edge-Cloud-Systemen: Vorgehen, Instrumente und Metriken.
Alt, B.; Bast, S.; Babel, M.; Buitmann, J.; Creutz, L.; Dartmann, G.; Ehaus, M.; Engel, T.; Gast, F.; Großegesse, N.; Guldner, A.; Jahnke, N.; Jilg, D.; Körner, M.; Leskow, J.; Mundorf, J.; Naumann, S.; Pehlken, A.; Schacht, M.; Schick, L.; Schlagenhauf, T.; Schultheis, A.; Schörner, K.; Straub, S.; Strüker, J.; Traphöner, R.; and Weiher, M.
2025.
Erstellt im Rahmen der Begleitforschung zum Technologieprogramm "Edge Datenwirtschaft"
Paper
link
bibtex
abstract
@misc{bewertung_edge_cloud_2025,
title = {{Bewertung von Edge-Cloud-Systemen: Vorgehen, Instrumente und Metriken}},
author = {Alt, Benjamin and Bast, Sebastian and Babel, Matthias and Buitmann, Julian and Creutz, Lars and Dartmann, Guido and Ehaus, Marvin and Engel, Thomas and Gast, Fabian and Großegesse, Nina and Guldner, Achim and Jahnke, Nils and Jilg, David and Körner, Marc-Fabian and Leskow, Justus and Mundorf, Johannes and Naumann, Stefan and Pehlken, Alexandra and Schacht, Marvin and Schick, Leo and Schlagenhauf, Tobias and Schultheis, Alexander and Schörner, Karsten and Straub, Sebastian and Strüker, Jens and Traphöner, Ralph and Weiher, Moritz-André},
editor = {Gabriel, Peter and Wittenbrink, Nicole},
institution = {Begleitforschung Edge Datenwirtschaft, im Auftrag des Bundesministeriums für Forschung, Technologie und Raumfahrt (BMFTR)},
publisher = {Institut für Innovation und Technik (iit) in der VDI/VDE Innovation + Technik GmbH},
address = {Berlin},
year = {2025},
note = {Erstellt im Rahmen der Begleitforschung zum Technologieprogramm "Edge Datenwirtschaft"},
url = {https://www.digitale-technologien.de/DT/Redaktion/DE/Downloads/Publikation/EDGE-Datenwirtschaft/2025_1107_EdgeCloudSystem.pdf},
abstract = {Edge-Cloud-Systeme bieten Unternehmen die Chance, innovative datenbasierte Dienste effizient und nachhaltig umzusetzen. Die Orientierungshilfe „Bewertung von Edge-Cloud-Systemen“ (ECS) bietet ein praxisnahes Vorgehensmodell mit Handlungsempfehlungen. Sie beschreibt, wie Edge- und Cloud-Komponenten sinnvoll kombiniert werden können, um Anforderungen wie geringe Latenz, Datenschutz und ökologische Nachhaltigkeit zu erfüllen. Im Mittelpunkt stehen konkrete Bewertungskriterien, Metriken und Instrumente, die helfen, ECS im jeweiligen Anwendungskontext systematisch zu analysieren. Beispiele aus realen Projekten zeigen, wie diese Kriterien operationalisiert werden können.}
}
Edge-Cloud-Systeme bieten Unternehmen die Chance, innovative datenbasierte Dienste effizient und nachhaltig umzusetzen. Die Orientierungshilfe „Bewertung von Edge-Cloud-Systemen“ (ECS) bietet ein praxisnahes Vorgehensmodell mit Handlungsempfehlungen. Sie beschreibt, wie Edge- und Cloud-Komponenten sinnvoll kombiniert werden können, um Anforderungen wie geringe Latenz, Datenschutz und ökologische Nachhaltigkeit zu erfüllen. Im Mittelpunkt stehen konkrete Bewertungskriterien, Metriken und Instrumente, die helfen, ECS im jeweiligen Anwendungskontext systematisch zu analysieren. Beispiele aus realen Projekten zeigen, wie diese Kriterien operationalisiert werden können.
EXAR: A Unified Experience-Grounded Agentic Reasoning Architecture.
Bergmann, R.; Brand, F.; Lenz, M.; and Malburg, L.
In Bichindaritz, I.; and López, B., editor(s),
Case-Based Reasoning Research and Development, volume 15662, of
Lecture Notes in Computer Science, pages 3–17, Cham, 2025. Springer Nature Switzerland
Paper
doi
link
bibtex
abstract
1 download
@inproceedings{Bergmann2025EXARUnifiedExperienceGrounded,
title = {{{EXAR}}: {{A Unified Experience-Grounded Agentic Reasoning Architecture}}},
shorttitle = {{{EXAR}}},
booktitle = {Case-{{Based Reasoning Research}} and {{Development}}},
author = {Bergmann, Ralph and Brand, Florian and Lenz, Mirko and Malburg, Lukas},
editor = {Bichindaritz, Isabelle and L{\'o}pez, Beatriz},
year = {2025},
series = {Lecture {{Notes}} in {{Computer Science}}},
volume = {15662},
pages = {3--17},
publisher = {Springer Nature Switzerland},
address = {Cham},
doi = {10.1007/978-3-031-96559-3_1},
abstract = {Current AI reasoning often relies on static pipelines (like the 4R cycle from Case-Based Reasoning (CBR) or standard Retrieval-Augmented Generation (RAG)) that limit adaptability. We argue it is time for a shift towards dynamic, experience-grounded agentic reasoning. This paper proposes EXAR, a new unified, experience-grounded architecture, conceptualizing reasoning not as a fixed sequence, but as a collaborative process orchestrated among specialized agents. EXAR integrates data and knowledge sources into a persistent Long-Term Memory utilized by diverse reasoning agents, which coordinate themselves via a Short-Term Memory. Governed by an Orchestrator and Meta Learner, EXAR enables flexible, context-aware reasoning strategies that adapt and improve over time, offering a blueprint for next-generation AI.},
isbn = {978-3-031-96559-3},
langid = {english},
url = {https://www.wi2.uni-trier.de/shared/publications/Bergmann2025EXARUnifiedExperienceGrounded.pdf}
}
Current AI reasoning often relies on static pipelines (like the 4R cycle from Case-Based Reasoning (CBR) or standard Retrieval-Augmented Generation (RAG)) that limit adaptability. We argue it is time for a shift towards dynamic, experience-grounded agentic reasoning. This paper proposes EXAR, a new unified, experience-grounded architecture, conceptualizing reasoning not as a fixed sequence, but as a collaborative process orchestrated among specialized agents. EXAR integrates data and knowledge sources into a persistent Long-Term Memory utilized by diverse reasoning agents, which coordinate themselves via a Short-Term Memory. Governed by an Orchestrator and Meta Learner, EXAR enables flexible, context-aware reasoning strategies that adapt and improve over time, offering a blueprint for next-generation AI.
A Framework for Supporting the Iterative Design of CBR Applications.
Jimenez-Diaz, G.; Lenz, M.; Malburg, L.; Díaz-Agudo, Belén; and Bergmann, R.
In Bichindaritz, I.; and López, B., editor(s),
Case-Based Reasoning Research and Development, volume 15662, of
Lecture Notes in Computer Science, pages 252–266, Cham, 2025. Springer Nature Switzerland
Paper
doi
link
bibtex
abstract
@inproceedings{Jimenez-Diaz2025FrameworkSupportingIterative,
title = {A {{Framework}} for~{{Supporting}} the~{{Iterative Design}} of~{{CBR Applications}}},
booktitle = {Case-{{Based Reasoning Research}} and {{Development}}},
author = {{Jimenez-Diaz}, Guillermo and Lenz, Mirko and Malburg, Lukas and {D{\'i}az-Agudo}, Bel{\'e}n and Bergmann, Ralph},
editor = {Bichindaritz, Isabelle and L{\'o}pez, Beatriz},
year = {2025},
series = {Lecture {{Notes}} in {{Computer Science}}},
volume = {15662},
pages = {252--266},
publisher = {Springer Nature Switzerland},
address = {Cham},
doi = {10.1007/978-3-031-96559-3_17},
abstract = {Iterative design is a well-known methodology that involves prototyping, testing, analyzing, and refining products or processes. Rapid prototyping and evaluation are crucial, enabling designers to quickly identify and resolve issues and iteratively improve the design. However, effective iterative design relies on tools that accelerate both prototype creation and visual evaluation. This paper aims to address this challenge by presenting a framework specifically designed to enhance the iterative design process for CBR applications. In this context, we emphasize the manual and knowledge-intensive task of defining similarity measures. Furthermore, we introduce a proof-of-concept implementation of this framework based on the CBRkit toolkit and the SimViz visualization tool. We studied the capabilities to support the iterative design of CBR applications through a case study in the prototypical cars domain.},
isbn = {978-3-031-96559-3},
langid = {english},
url = {https://www.wi2.uni-trier.de/shared/publications/Jimenez-Diaz2025FrameworkSupportingIterative.pdf}
}
Iterative design is a well-known methodology that involves prototyping, testing, analyzing, and refining products or processes. Rapid prototyping and evaluation are crucial, enabling designers to quickly identify and resolve issues and iteratively improve the design. However, effective iterative design relies on tools that accelerate both prototype creation and visual evaluation. This paper aims to address this challenge by presenting a framework specifically designed to enhance the iterative design process for CBR applications. In this context, we emphasize the manual and knowledge-intensive task of defining similarity measures. Furthermore, we introduce a proof-of-concept implementation of this framework based on the CBRkit toolkit and the SimViz visualization tool. We studied the capabilities to support the iterative design of CBR applications through a case study in the prototypical cars domain.
LLsiM: Large Language Models for Similarity Assessment in Case-Based Reasoning.
Lenz, M.; Hoffmann, M.; and Bergmann, R.
In Bichindaritz, I.; and López, B., editor(s),
Case-Based Reasoning Research and Development, volume 15662, of
Lecture Notes in Computer Science, pages 126–141, Cham, 2025. Springer Nature Switzerland
Paper
doi
link
bibtex
abstract
@inproceedings{Lenz2025LLsiMLargeLanguage,
title = {{{LLsiM}}: {{Large Language Models}} for~{{Similarity Assessment}} in~{{Case-Based Reasoning}}},
shorttitle = {{{LLsiM}}},
booktitle = {Case-{{Based Reasoning Research}} and {{Development}}},
author = {Lenz, Mirko and Hoffmann, Maximilian and Bergmann, Ralph},
editor = {Bichindaritz, Isabelle and L{\'o}pez, Beatriz},
year = {2025},
series = {Lecture {{Notes}} in {{Computer Science}}},
volume = {15662},
pages = {126--141},
publisher = {Springer Nature Switzerland},
address = {Cham},
doi = {10.1007/978-3-031-96559-3_9},
abstract = {In Case-Based Reasoning (CBR), past experience is used to solve new problems. Determining the most relevant cases is a crucial aspect of this process and is typically based on one or multiple manually-defined similarity measures, requiring deep domain knowledge. To overcome the knowledge-acquisition bottleneck, we propose the use of Large Language Models (LLMs) to automatically assess similarities between cases. We present three distinct approaches where the model is used for different tasks: (i) to predict similarity scores, (ii) to assess pairwise preferences, and (iii) to automatically configure similarity measures. Our conceptual work is accompanied by an open-source Python implementation that we use to evaluate the approaches on three different domains by comparing them to manually crafted similarity measures. Our results show that directly using LLM-based scores does not align well with the baseline rankings, but letting the LLM automatically configure the measures yields rankings that closely resemble the expert-defined ones.},
isbn = {978-3-031-96559-3},
langid = {english},
url = {https://www.wi2.uni-trier.de/shared/publications/Lenz2025LLsiMLargeLanguage.pdf}
}
In Case-Based Reasoning (CBR), past experience is used to solve new problems. Determining the most relevant cases is a crucial aspect of this process and is typically based on one or multiple manually-defined similarity measures, requiring deep domain knowledge. To overcome the knowledge-acquisition bottleneck, we propose the use of Large Language Models (LLMs) to automatically assess similarities between cases. We present three distinct approaches where the model is used for different tasks: (i) to predict similarity scores, (ii) to assess pairwise preferences, and (iii) to automatically configure similarity measures. Our conceptual work is accompanied by an open-source Python implementation that we use to evaluate the approaches on three different domains by comparing them to manually crafted similarity measures. Our results show that directly using LLM-based scores does not align well with the baseline rankings, but letting the LLM automatically configure the measures yields rankings that closely resemble the expert-defined ones.
Advanced Search Techniques for Determining Optimal Sequences of Adaptation Rules in Process-Oriented Case-Based Reasoning.
Hotz, M.; Malburg, L.; and Bergmann, R.
In Bichindaritz, I.; and Lopez, B., editor(s),
Case-Based Reasoning Research and Development - 33rd International Conference, ICCBR 2025, Biarritz, France, June 30 - July 3, 2025, Proceedings, volume 15662, of
Lecture Notes in Computer Science, pages 236–251, 2025. Springer.
Paper
doi
link
bibtex
abstract
3 downloads
@inproceedings{Hotz_AdvancedSearchForRules_2025,
author = {Maxim Hotz and Lukas Malburg and Ralph Bergmann},
title = {{Advanced Search Techniques for Determining Optimal Sequences of Adaptation Rules in Process-Oriented Case-Based Reasoning}},
booktitle = {Case-Based Reasoning Research and Development - 33rd International Conference, {ICCBR} 2025, Biarritz, France, June 30 - July 3, 2025, Proceedings},
series = {Lecture Notes in Computer Science},
pages = {236--251},
volume = {15662},
editor = {Bichindaritz, Isabelle and Lopez, Beatriz},
publisher = {Springer.},
doi = {10.1007/978-3-031-96559-3\_16},
year = {2025},
keywords = {Process-Oriented Case-Based Reasoning, Adaptive Workflow Management, Rule-Based Adaptation, Constrained Optimization Planning, Genetic Algorithms},
abstract = {The application of adaptation knowledge still poses a major challenge in modern Case-Based Reasoning (CBR) systems. This is especially the case in the subdomain of Process-Oriented Case-Based Reasoning (POCBR), where cases represent procedural experimental knowledge. Current adaptation methods in this field make use of proprietary local search techniques to apply adaptation knowledge, which is inefficient and does not allow exploitation of advanced search and optimization techniques. Therefore, this work presents an approach to transform rule-based adaptation into a planning problem that is solvable with both Constrained Optimization Planning (COP) and Genetic Algorithms (GA). The results of an experimental evaluation indicate that this transformation is feasible, although it does not achieve significant improvements in terms of adaptation quality without fine-tuning the default search parameters. However, integration into state-of-the-art planning frameworks builds the basis for using a variety of additional features and updates, enhancing the modeling and configuration process.},
url = {https://www.wi2.uni-trier.de/shared/publications/2025_ICCBR_AdvancedSearchForRules_HotzEtAl.pdf}
}
The application of adaptation knowledge still poses a major challenge in modern Case-Based Reasoning (CBR) systems. This is especially the case in the subdomain of Process-Oriented Case-Based Reasoning (POCBR), where cases represent procedural experimental knowledge. Current adaptation methods in this field make use of proprietary local search techniques to apply adaptation knowledge, which is inefficient and does not allow exploitation of advanced search and optimization techniques. Therefore, this work presents an approach to transform rule-based adaptation into a planning problem that is solvable with both Constrained Optimization Planning (COP) and Genetic Algorithms (GA). The results of an experimental evaluation indicate that this transformation is feasible, although it does not achieve significant improvements in terms of adaptation quality without fine-tuning the default search parameters. However, integration into state-of-the-art planning frameworks builds the basis for using a variety of additional features and updates, enhancing the modeling and configuration process.
Case-Based Activity Detection from Segmented Internet of Things Data.
Seiger, R.; Schultheis, A.; and Bergmann, R.
In
Case-Based Reasoning Research and Development - 33rd International Conference, ICCBR 2025, Biarritz, France, June 30 - July 3rd, 2025, Proceedings, of
Lecture Notes in Computer Science, pages 438–453, 2025. Springer.
Paper
doi
link
bibtex
abstract
2 downloads
@inproceedings{SeigerSB2025,
author = {Ronny Seiger and Alexander Schultheis and Ralph Bergmann},
title = {{Case-Based Activity Detection from Segmented Internet of Things Data}},
booktitle = {Case-Based Reasoning Research and Development - 33rd International Conference, {ICCBR} 2025, Biarritz, France, June 30 - July 3rd, 2025, Proceedings},
series = {Lecture Notes in Computer Science},
pages = {438--453},
publisher = {Springer.},
year = {2025},
doi = {10.1007/978-3-031-96559-3_29},
keywords = {Temporal Case-Based Reasoning, Time Series Data, Activity Detection, Internet of Things},
abstract = {The use of Internet of Things (IoT) technologies drives the automation of business processes. However, such environments often lack process awareness and corresponding systems to monitor process executions. Due to its too fine-grained nature and variations, the direct use of data from IoT devices for monitoring is problematic, requiring an event abstraction step to lift the data to the business process level. This work investigates the application of Temporal Case-Based Reasoning (TCBR) as a novel experience-based approach to detect process activity executions in IoT data. The proposed TCBR approach uses activity signatures - representations of process and IoT data for an activity prototype - as a case base to classify unknown IoT time series data from a smart factory. A data flow architecture is presented that supports analysts in selecting a suitable activity prototype and evaluating its quality for activity detection. The results enable both, the development of high-quality activity detection services and the identification of improvement opportunities in IoT monitoring systems. The approach is evaluated using data produced by a smart factory. The results indicate that the TCBR methods used are very suitable for detecting activities in this IoT use case.},
url = {https://www.wi2.uni-trier.de/shared/publications/2025_ICCBR_SeigerEtAL.pdf}
}
The use of Internet of Things (IoT) technologies drives the automation of business processes. However, such environments often lack process awareness and corresponding systems to monitor process executions. Due to its too fine-grained nature and variations, the direct use of data from IoT devices for monitoring is problematic, requiring an event abstraction step to lift the data to the business process level. This work investigates the application of Temporal Case-Based Reasoning (TCBR) as a novel experience-based approach to detect process activity executions in IoT data. The proposed TCBR approach uses activity signatures - representations of process and IoT data for an activity prototype - as a case base to classify unknown IoT time series data from a smart factory. A data flow architecture is presented that supports analysts in selecting a suitable activity prototype and evaluating its quality for activity detection. The results enable both, the development of high-quality activity detection services and the identification of improvement opportunities in IoT monitoring systems. The approach is evaluated using data produced by a smart factory. The results indicate that the TCBR methods used are very suitable for detecting activities in this IoT use case.
Integration of Time Series Embedding for Efficient Retrieval in Case-Based Reasoning.
Weich, J.; Schultheis, A.; Hoffmann, M.; and Bergmann, R.
In
Case-Based Reasoning Research and Development - 33rd International Conference, ICCBR 2025, Biarritz, France, June 30 - July 3rd, 2025, Proceedings, of
Lecture Notes in Computer Science, pages 328–344, 2025. Springer.
Paper
doi
link
bibtex
abstract
@inproceedings{WeichSHB2025,
author = {Justin Weich and Alexander Schultheis and Maximilian Hoffmann and Ralph Bergmann},
title = {{Integration of Time Series Embedding for Efficient Retrieval in Case-Based Reasoning}},
booktitle = {Case-Based Reasoning Research and Development - 33rd International Conference, {ICCBR} 2025, Biarritz, France, June 30 - July 3rd, 2025, Proceedings},
series = {Lecture Notes in Computer Science},
pages = {328--344},
publisher = {Springer.},
year = {2025},
doi = {10.1007/978-3-031-96559-3_22},
keywords = {Temporal Case-Based Reasoning, Time Series Data, Time Series Embedding, Time Series Similarity Measure, Siamese Neural Networks},
abstract = {The increasing volume of time series data in Industry 4.0 applications creates substantial challenges for real-time data analysis. Such analyses that are conducted in the research area of Temporal Case-Based Reasoning (TCBR) face performance problems due to complex similarity measures. One potential approach already proven in other domains for addressing these problems is the usage of embedding techniques for time series data, which map these data into a simplified vector representation. Therefore, this paper investigates the integration of time series embedding techniques in the context of Case-Based Reasoning (CBR) to improve retrieval efficiency. Therefore, requirements for the application of embedding techniques in CBR are derived. A systematic literature study identifies possible approaches that are analyzed based on the requirements, with the result that no approach is suitable for the application. Therefore, a novel embedding architecture is proposed, using a Siamese neural network approach that can be trained with similarity values. The architecture is prototypically implemented in the ProCAKE framework and evaluated in an Internet of Things use case from a smart factory. The results demonstrate that the embedding-based retrieval achieves classification performance comparable to traditional similarity measures while significantly reducing retrieval time.},
url = {https://www.wi2.uni-trier.de/shared/publications/2025_ICCBR_WeichEtAL.pdf}
}
The increasing volume of time series data in Industry 4.0 applications creates substantial challenges for real-time data analysis. Such analyses that are conducted in the research area of Temporal Case-Based Reasoning (TCBR) face performance problems due to complex similarity measures. One potential approach already proven in other domains for addressing these problems is the usage of embedding techniques for time series data, which map these data into a simplified vector representation. Therefore, this paper investigates the integration of time series embedding techniques in the context of Case-Based Reasoning (CBR) to improve retrieval efficiency. Therefore, requirements for the application of embedding techniques in CBR are derived. A systematic literature study identifies possible approaches that are analyzed based on the requirements, with the result that no approach is suitable for the application. Therefore, a novel embedding architecture is proposed, using a Siamese neural network approach that can be trained with similarity values. The architecture is prototypically implemented in the ProCAKE framework and evaluated in an Internet of Things use case from a smart factory. The results demonstrate that the embedding-based retrieval achieves classification performance comparable to traditional similarity measures while significantly reducing retrieval time.
Challenges in Data Quality Management for IoT-Enhanced Event Logs.
Bertrand, Y.; Schultheis, A.; Malburg, L.; Grüger, J.; Serral Asensio, E.; and Bergmann, R.
In
Research Challenges in Information Science - 19th International Conference, RCIS 2025, Guimarães, Seville, Spain, 20 - 23 May, 2025, Proceedings, of
Lecture Notes in Business Information Processing, pages 20–36, 2025. Springer
Paper
doi
link
bibtex
abstract
3 downloads
@inproceedings{BertrandSMGSAB2025,
author = {Yannis Bertrand and Alexander Schultheis and Lukas Malburg and Joscha Grüger and Estefan{\'{\i}}a {Serral Asensio} and Ralph Bergmann},
title = {{Challenges in Data Quality Management for IoT-Enhanced Event Logs}},
booktitle = {Research Challenges in Information Science - 19th International Conference, {RCIS} 2025, Guimar{\~{a}}es, Seville, Spain, 20 - 23 May, 2025, Proceedings},
series = {Lecture Notes in Business Information Processing},
publisher = {Springer},
year = {2025},
doi = {10.1007/978-3-031-92474-3_2},
pages = {20--36},
isbn = {978-3-031-92474-3},
keywords = {Business Process Management, Internet of Things, Data Quality Management, Data Quality Issues},
abstract = {Modern organizations make frequent use of Internet of Things (IoT) devices, such as sensors and actuators, to monitor and support their so-called IoT-enhanced Business Processes (BPs). These IoT devices collect vast amounts of data which, when processed appropriately, can yield crucial insights into the working of the BPs. However, IoT data, such as sensor data, is notoriously of poor quality, e.g., suffering from noise or having some missing data points. These problems are referred to as Data Quality Issues (DQIs), which often interfere with the analysis of IoT data in an industrial context. In this paper, we present a list of challenges that have to be tackled to achieve Data Quality (DQ) management in IoT-enhanced event logs. These challenges are derived and refined by leveraging expert knowledge and experience in DQIs within a focus group interview. In addition, we provide directions for solutions to these challenges based on input from the focus group interview and the literature. Finally, we discuss the challenges and their impact on typical DQ management tasks. The insights provided can help guide future research to achieve better DQ in event logs of IoT-enhanced BPs.},
url = {https://www.wi2.uni-trier.de/shared/publications/2025_RCIS_BertrandEtAl.pdf}
}
Modern organizations make frequent use of Internet of Things (IoT) devices, such as sensors and actuators, to monitor and support their so-called IoT-enhanced Business Processes (BPs). These IoT devices collect vast amounts of data which, when processed appropriately, can yield crucial insights into the working of the BPs. However, IoT data, such as sensor data, is notoriously of poor quality, e.g., suffering from noise or having some missing data points. These problems are referred to as Data Quality Issues (DQIs), which often interfere with the analysis of IoT data in an industrial context. In this paper, we present a list of challenges that have to be tackled to achieve Data Quality (DQ) management in IoT-enhanced event logs. These challenges are derived and refined by leveraging expert knowledge and experience in DQIs within a focus group interview. In addition, we provide directions for solutions to these challenges based on input from the focus group interview and the literature. Finally, we discuss the challenges and their impact on typical DQ management tasks. The insights provided can help guide future research to achieve better DQ in event logs of IoT-enhanced BPs.
Combining informed data-driven anomaly detection with knowledge graphs for root cause analysis in predictive maintenance.
Klein, P.; Malburg, L.; and Bergmann, R.
Engineering Applications of Artificial Intelligence, 145: 110152. 2025.
Paper
doi
link
bibtex
abstract
4 downloads
@article{Klein_InformedAnomalyDetection_2025,
title = {{Combining informed data-driven anomaly detection with knowledge graphs for root cause analysis in predictive maintenance}},
journal = {Engineering Applications of Artificial Intelligence},
volume = {145},
pages = {110152},
year = {2025},
issn = {0952-1976},
doi = {10.1016/j.engappai.2025.110152},
author = {Patrick Klein and Lukas Malburg and Ralph Bergmann},
keywords = {Hybrid artificial intelligence, Knowledge-based diagnosis, Data-driven anomaly detection, Predictive maintenance, Knowledge graph, Semantic web technologies},
abstract = {Industry 4.0 has facilitated the access to sensor and actuator data from manufacturing systems, leading to studies on data-driven anomaly detection, but limited attention has been paid to finding root causes and automating this process using formalized expert knowledge. This is crucial due to the scarcity of qualified engineers and the time-consuming nature of diagnosing issues in large production systems. To address this gap, we present a framework that combines data-driven anomaly detection with a knowledge graph that provides domain knowledge by leveraging typical explanations of such models (i.e., data streams potentially caused the detection) for further diagnosis. The framework’s usefulness to infer affected components or data set labels has been evaluated using two deep anomaly detection approaches. For knowledge-based diagnosis, three query strategies that utilize various knowledge graph relationships are implemented through three Artificial Intelligence (AI) techniques. The proposed anomaly detection approach, informed by integrating expert knowledge via the graph structure of the knowledge graph and node embeddings for encoding time series, outperforms baselines and a deep autoencoder in detecting anomalies and in identifying anomalous data streams. In subsequent diagnosis, it achieves the best performance on a complete knowledge graph in combination with a graph pattern matching query by identifying the label or affected component in 60% of detected anomalies by providing 4.1 labels or 2.3 components until the correct one is identified. In case of a corrupted one, Symbolic-Driven Neural Reasoning (SDNR) and Case-Based Reasoning (CBR) with knowledge graph embeddings demonstrate advantages by halving the number of incorrect labels and unaffected components.},
url = {https://www.wi2.uni-trier.de/shared/publications/2025_EAAI_KleinEtAl.pdf}
}
Industry 4.0 has facilitated the access to sensor and actuator data from manufacturing systems, leading to studies on data-driven anomaly detection, but limited attention has been paid to finding root causes and automating this process using formalized expert knowledge. This is crucial due to the scarcity of qualified engineers and the time-consuming nature of diagnosing issues in large production systems. To address this gap, we present a framework that combines data-driven anomaly detection with a knowledge graph that provides domain knowledge by leveraging typical explanations of such models (i.e., data streams potentially caused the detection) for further diagnosis. The framework’s usefulness to infer affected components or data set labels has been evaluated using two deep anomaly detection approaches. For knowledge-based diagnosis, three query strategies that utilize various knowledge graph relationships are implemented through three Artificial Intelligence (AI) techniques. The proposed anomaly detection approach, informed by integrating expert knowledge via the graph structure of the knowledge graph and node embeddings for encoding time series, outperforms baselines and a deep autoencoder in detecting anomalies and in identifying anomalous data streams. In subsequent diagnosis, it achieves the best performance on a complete knowledge graph in combination with a graph pattern matching query by identifying the label or affected component in 60% of detected anomalies by providing 4.1 labels or 2.3 components until the correct one is identified. In case of a corrupted one, Symbolic-Driven Neural Reasoning (SDNR) and Case-Based Reasoning (CBR) with knowledge graph embeddings demonstrate advantages by halving the number of incorrect labels and unaffected components.
Augmentation of Semantic Processes for Deep Learning Applications.
Hoffmann, M.; Malburg, L.; and Bergmann, R.
Applied Artificial Intelligence, 39(1): 2506788. 2025.
Paper
doi
link
bibtex
4 downloads
@article{Hoffmann_Process_Augmentation_2025,
author = {Maximilian Hoffmann and Lukas Malburg and Ralph Bergmann},
title = {{Augmentation of Semantic Processes for Deep Learning Applications}},
journal = {Applied Artificial Intelligence},
volume = {39},
number = {1},
pages = {2506788},
year = {2025},
publisher = {Taylor \& Francis},
doi = {10.1080/08839514.2025.2506788},
URL = {https://www.wi2.uni-trier.de/shared/publications/2025_AAI_HoffmannEtAl.pdf}
}
ArgueMapper Assistant: Interactive Argument Mining Using Generative Language Models.
Lenz, M.; and Bergmann, R.
In Bramer, M.; and Stahl, F., editor(s),
Artificial Intelligence XLI, volume 15446, of
Lecture Notes in Computer Science, pages 189–203, Cham, 2025. Springer Nature Switzerland
Paper
doi
link
bibtex
abstract
@inproceedings{Lenz2025ArgueMapperAssistantInteractive,
title = {{{ArgueMapper Assistant}}: {{Interactive Argument Mining Using Generative Language Models}}},
shorttitle = {{{ArgueMapper Assistant}}},
booktitle = {Artificial {{Intelligence XLI}}},
author = {Lenz, Mirko and Bergmann, Ralph},
editor = {Bramer, Max and Stahl, Frederic},
year = {2025},
series = {Lecture {{Notes}} in {{Computer Science}}},
volume = {15446},
pages = {189--203},
publisher = {Springer Nature Switzerland},
address = {Cham},
doi = {10.1007/978-3-031-77915-2_14},
abstract = {Structured arguments are a valuable resource for analyzing and understanding complex topics. However, manual annotation is time-consuming and often not feasible for large datasets, and automated approaches are less accurate. To address this issue, we propose an interactive argument mining system that takes advantage of generative language models to support humans in the creation of argument graphs. We present the open source ArgueMapper Assistant featuring two prompting strategies and evaluate it on a real-world news dataset. The resulting corpus containing 88 argument graphs is publicly available as well. With generative models, the annotation time is reduced by about 20\% while the number of errors is slightly increased (mostly due to missing argumentative units and wrong relation types). A survey provides insights into the usefulness and reliability of the assistant features and shows that participants prefer to use the assistant in the future.},
isbn = {978-3-031-77915-2},
langid = {english},
url = {https://www.wi2.uni-trier.de/shared/publications/Lenz2025ArgueMapperAssistantInteractive.pdf}
}
Structured arguments are a valuable resource for analyzing and understanding complex topics. However, manual annotation is time-consuming and often not feasible for large datasets, and automated approaches are less accurate. To address this issue, we propose an interactive argument mining system that takes advantage of generative language models to support humans in the creation of argument graphs. We present the open source ArgueMapper Assistant featuring two prompting strategies and evaluate it on a real-world news dataset. The resulting corpus containing 88 argument graphs is publicly available as well. With generative models, the annotation time is reduced by about 20% while the number of errors is slightly increased (mostly due to missing argumentative units and wrong relation types). A survey provides insights into the usefulness and reliability of the assistant features and shows that participants prefer to use the assistant in the future.
Hybrid AI for Process Management - Improving Similarity Assessment in Process-Oriented Case-Based Reasoning via Deep Learning.
Hoffmann, M.
Ph.D. Thesis, University of Trier, Germany, 2025.
Licensed under Creative Commons 4.0 BY-NC.
Paper
link
bibtex
@phdthesis{Hoffmann25,
title = {{Hybrid AI for Process Management - Improving Similarity Assessment in Process-Oriented Case-Based Reasoning via Deep Learning}},
author = {Hoffmann, Maximilian},
year = 2025,
url = {https://ubt.opus.hbz-nrw.de/files/2518/Dissertation_Hoffmann.pdf},
school = {University of Trier, Germany},
note = {Licensed under Creative Commons 4.0 BY-NC.}
}