LLsiM: Large Language Models for Similarity Assessment in Case-Based Reasoning. Lenz, M., Hoffmann, M., & Bergmann, R. In Bichindaritz, I. & López, B., editors, Case-Based Reasoning Research and Development, volume 15662, of Lecture Notes in Computer Science, pages 126–141, Cham, 2025. Springer Nature Switzerland.
Paper doi abstract bibtex 1 download In Case-Based Reasoning (CBR), past experience is used to solve new problems. Determining the most relevant cases is a crucial aspect of this process and is typically based on one or multiple manually-defined similarity measures, requiring deep domain knowledge. To overcome the knowledge-acquisition bottleneck, we propose the use of Large Language Models (LLMs) to automatically assess similarities between cases. We present three distinct approaches where the model is used for different tasks: (i) to predict similarity scores, (ii) to assess pairwise preferences, and (iii) to automatically configure similarity measures. Our conceptual work is accompanied by an open-source Python implementation that we use to evaluate the approaches on three different domains by comparing them to manually crafted similarity measures. Our results show that directly using LLM-based scores does not align well with the baseline rankings, but letting the LLM automatically configure the measures yields rankings that closely resemble the expert-defined ones.
@inproceedings{Lenz2025LLsiMLargeLanguage,
title = {{{LLsiM}}: {{Large Language Models}} for~{{Similarity Assessment}} in~{{Case-Based Reasoning}}},
shorttitle = {{{LLsiM}}},
booktitle = {Case-{{Based Reasoning Research}} and {{Development}}},
author = {Lenz, Mirko and Hoffmann, Maximilian and Bergmann, Ralph},
editor = {Bichindaritz, Isabelle and L{\'o}pez, Beatriz},
year = {2025},
series = {Lecture {{Notes}} in {{Computer Science}}},
volume = {15662},
pages = {126--141},
publisher = {Springer Nature Switzerland},
address = {Cham},
doi = {10.1007/978-3-031-96559-3_9},
abstract = {In Case-Based Reasoning (CBR), past experience is used to solve new problems. Determining the most relevant cases is a crucial aspect of this process and is typically based on one or multiple manually-defined similarity measures, requiring deep domain knowledge. To overcome the knowledge-acquisition bottleneck, we propose the use of Large Language Models (LLMs) to automatically assess similarities between cases. We present three distinct approaches where the model is used for different tasks: (i) to predict similarity scores, (ii) to assess pairwise preferences, and (iii) to automatically configure similarity measures. Our conceptual work is accompanied by an open-source Python implementation that we use to evaluate the approaches on three different domains by comparing them to manually crafted similarity measures. Our results show that directly using LLM-based scores does not align well with the baseline rankings, but letting the LLM automatically configure the measures yields rankings that closely resemble the expert-defined ones.},
isbn = {978-3-031-96559-3},
langid = {english},
url = {https://www.wi2.uni-trier.de/shared/publications/Lenz2025LLsiMLargeLanguage.pdf}
}
Downloads: 1
{"_id":"kZCH9Kj398YHQGvTY","bibbaseid":"lenz-hoffmann-bergmann-llsimlargelanguagemodelsforsimilarityassessmentincasebasedreasoning-2025","author_short":["Lenz, M.","Hoffmann, M.","Bergmann, R."],"bibdata":{"bibtype":"inproceedings","type":"inproceedings","title":"LLsiM: Large Language Models for Similarity Assessment in Case-Based Reasoning","shorttitle":"LLsiM","booktitle":"Case-Based Reasoning Research and Development","author":[{"propositions":[],"lastnames":["Lenz"],"firstnames":["Mirko"],"suffixes":[]},{"propositions":[],"lastnames":["Hoffmann"],"firstnames":["Maximilian"],"suffixes":[]},{"propositions":[],"lastnames":["Bergmann"],"firstnames":["Ralph"],"suffixes":[]}],"editor":[{"propositions":[],"lastnames":["Bichindaritz"],"firstnames":["Isabelle"],"suffixes":[]},{"propositions":[],"lastnames":["López"],"firstnames":["Beatriz"],"suffixes":[]}],"year":"2025","series":"Lecture Notes in Computer Science","volume":"15662","pages":"126–141","publisher":"Springer Nature Switzerland","address":"Cham","doi":"10.1007/978-3-031-96559-3_9","abstract":"In Case-Based Reasoning (CBR), past experience is used to solve new problems. Determining the most relevant cases is a crucial aspect of this process and is typically based on one or multiple manually-defined similarity measures, requiring deep domain knowledge. To overcome the knowledge-acquisition bottleneck, we propose the use of Large Language Models (LLMs) to automatically assess similarities between cases. We present three distinct approaches where the model is used for different tasks: (i) to predict similarity scores, (ii) to assess pairwise preferences, and (iii) to automatically configure similarity measures. Our conceptual work is accompanied by an open-source Python implementation that we use to evaluate the approaches on three different domains by comparing them to manually crafted similarity measures. Our results show that directly using LLM-based scores does not align well with the baseline rankings, but letting the LLM automatically configure the measures yields rankings that closely resemble the expert-defined ones.","isbn":"978-3-031-96559-3","langid":"english","url":"https://www.wi2.uni-trier.de/shared/publications/Lenz2025LLsiMLargeLanguage.pdf","bibtex":"@inproceedings{Lenz2025LLsiMLargeLanguage,\n title = {{{LLsiM}}: {{Large Language Models}} for~{{Similarity Assessment}} in~{{Case-Based Reasoning}}},\n shorttitle = {{{LLsiM}}},\n booktitle = {Case-{{Based Reasoning Research}} and {{Development}}},\n author = {Lenz, Mirko and Hoffmann, Maximilian and Bergmann, Ralph},\n editor = {Bichindaritz, Isabelle and L{\\'o}pez, Beatriz},\n year = {2025},\n series = {Lecture {{Notes}} in {{Computer Science}}},\n volume = {15662},\n pages = {126--141},\n publisher = {Springer Nature Switzerland},\n address = {Cham},\n doi = {10.1007/978-3-031-96559-3_9},\n abstract = {In Case-Based Reasoning (CBR), past experience is used to solve new problems. Determining the most relevant cases is a crucial aspect of this process and is typically based on one or multiple manually-defined similarity measures, requiring deep domain knowledge. To overcome the knowledge-acquisition bottleneck, we propose the use of Large Language Models (LLMs) to automatically assess similarities between cases. We present three distinct approaches where the model is used for different tasks: (i) to predict similarity scores, (ii) to assess pairwise preferences, and (iii) to automatically configure similarity measures. Our conceptual work is accompanied by an open-source Python implementation that we use to evaluate the approaches on three different domains by comparing them to manually crafted similarity measures. Our results show that directly using LLM-based scores does not align well with the baseline rankings, but letting the LLM automatically configure the measures yields rankings that closely resemble the expert-defined ones.},\n isbn = {978-3-031-96559-3},\n langid = {english},\n url = {https://www.wi2.uni-trier.de/shared/publications/Lenz2025LLsiMLargeLanguage.pdf}\n}\n\n\n\n","author_short":["Lenz, M.","Hoffmann, M.","Bergmann, R."],"editor_short":["Bichindaritz, I.","López, B."],"key":"Lenz2025LLsiMLargeLanguage","id":"Lenz2025LLsiMLargeLanguage","bibbaseid":"lenz-hoffmann-bergmann-llsimlargelanguagemodelsforsimilarityassessmentincasebasedreasoning-2025","role":"author","urls":{"Paper":"https://www.wi2.uni-trier.de/shared/publications/Lenz2025LLsiMLargeLanguage.pdf"},"metadata":{"authorlinks":{}},"downloads":1},"bibtype":"inproceedings","biburl":"https://web.wi2.uni-trier.de/publications/WI2Publikationen.bib","dataSources":["MSp3DzP4ToPojqkFy","CnwPa99ZchEF4SZgh"],"keywords":[],"search_terms":["llsim","large","language","models","similarity","assessment","case","based","reasoning","lenz","hoffmann","bergmann"],"title":"LLsiM: Large Language Models for Similarity Assessment in Case-Based Reasoning","year":2025,"downloads":1}