Legal Prompting: Teaching a Language Model to Think Like a Lawyer. Yu, F., Quartey, L., & Schilder, F. December, 2022. arXiv:2212.01326 [cs]
Paper doi abstract bibtex Large language models that are capable of zero or few-shot prompting approaches have given rise to the new research area of prompt engineering. Recent advances showed that for example Chain-of-Thought (CoT) prompts can improve arithmetic or common sense tasks significantly. We explore how such approaches fare with legal reasoning tasks and take the COLIEE entailment task based on the Japanese Bar exam for testing zero-shot/few-shot and fine-tuning approaches. Our findings show that while CoT prompting and fine-tuning with explanations approaches show improvements, the best results are produced by prompts that are derived from specific legal reasoning techniques such as IRAC (Issue, Rule, Application, Conclusion). Based on our experiments we improve the 2021 best result from 0.7037 accuracy to 0.8148 accuracy and beat the 2022 best system of 0.6789 accuracy with an accuracy of 0.7431.
@misc{yuLegalPromptingTeaching2022,
title = {Legal {Prompting}: {Teaching} a {Language} {Model} to {Think} {Like} a {Lawyer}},
shorttitle = {Legal {Prompting}},
url = {http://arxiv.org/abs/2212.01326},
doi = {10.48550/arXiv.2212.01326},
abstract = {Large language models that are capable of zero or few-shot prompting approaches have given rise to the new research area of prompt engineering. Recent advances showed that for example Chain-of-Thought (CoT) prompts can improve arithmetic or common sense tasks significantly. We explore how such approaches fare with legal reasoning tasks and take the COLIEE entailment task based on the Japanese Bar exam for testing zero-shot/few-shot and fine-tuning approaches. Our findings show that while CoT prompting and fine-tuning with explanations approaches show improvements, the best results are produced by prompts that are derived from specific legal reasoning techniques such as IRAC (Issue, Rule, Application, Conclusion). Based on our experiments we improve the 2021 best result from 0.7037 accuracy to 0.8148 accuracy and beat the 2022 best system of 0.6789 accuracy with an accuracy of 0.7431.},
urldate = {2024-07-29},
publisher = {arXiv},
author = {Yu, Fangyi and Quartey, Lee and Schilder, Frank},
month = dec,
year = {2022},
note = {arXiv:2212.01326 [cs]},
keywords = {Computer Science - Artificial Intelligence, Computer Science - Computation and Language, I.2.7},
}
Downloads: 0
{"_id":"TkrxH8uTjWf2xhQtB","bibbaseid":"yu-quartey-schilder-legalpromptingteachingalanguagemodeltothinklikealawyer-2022","author_short":["Yu, F.","Quartey, L.","Schilder, F."],"bibdata":{"bibtype":"misc","type":"misc","title":"Legal Prompting: Teaching a Language Model to Think Like a Lawyer","shorttitle":"Legal Prompting","url":"http://arxiv.org/abs/2212.01326","doi":"10.48550/arXiv.2212.01326","abstract":"Large language models that are capable of zero or few-shot prompting approaches have given rise to the new research area of prompt engineering. Recent advances showed that for example Chain-of-Thought (CoT) prompts can improve arithmetic or common sense tasks significantly. We explore how such approaches fare with legal reasoning tasks and take the COLIEE entailment task based on the Japanese Bar exam for testing zero-shot/few-shot and fine-tuning approaches. Our findings show that while CoT prompting and fine-tuning with explanations approaches show improvements, the best results are produced by prompts that are derived from specific legal reasoning techniques such as IRAC (Issue, Rule, Application, Conclusion). Based on our experiments we improve the 2021 best result from 0.7037 accuracy to 0.8148 accuracy and beat the 2022 best system of 0.6789 accuracy with an accuracy of 0.7431.","urldate":"2024-07-29","publisher":"arXiv","author":[{"propositions":[],"lastnames":["Yu"],"firstnames":["Fangyi"],"suffixes":[]},{"propositions":[],"lastnames":["Quartey"],"firstnames":["Lee"],"suffixes":[]},{"propositions":[],"lastnames":["Schilder"],"firstnames":["Frank"],"suffixes":[]}],"month":"December","year":"2022","note":"arXiv:2212.01326 [cs]","keywords":"Computer Science - Artificial Intelligence, Computer Science - Computation and Language, I.2.7","bibtex":"@misc{yuLegalPromptingTeaching2022,\n\ttitle = {Legal {Prompting}: {Teaching} a {Language} {Model} to {Think} {Like} a {Lawyer}},\n\tshorttitle = {Legal {Prompting}},\n\turl = {http://arxiv.org/abs/2212.01326},\n\tdoi = {10.48550/arXiv.2212.01326},\n\tabstract = {Large language models that are capable of zero or few-shot prompting approaches have given rise to the new research area of prompt engineering. Recent advances showed that for example Chain-of-Thought (CoT) prompts can improve arithmetic or common sense tasks significantly. We explore how such approaches fare with legal reasoning tasks and take the COLIEE entailment task based on the Japanese Bar exam for testing zero-shot/few-shot and fine-tuning approaches. Our findings show that while CoT prompting and fine-tuning with explanations approaches show improvements, the best results are produced by prompts that are derived from specific legal reasoning techniques such as IRAC (Issue, Rule, Application, Conclusion). Based on our experiments we improve the 2021 best result from 0.7037 accuracy to 0.8148 accuracy and beat the 2022 best system of 0.6789 accuracy with an accuracy of 0.7431.},\n\turldate = {2024-07-29},\n\tpublisher = {arXiv},\n\tauthor = {Yu, Fangyi and Quartey, Lee and Schilder, Frank},\n\tmonth = dec,\n\tyear = {2022},\n\tnote = {arXiv:2212.01326 [cs]},\n\tkeywords = {Computer Science - Artificial Intelligence, Computer Science - Computation and Language, I.2.7},\n}\n\n","author_short":["Yu, F.","Quartey, L.","Schilder, F."],"key":"yuLegalPromptingTeaching2022","id":"yuLegalPromptingTeaching2022","bibbaseid":"yu-quartey-schilder-legalpromptingteachingalanguagemodeltothinklikealawyer-2022","role":"author","urls":{"Paper":"http://arxiv.org/abs/2212.01326"},"keyword":["Computer Science - Artificial Intelligence","Computer Science - Computation and Language","I.2.7"],"metadata":{"authorlinks":{}}},"bibtype":"misc","biburl":"https://bibbase.org/f/vr5ooa48xeYes5KDD/ailaw.bib","dataSources":["7FkfQdR6FwGXEAZFa","QHxajSYCsDY5s5PEr"],"keywords":["computer science - artificial intelligence","computer science - computation and language","i.2.7"],"search_terms":["legal","prompting","teaching","language","model","think","lawyer","yu","quartey","schilder"],"title":"Legal Prompting: Teaching a Language Model to Think Like a Lawyer","year":2022}