ChatGPT may Pass the Bar Exam soon, but has a Long Way to Go for the LexGLUE benchmark. Chalkidis, I. March, 2023. arXiv:2304.12202 [cs]
ChatGPT may Pass the Bar Exam soon, but has a Long Way to Go for the LexGLUE benchmark [link]Paper  doi  abstract   bibtex   
Following the hype around OpenAI's ChatGPT conversational agent, the last straw in the recent development of Large Language Models (LLMs) that demonstrate emergent unprecedented zero-shot capabilities, we audit the latest OpenAI's GPT-3.5 model, `gpt-3.5-turbo', the first available ChatGPT model, in the LexGLUE benchmark in a zero-shot fashion providing examples in a templated instruction-following format. The results indicate that ChatGPT achieves an average micro-F1 score of 47.6% across LexGLUE tasks, surpassing the baseline guessing rates. Notably, the model performs exceptionally well in some datasets, achieving micro-F1 scores of 62.8% and 70.2% in the ECtHR B and LEDGAR datasets, respectively. The code base and model predictions are available for review on https://github.com/coastalcph/zeroshot_lexglue.
@misc{chalkidisChatGPTMayPass2023,
	title = {{ChatGPT} may {Pass} the {Bar} {Exam} soon, but has a {Long} {Way} to {Go} for the {LexGLUE} benchmark},
	url = {http://arxiv.org/abs/2304.12202},
	doi = {10.48550/arXiv.2304.12202},
	abstract = {Following the hype around OpenAI's ChatGPT conversational agent, the last straw in the recent development of Large Language Models (LLMs) that demonstrate emergent unprecedented zero-shot capabilities, we audit the latest OpenAI's GPT-3.5 model, `gpt-3.5-turbo', the first available ChatGPT model, in the LexGLUE benchmark in a zero-shot fashion providing examples in a templated instruction-following format. The results indicate that ChatGPT achieves an average micro-F1 score of 47.6\% across LexGLUE tasks, surpassing the baseline guessing rates. Notably, the model performs exceptionally well in some datasets, achieving micro-F1 scores of 62.8\% and 70.2\% in the ECtHR B and LEDGAR datasets, respectively. The code base and model predictions are available for review on https://github.com/coastalcph/zeroshot\_lexglue.},
	urldate = {2023-06-12},
	publisher = {arXiv},
	author = {Chalkidis, Ilias},
	month = mar,
	year = {2023},
	note = {arXiv:2304.12202 [cs]},
	keywords = {Computer Science - Computation and Language},
	annote = {Comment: Working paper},
}

Downloads: 0