Polyglot: An Extensible Framework to Benchmark Code Translation with LLMs. Vieira, M., Shah, P. A., Shah, B., & Krasniqi, R. In 2025 40th IEEE/ACM International Conference on Automated Software Engineering (ASE), pages 2363–2375, 2025. ISSN: 2643-1572
Polyglot: An Extensible Framework to Benchmark Code Translation with LLMs [link]Paper  doi  abstract   bibtex   
Large Language Models (LLMs) show great potential for automating code-related tasks. However, sound assessments are necessary to understand their true capabilities, particularly in code translation, where reliability is crucial. We introduce Polyglot, an automated, multi-language framework for evaluating the translation quality of LLMs between different programming languages. Leveraging the IBM CodeNet Project, an extensive collection of coding problems in multiple languages, we assess translation quality using syntactic correctness, execution reliability, semantic preservation, and static code metrics. Our evaluation focuses on translating C to Java, Python, and Rust, languages that follow distinct paradigms and represent alternatives to modernize C-based systems. We evaluate open-source LLMs using three prompting strategies to understand the impact on translation performance. Our findings highlight that while LLMs show promising results for simple code translation, their limitations regarding complex logic and distinct language paradigms require further analysis.
@inproceedings{vieira_polyglot_2025,
	title = {Polyglot: {An} {Extensible} {Framework} to {Benchmark} {Code} {Translation} with {LLMs}},
	shorttitle = {Polyglot},
	url = {https://ieeexplore.ieee.org/document/11334550},
	doi = {10.1109/ASE63991.2025.00195},
	abstract = {Large Language Models (LLMs) show great potential for automating code-related tasks. However, sound assessments are necessary to understand their true capabilities, particularly in code translation, where reliability is crucial. We introduce Polyglot, an automated, multi-language framework for evaluating the translation quality of LLMs between different programming languages. Leveraging the IBM CodeNet Project, an extensive collection of coding problems in multiple languages, we assess translation quality using syntactic correctness, execution reliability, semantic preservation, and static code metrics. Our evaluation focuses on translating C to Java, Python, and Rust, languages that follow distinct paradigms and represent alternatives to modernize C-based systems. We evaluate open-source LLMs using three prompting strategies to understand the impact on translation performance. Our findings highlight that while LLMs show promising results for simple code translation, their limitations regarding complex logic and distinct language paradigms require further analysis.},
	urldate = {2026-02-02},
	booktitle = {2025 40th {IEEE}/{ACM} {International} {Conference} on {Automated} {Software} {Engineering} ({ASE})},
	author = {Vieira, Marco and Shah, Priyam Ashish and Shah, Bhavain and Krasniqi, Rrezarta},
	year = {2025},
	note = {ISSN: 2643-1572},
	keywords = {Conference Full Papers},
	pages = {2363--2375},
}

Downloads: 0