Skip to main navigation Skip to search Skip to main content

UniPar: A Unified LLM-Based Framework for Parallel and Accelerated Code Translation in HPC

  • Tomer Bitan
  • , Tal Kadosh
  • , Erel Kaplan
  • , Shira Meiri
  • , Le Chen
  • , Peter Morales
  • , Niranjan Hasabnis
  • , Gal Oren

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Translating programs between various parallel programming languages is an important problem in the high-performance computing (HPC) community, with implications for industry and academia. Existing tools for this problem are either too narrow in scope (translate between specific languages) and/or outdated (requiring maintenance). Recent explosive growth in the popularity of large language models (LLMs) and their ability to generate and translate code offers a potential alternative approach. Toward that end, we first need to systematically evaluate the ability of LLMs to translate between parallel languages.In this work, we introduce UniPar, a systematic evaluation framework for LLM-based parallel code translation. Specifically, in this work, we target translations between serial code, CUDA, and OpenMP. Our goal is to assess how well current instruction-tuned LLMs - specifically GPT-4o-mini and LLaMA-3.3-70B-Instruct - can be used out of the box or enhanced through known strategies. We evaluated four major usage modes: hyperparameter optimization for decoding, zero- and few-shot prompting, supervised fine-tuning, and iterative feedback through compiler-based repair. As a part of the evaluation, we construct a new dataset called ParaTrans, covering both serial-to-parallel translation and cross-paradigm transformations.Our findings reveal that while off-the-shelf models struggle under the default settings (e.g., GPT-4o-mini achieves only 46% compilation and 15% functional correctness), our UniPar methodology - combining fine-tuning, hyperparameter tuning, and compiler-guided repair - improves performance by up to 2X (69% compilation and 33% correctness). We believe that our findings will provide useful insights for researchers to further improve LLMs for the parallel language translation problem.UniPar source code and ParaTrans dataset are available at our GitHub repository.

Original languageEnglish
Title of host publication2025 IEEE High Performance Extreme Computing Conference, HPEC 2025
PublisherInstitute of Electrical and Electronics Engineers
ISBN (Electronic)9798331578442
DOIs
StatePublished - 1 Jan 2025
Event2025 IEEE High Performance Extreme Computing Conference, HPEC 2025 - Virtual, Online
Duration: 15 Sep 202519 Sep 2025

Publication series

Name2025 IEEE High Performance Extreme Computing Conference, HPEC 2025

Conference

Conference2025 IEEE High Performance Extreme Computing Conference, HPEC 2025
CityVirtual, Online
Period15/09/2519/09/25

ASJC Scopus subject areas

  • Artificial Intelligence
  • Computational Theory and Mathematics
  • Computer Science Applications
  • Hardware and Architecture
  • Computational Mathematics
  • Control and Optimization

Fingerprint

Dive into the research topics of 'UniPar: A Unified LLM-Based Framework for Parallel and Accelerated Code Translation in HPC'. Together they form a unique fingerprint.

Cite this