An Empirical Comparison of Code Generation Approaches for Ansible

Benjamin Darnell, Hetarth Chopra, Aaron Councilman, David Grove, Yu Xiong Wang, Vikram Adve

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

The rapid proliferation of LLM-based programming assistants has enabled fast and accurate automatic code generation for general purpose programming languages. Domain-specific languages like Ansible, a DSL for IT Automation, have seen a lack of support despite being critical to many fields, due to limited public-domain code for training models and a lack of interest from tool developers. To address this issue, we collect a novel dataset of permissively licensed Ansible code, and use it to create Warp, an LLM for code fine-tuned to produce Ansible tasks from a natural language prompt. We evaluate state-of-the-art tools for LLM-based code generation models, comparing multiple common strategies, including fine-tuning base models on Ansible code and retrieval-augmented-generation using documentation, in order to understand challenges with existing methodology and identify future research directions to enable better code generation for DSLs.

Original languageEnglish (US)
Title of host publicationProceedings - 2024 IEEE/ACM 2nd International Workshop on Interpretability, Robustness, and Benchmarking in Neural Software Engineering, InteNSE 2024
PublisherAssociation for Computing Machinery
Pages1-6
Number of pages6
ISBN (Electronic)9798400705649
DOIs
StatePublished - Apr 15 2024
Event2nd International Workshop on Interpretability, Robustness, and Benchmarking in Neural Software Engineering, InteNSE 2024, co-located with ICSE 2024 - Lisbon, Portugal
Duration: Apr 15 2024 → …

Publication series

NameProceedings - 2024 IEEE/ACM 2nd International Workshop on Interpretability, Robustness, and Benchmarking in Neural Software Engineering, InteNSE 2024

Conference

Conference2nd International Workshop on Interpretability, Robustness, and Benchmarking in Neural Software Engineering, InteNSE 2024, co-located with ICSE 2024
Country/TerritoryPortugal
CityLisbon
Period4/15/24 → …

Keywords

  • ansible
  • code generation
  • domain specific languages
  • large language models

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Safety, Risk, Reliability and Quality

Fingerprint

Dive into the research topics of 'An Empirical Comparison of Code Generation Approaches for Ansible'. Together they form a unique fingerprint.

Cite this