File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: IaC-Eval: A Code Generation Benchmark for Cloud Infrastructure-as-Code Programs

TitleIaC-Eval: A Code Generation Benchmark for Cloud Infrastructure-as-Code Programs
Authors
Issue Date2024
Citation
Advances in Neural Information Processing Systems, 2024, v. 37 How to Cite?
AbstractInfrastructure-as-Code (IaC), an important component of cloud computing, allows the definition of cloud infrastructure in high-level programs. However, developing IaC programs is challenging, complicated by factors that include the burgeoning complexity of the cloud ecosystem (e.g., diversity of cloud services and workloads), and the relative scarcity of IaC-specific code examples and public repositories. While large language models (LLMs) have shown promise in general code generation and could potentially aid in IaC development, no benchmarks currently exist for evaluating their ability to generate IaC code. We present IaC-Eval, a first step in this research direction. IaC-Eval's dataset includes 458 human-curated scenarios covering a wide range of popular AWS services, at varying difficulty levels. Each scenario mainly comprises a natural language IaC problem description and an infrastructure intent specification. The former is fed as user input to the LLM, while the latter is a general notion used to verify if the generated IaC program conforms to the user's intent; by making explicit the problem's requirements that can encompass various cloud services, resources and internal infrastructure details. Our in-depth evaluation shows that contemporary LLMs perform poorly on IaC-Eval, with the top-performing model, GPT-4, obtaining a pass@1 accuracy of 19.36%. In contrast, it scores 86.6% on EvalPlus, a popular Python code generation benchmark, highlighting a need for advancements in this domain. We open-source the IaC-Eval dataset and evaluation framework at https://github.com/autoiac-project/iac-eval to enable future research on LLM-based IaC code generation.
Persistent Identifierhttp://hdl.handle.net/10722/362998
ISSN
2020 SCImago Journal Rankings: 1.399

 

DC FieldValueLanguage
dc.contributor.authorKon, Patrick Tser Jern-
dc.contributor.authorLiu, Jiachen-
dc.contributor.authorQiu, Yiming-
dc.contributor.authorFan, Weijun-
dc.contributor.authorLin, Ting He Lei-
dc.contributor.authorZhang, Haoran-
dc.contributor.authorPark, Owen M.-
dc.contributor.authorElengikal, George S.-
dc.contributor.authorKang, Yuxin-
dc.contributor.authorChen, Ang-
dc.contributor.authorChowdhury, Mosharaf-
dc.contributor.authorLee, Myungjin-
dc.contributor.authorWang, Xinyu-
dc.date.accessioned2025-10-10T07:43:58Z-
dc.date.available2025-10-10T07:43:58Z-
dc.date.issued2024-
dc.identifier.citationAdvances in Neural Information Processing Systems, 2024, v. 37-
dc.identifier.issn1049-5258-
dc.identifier.urihttp://hdl.handle.net/10722/362998-
dc.description.abstractInfrastructure-as-Code (IaC), an important component of cloud computing, allows the definition of cloud infrastructure in high-level programs. However, developing IaC programs is challenging, complicated by factors that include the burgeoning complexity of the cloud ecosystem (e.g., diversity of cloud services and workloads), and the relative scarcity of IaC-specific code examples and public repositories. While large language models (LLMs) have shown promise in general code generation and could potentially aid in IaC development, no benchmarks currently exist for evaluating their ability to generate IaC code. We present IaC-Eval, a first step in this research direction. IaC-Eval's dataset includes 458 human-curated scenarios covering a wide range of popular AWS services, at varying difficulty levels. Each scenario mainly comprises a natural language IaC problem description and an infrastructure intent specification. The former is fed as user input to the LLM, while the latter is a general notion used to verify if the generated IaC program conforms to the user's intent; by making explicit the problem's requirements that can encompass various cloud services, resources and internal infrastructure details. Our in-depth evaluation shows that contemporary LLMs perform poorly on IaC-Eval, with the top-performing model, GPT-4, obtaining a pass@1 accuracy of 19.36%. In contrast, it scores 86.6% on EvalPlus, a popular Python code generation benchmark, highlighting a need for advancements in this domain. We open-source the IaC-Eval dataset and evaluation framework at https://github.com/autoiac-project/iac-eval to enable future research on LLM-based IaC code generation.-
dc.languageeng-
dc.relation.ispartofAdvances in Neural Information Processing Systems-
dc.titleIaC-Eval: A Code Generation Benchmark for Cloud Infrastructure-as-Code Programs-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-105000522034-
dc.identifier.volume37-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats