Loading Now

Summary of Grounding Data Science Code Generation with Input-output Specifications, by Yeming Wen et al.


Grounding Data Science Code Generation with Input-Output Specifications

by Yeming Wen, Pengcheng Yin, Kensen Shi, Henryk Michalewski, Swarat Chaudhuri, Alex Polozov

First submitted to arxiv on: 12 Feb 2024

Categories

  • Main: Machine Learning (cs.LG)
  • Secondary: Programming Languages (cs.PL); Software Engineering (cs.SE)

     Abstract of paper      PDF of paper


GrooveSquid.com Paper Summaries

GrooveSquid.com’s goal is to make artificial intelligence research accessible by summarizing AI papers in simpler terms. Each summary below covers the same AI paper, written at different levels of difficulty. The medium difficulty and low difficulty versions are original summaries written by GrooveSquid.com, while the high difficulty version is the paper’s original abstract. Feel free to learn from the version that suits you best!

Summary difficulty Written by Summary
High Paper authors High Difficulty Summary
Read the original abstract here
Medium GrooveSquid.com (original content) Medium Difficulty Summary
This paper addresses a crucial challenge in large language model (LLM) programming: aligning outputs with both natural language prompts and input-output specifications. Current LLMs struggle to generate accurate code when given ambiguous NL prompts, requiring additional I/O specifications. The authors propose GIFT4Code, an instruction fine-tuning approach that utilizes synthetic data produced by the LLM and execution-derived feedback in the form of program I/O specifications. This method facilitates learning signals for the LLM, significantly improving its ability to generate executable code aligned with user specifications.
Low GrooveSquid.com (original content) Low Difficulty Summary
Large language models can generate code from natural language prompts, but they struggle when prompts are ambiguous. To fix this, researchers have developed a new way to fine-tune these models using data produced by the model itself and feedback from how the code works. This method helps the model learn what it means to follow instructions correctly. The authors tested their approach on two challenging data science tasks and found that it greatly improved the quality of generated code.

Keywords

* Artificial intelligence  * Fine tuning  * Large language model  * Synthetic data