Wed 17 Jul 2024 16:18 - 16:36 at Pitomba - AI4SE 2 Chair(s): Jingyue Li

Testing plays a pivotal role in ensuring software quality, yet conventional Search Based Software Testing (SBST) methods often struggle with complex software units, achieving suboptimal test coverage. Recent work using large language models (LLMs) for test generation have focused on improving generation quality through optimizing the test generation context and correcting errors in model outputs, but use fixed prompting strategies that prompt the model to generate tests without additional guidance. As a result LLM-generated testsuites still suffer from low coverage.

In this paper, we present SymPrompt, a code-aware prompting strategy for LLMs in test generation. SymPrompt’s approach is based on recent work that demonstrates LLMs can solve more complex logical problems when prompted to reason about the problem in a multi-step fashion. We apply this methodology to test generation by deconstructing the testsuite generation process into a multi-stage sequence, each of which is driven by a specific prompt aligned with the execution paths of the method under test, and exposing relevant type and dependency focal context to the model. Our approach enables pretrained LLMs to generate more complete test cases without any additional training. We implement SymPrompt using the TreeSitter parsing framework and evaluate on a benchmark challenging methods from open source Python projects. SymPrompt enhances correct test generations by a factor of 5 and bolsters relative coverage by 26% for CodeGen2. Notably, when applied to GPT-4, symbolic path prompts improve coverage by a factor of over $2\times$ compared to baseline prompting strategies.

Wed 17 Jul

Displayed time zone: Brasilia, Distrito Federal, Brazil change

16:00 - 18:00
AI4SE 2Industry Papers / Research Papers at Pitomba
Chair(s): Jingyue Li Norwegian University of Science and Technology (NTNU)
16:00
18m
Talk
MonitorAssistant: Simplifying Cloud Service Monitoring via Large Language Models
Industry Papers
Zhaoyang Yu Tsinghua University, Minghua Ma Microsoft Research, Chaoyun Zhang Microsoft, Si Qin Microsoft Research, Yu Kang Microsoft Research, Chetan Bansal Microsoft Research, Saravan Rajmohan Microsoft, Yingnong Dang Microsoft Azure, Changhua Pei Computer Network Information Center at Chinese Academy of Sciences, Dan Pei Tsinghua University, Qingwei Lin Microsoft, Dongmei Zhang Microsoft Research
16:18
18m
Talk
Code-Aware Prompting: A study of Coverage guided Test Generation in Regression Setting using LLM
Research Papers
Gabriel Ryan Columbia University, Siddhartha Jain AWS AI Labs, Mingyue Shang AWS AI Labs, Shiqi Wang AWS AI Labs, Xiaofei Ma AWS AI Labs, Murali Krishna Ramanathan AWS AI Labs, Baishakhi Ray Columbia University, New York; AWS AI Lab
16:36
18m
Talk
A Machine Learning-Based Error Mitigation Approach for Reliable Software Development on IBM’s Quantum Computers
Industry Papers
Asmar Muqeet Simula Research Laboratory and University of Oslo, Shaukat Ali Simula Research Laboratory and Oslo Metropolitan University, Tao Yue Beihang University, Paolo Arcaini National Institute of Informatics
16:54
18m
Talk
Multi-line AI-assisted Code Authoring
Industry Papers
Omer Dunay Meta Platforms, Inc., Daniel Cheng Meta Platforms Inc., Adam Tait Meta Platforms, Inc., Parth Thakkar Meta Platforms, Inc., Peter C Rigby Meta / Concordia University, Andy Chiu Meta Platforms, Inc., Imad Ahmad Meta Platforms, Inc., Arun Ganesan Meta Platforms, Inc., Chandra Sekhar Maddila Meta Platforms, Inc., Vijayaraghavan Murali Meta Platforms Inc., Ali Tayyebi Meta Platforms Inc., Nachiappan Nagappan Meta Platforms, Inc.
17:12
18m
Talk
Combating Missed Recalls in E-commerce Search: a CoT-prompting Testing Approach
Industry Papers
Shengnan Wu School of Computer Science, Fudan University, Yongxiang Hu Fudan University, Yingchuan Wang School of Computer Science, Fudan University, Jiazhen Gu The Chinese University of Hong Kong, Jin Meng Meituan Inc., Liujie Fan Meituan Inc., Zhongshi Luan Meituan Inc., Xin Wang Fudan University, Yangfan Zhou Fudan University
Pre-print
17:30
18m
Talk
Automated Unit Test Improvement using Large Language Models at Meta
Industry Papers
Mark Harman Meta Platforms, Inc. and UCL, Jubin Chheda Meta platforms, Anastasia Finogenova Meta platforms, Inna Harper Meta, Alexandru Marginean Meta platforms, Shubho Sengupta Meta platforms, Eddy Wang Meta platforms, Nadia Alshahwan Meta Platforms, Beliz Gokkaya Meta Platforms