Imperfect Code Generation: Uncovering Weaknesses in Automatic Code Generation by Large Language Models.

International Conference on Software Engineering(2024)

引用 0|浏览19
暂无评分
摘要
The task of code generation has received significant attention in recent years, especially when the pre-trained large language models (LLMs) for code have consistently achieved state-of-the-art performance. However, there is currently a lack of a comprehensive weakness taxonomy in the field, uncovering weaknesses in automatic code generation by LLMs. This may lead the community to invest excessive efforts into well-known hotspots while neglecting many crucial yet unrecognized issues that deserve more attention. To bridge this gap, we conduct a systematic study on analyzing the weaknesses based on three state-of-the-art LLMs across three widely-used code generation datasets. Our study identifies eight types of weaknesses and assesses their prevalence across each LLM and dataset, aiming to inform and shape the trajectory of future research in the domain.
更多
查看译文
关键词
Automatic Generation,Code Generation,Large Language Models,Automatic Code Generation,Research Domain,Pre-trained Language Models,Weak Type,Comprehensive Taxonomy,Benchmark,Semantic,False Negative,Source Code,Gold-plated,Code Snippets,Clone Detection
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要