Can Large Language Models Transform Natural Language Intent into Formal Method Postconditions?
arxiv(2023)
摘要
Informal natural language that describes code functionality, such as code
comments or function documentation, may contain substantial information about a
programs intent. However, there is typically no guarantee that a programs
implementation and natural language documentation are aligned. In the case of a
conflict, leveraging information in code-adjacent natural language has the
potential to enhance fault localization, debugging, and code trustworthiness.
In practice, however, this information is often underutilized due to the
inherent ambiguity of natural language which makes natural language intent
challenging to check programmatically. The emergent abilities of Large Language
Models (LLMs) have the potential to facilitate the translation of natural
language intent to programmatically checkable assertions. However, it is
unclear if LLMs can correctly translate informal natural language
specifications into formal specifications that match programmer intent.
Additionally, it is unclear if such translation could be useful in practice. In
this paper, we describe nl2postcond, the problem of leveraging LLMs for
transforming informal natural language to formal method postconditions,
expressed as program assertions. We introduce and validate metrics to measure
and compare different nl2postcond approaches, using the correctness and
discriminative power of generated postconditions. We then use qualitative and
quantitative methods to assess the quality of nl2postcond postconditions,
finding that they are generally correct and able to discriminate incorrect
code. Finally, we find that nl2postcond via LLMs has the potential to be
helpful in practice; nl2postcond generated postconditions were able to catch 64
real-world historical bugs from Defects4J.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要