Interactive-KBQA: Multi-Turn Interactions for Knowledge Base Question Answering with Large Language Models
CoRR(2024)
摘要
This study explores the realm of knowledge-base question answering (KBQA).
KBQA is considered a challenging task, particularly in parsing intricate
questions into executable logical forms. Traditional semantic parsing
(SP)-based methods require extensive data annotations, which result in
significant costs. Recently, the advent of few-shot in-context learning,
powered by large language models (LLMs), has showcased promising capabilities.
Yet, fully leveraging LLMs to parse questions into logical forms in
low-resource scenarios poses a substantial challenge. To tackle these hurdles,
we introduce Interactive-KBQA, a framework designed to generate logical forms
through direct interaction with knowledge bases (KBs). Within this framework,
we have developed three generic APIs for KB interaction. For each category of
complex question, we devised exemplars to guide LLMs through the reasoning
processes. Our method achieves competitive results on the WebQuestionsSP,
ComplexWebQuestions, KQA Pro, and MetaQA datasets with a minimal number of
examples (shots). Importantly, our approach supports manual intervention,
allowing for the iterative refinement of LLM outputs. By annotating a dataset
with step-wise reasoning processes, we showcase our model's adaptability and
highlight its potential for contributing significant enhancements to the field.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要