XPrompt:Explaining Large Language Model's Generation via Joint Prompt Attribution
CoRR(2024)
摘要
Large Language Models (LLMs) have demonstrated impressive performances in
complex text generation tasks. However, the contribution of the input prompt to
the generated content still remains obscure to humans, underscoring the
necessity of elucidating and explaining the causality between input and output
pairs. Existing works for providing prompt-specific explanation often confine
model output to be classification or next-word prediction. Few initial attempts
aiming to explain the entire language generation often treat input prompt texts
independently, ignoring their combinatorial effects on the follow-up
generation. In this study, we introduce a counterfactual explanation framework
based on joint prompt attribution, XPrompt, which aims to explain how a few
prompt texts collaboratively influences the LLM's complete generation.
Particularly, we formulate the task of prompt attribution for generation
interpretation as a combinatorial optimization problem, and introduce a
probabilistic algorithm to search for the casual input combination in the
discrete space. We define and utilize multiple metrics to evaluate the produced
explanations, demonstrating both faithfulness and efficiency of our framework.
更多查看译文
AI 理解论文
溯源树
样例
![](https://originalfileserver.aminer.cn/sys/aminer/pubs/mrt_preview.jpeg)
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要