TriSum: Learning Summarization Ability from Large Language Models with Structured Rationale
arxiv(2024)
摘要
The advent of large language models (LLMs) has significantly advanced natural
language processing tasks like text summarization. However, their large size
and computational demands, coupled with privacy concerns in data transmission,
limit their use in resource-constrained and privacy-centric settings. To
overcome this, we introduce TriSum, a framework for distilling LLMs' text
summarization abilities into a compact, local model. Initially, LLMs extract a
set of aspect-triple rationales and summaries, which are refined using a
dual-scoring method for quality. Next, a smaller local model is trained with
these tasks, employing a curriculum learning strategy that evolves from simple
to complex tasks. Our method enhances local model performance on various
benchmarks (CNN/DailyMail, XSum, and ClinicalTrial), outperforming baselines by
4.5
providing insights into the summarization rationale.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要