MergeNet: Knowledge Migration across Heterogeneous Models, Tasks, and Modalities
arxiv(2024)
摘要
In this study, we focus on heterogeneous knowledge transfer across entirely
different model architectures, tasks, and modalities. Existing knowledge
transfer methods (e.g., backbone sharing, knowledge distillation) often hinge
on shared elements within model structures or task-specific features/labels,
limiting transfers to complex model types or tasks. To overcome these
challenges, we present MergeNet, which learns to bridge the gap of parameter
spaces of heterogeneous models, facilitating the direct interaction,
extraction, and application of knowledge within these parameter spaces. The
core mechanism of MergeNet lies in the parameter adapter, which operates by
querying the source model's low-rank parameters and adeptly learning to
identify and map parameters into the target model. MergeNet is learned
alongside both models, allowing our framework to dynamically transfer and adapt
knowledge relevant to the current stage, including the training trajectory
knowledge of the source model. Extensive experiments on heterogeneous knowledge
transfer demonstrate significant improvements in challenging settings, where
representative approaches may falter or prove less applicable.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要