本文首次提出MoE-Adapters,一种参数高效的训练框架,旨在缓解视觉-语言模型(VLM)增量学习中的长期遗忘问题。我们的MoE-Adapters利用逐步添加的路由器来激活并整合来自预定义静态专家集的专属专家适配器,使预训练的CLIP能够高效适应新任务。为了保持VLM的零样本能力,引入了分布判别自动选择器(DDAS),它能自动将分布内和分布外输入分别路由至MoE-Adapters和原始CLIP。然而,依赖静态专家集和独立分布选择器可能导致参数冗余和训练复杂度增加。对此,我们进一步扩展了MoE-Adapters++框架,引入动态MoE适配器,使专家能在持续学习过程中自适应参与。此外,提出了潜在嵌入自动选择器(LEAS),该选择器融合分布选择于CLIP内部,构建更统一的架构。大量多样化设置的实验表明,该方法在持续提升训练效率的同时,始终超过了以往最先进方法的性能。
Building similarity graph...
Analyzing shared references across papers
Loading...
Jiazuo Yu
Zichen Huang
Yunzhi Zhuge
IEEE Transactions on Pattern Analysis and Machine Intelligence
Tsinghua University
Dalian University of Technology
University of Electronic Science and Technology of China
Building similarity graph...
Analyzing shared references across papers
Loading...
Yu等人(周三,)研究了这一问题。
www.synapsesocial.com/papers/68a3633d0a429f7973329f0c — DOI: https://doi.org/10.1109/tpami.2025.3597942
Synapse has enriched 4 closely related papers on similar clinical questions. Consider them for comparative context: