If you have any questions about the library, please feel free to contact us. Email: [email protected]
A comprehensive list of papers about '[Knowledge Fusion: A Comprehensive Survey.]'.
As the comprehensive capabilities of foundational large models rapidly improve, similar general abilities have emerged across different models, making capability transfer and fusion between them more feasible. Knowledge fusion aims to integrate existing LLMs of diverse architectures and capabilities into a more powerful model through efficient methods such as knowledge distillation, model merging, mixture of experts, and PEFT, thereby reducing the need for costly LLM development and adaptation. We provide a comprehensive overview of model merging methods and theories, covering their applications across various fields and scenarios, including LLMs, MLLMs, image generation, model compression, continual learning, and more. Finally, we highlight the challenges of knowledge fusion and explore future research directions.
Paper Title | Code | Publication & Date |
---|---|---|
Soft merging of experts with adaptive routing | TMLR 2024 | |
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models | DeepSeekMoE | ArXiv 24.01 |
Multiple Expert Brainstorming for Domain Adaptive Person Re-identification | MEB-Net | ECCV 2020 |
Merging Vision Transformers from Different Tasks and Domains | ArXiv 23.12 |
Paper Title | Code | Publication & Date |
---|---|---|
Model Reprogramming: Resource-Efficient Cross-Domain Machine Learning | AAAI 2024 | |
Towards Efficient Task-Driven Model Reprogramming with Foundation Models | ArXiv 23.06 | |
Deep Graph Reprogramming | ycjing | CVPR 2023 |
From English to More Languages: Parameter-Efficient Model Reprogramming for Cross-Lingual Speech Recognition | ICASSP 2023 | |
Fairness Reprogramming | USBC-NLP | NeurIPS 2022 |
Voice2Series: Reprogramming Acoustic Models for Time Series Classification | V2S | ICML 2021 |
Paper Title | Code | Publication & Date |
---|---|---|
EMR-Merging: Tuning-Free High-Performance Model Merging | EMR_Merging | NeurIPS 2024 spolight |
Model Composition for Multimodal Large Language Models | THUNLP | ACL 2024 |
Localizing Task Information for Improved Model Merging and Compression | tall_masks | ICML 2024 |
Adapting a Single Network to Multiple Tasks by Learning to Mask Weights | Piggyback | ECCV 2018 |
Paper Title | Code | Publication & Date |
---|---|---|
Amalgamating Knowledge From Heterogeneous Graph Neural Networks | ycjing | CVPR 2021 |
Paper Title | Code | Publication & Date |
---|---|---|
Advances in Robust Federated Learning: Heterogeneity Considerations | ArXiv 24.05 | |
Towards Personalized Federated Learning via Heterogeneous Model Reassembly | pFedHR | NeurIPS 2023 |
Stitchable Neural Networks | snnet | CVPR 2023 Highlight |
Instant Soup Cheap Pruning Ensembles in A Single Pass Can Draw Lottery Tickets from Large Models | instant_soup | ICML 2023 |
Deep Incubation: Training Large Models by Divide-and-Conquering | Deep-Incubation | ArXiv 22.12 |
Deep Model Reassembly | DeRy | NeurIPS 2022 |
GAN Cocktail: Mixing GANs without Dataset Access | GAN-cocktail | ECCV 2022 |
Paper Title | Code | Publication & Date |
---|---|---|
It’s Morphing Time: Unleashing the Potential of Multiple LLMs via Multi-objective Optimization | - | ArXiv 24.07 |
Evolutionary Optimization of Model Merging Recipes | EvoLLM | ArXiv 24.03 |
Knowledge Fusion By Evolving Weights of Language Models | Model_Evolver | ACL 2024 |
Population-based evolutionary gaming for unsupervised person re-identification | - | IJCV 2023 |
Paper Title | Code | Publication & Date |
---|---|---|
Evaluating the External and Parametric Knowledge Fusion of Large Language Models | ArXiv 24.05 | |
Knowledge Fusion and Semantic Knowledge Ranking for Open Domain Question Answering | ArXiv 20.04 |
Paper Title | Code | Publication & Date |
---|---|---|
You Only Merge Once: Learning the Pareto Set of Preference-Aware Model Merging | ArXiv 24.08 | |
Towards Efficient Pareto Set Approximation via Mixture of Experts Based Model Fusion | ArXiv 24.06 | |
MAP: Low-compute Model Merging with Amortized Pareto Fronts via Quadratic Approximation | ArXiv 24.06 |
Paper Title | Code | Publication & Date |
---|---|---|
Adaptive Discovering and Merging for Incremental Novel Class Discovery | AAAI 2024 | |
Knowledge Fusion and Semantic Knowledge Ranking for Open Domain Question Answering | ArXiv 20.04 |
Paper Title | Code | Publication & Date |
---|---|---|
A Survey on Model MoErging: Recycling and Routing Among Specialized Experts for Collaborative Learning | ArXiv 24.08 | |
Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities | Yang | ArXiv 24.08 |
Merge, Ensemble, and Cooperate! A Survey on Collaborative Strategies in the Era of Large Language Models | ArXiv 24.07 | |
Arcee's MergeKit: A Toolkit for Merging Large Language Models | MergeKit | ArXiv 24.03 |
Learn From Model Beyond Fine-Tuning: A Survey | LFM | ArXiv 23.10 |
Deep Model Fusion: A Survey | ArXiv 23.09 | |
A curated paper list of Model Merging methods | ycjing | GitHub |
Junlin Lee; Qi Tang; Runhua Jiang.
Star History
We invite all researchers to contribute to this repository, 'Knowledge Fusion: The Integration of Model Capabilities'. If you have any questions about the library, please feel free to contact us.
Email: [email protected]