Tools for merging pretrained large language models.
-
Updated
Oct 5, 2024 - Python
Tools for merging pretrained large language models.
Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. arXiv:2408.07666.
Code release for Dataless Knowledge Fusion by Merging Weights of Language Models (https://openreview.net/forum?id=FCnohuR6AnM)
ZhiJian: A Unifying and Rapidly Deployable Toolbox for Pre-trained Model Reuse
FusionBench: A Comprehensive Benchmark of Deep Model Fusion
AdaMerging: Adaptive Model Merging for Multi-Task Learning. ICLR, 2024.
Official repository of "Localizing Task Information for Improved Model Merging and Compression" [ICML 2024]
Representation Surgery for Multi-Task Model Merging. ICML, 2024.
DELLA-Merging: Reducing Interference in Model Merging through Magnitude-Based Sampling
All-in-one UI for merged LLMs in Hugging Face
flow-merge is a powerful Python library that enables seamless merging of multiple transformer-based language models using the most popular merge methods such as model soups, SLERP, ties-MERGING or DARE.
Merge transformers without using like a bajillion GB of RAM
[ECCV 2024] MagMax: Leveraging Model Merging for Seamless Continual Learning (official repository)
An easy-to-use Python library for merging PyTorch models.
Mergecraft is a simple library to streamline model merging operations, with seamless integration with HuggingFace🤗
Advanced Transfer Learning project with the purpose of obtaining the best model by mixing three twice-fine-tuned models.
Add a description, image, and links to the model-merging topic page so that developers can more easily learn about it.
To associate your repository with the model-merging topic, visit your repo's landing page and select "manage topics."