Model merging and safety alignment: one bad model spoils the bunch
Merging Large Language Models (LLMs) is a cost-effective technique for combining multiple expert LLMs into a single versatile model, retaining the expertise of the original ones. However, current approaches often overlook the importance of safety alignment during merging, leading to highly misaligne...
Main Authors: | Hammoud, HAAK, Michieli, U, Pizzati, F, Torr, P, Bibi, A, Ghanem, B, Ozay, M |
---|---|
Format: | Conference item |
Language: | English |
Published: |
Association for Computational Linguistics
2024
|
Similar Items
-
Bi-factorial preference optimization: balancing safety-helpfulness in language models
by: Zhang, W, et al.
Published: (2025) -
Merging of gold and semiconductor at nanoscale
by: Kamysbayev Vladislav
Published: (2014) -
Turbulent mixing of merging plumes
by: Lee, Selina Chien Cheah.
Published: (2008) -
Merge : the hybridization of architecture, infrastructure, and landscape
by: Fausto, Ariel
Published: (2007) -
Atomic Habits : Tiny Changes, Remarkable Results : An Easy & Proven Way to Build Good Habits & Break Bad Ones /
by: Clear, James, author 656048
Published: (2018)