Artwork

Content provided by Nicolay Gerold. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Nicolay Gerold or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!

#054 Building Frankenstein Models with Model Merging and the Future of AI

1:06:55
 
Share
 

Manage episode 497224954 series 3585930
Content provided by Nicolay Gerold. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Nicolay Gerold or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.

Nicolay here,most AI conversations focus on training bigger models with more compute. This one explores the counterintuitive world where averaging weights from different models creates better performance than expensive post-training.

Today I have the chance to talk to Maxime Labonne, who's a researcher at Liquid AI and the architect of some of the most popular open source models on Hugging Face.

He went from researching neural networks for cybersecurity to building "Frankenstein models" through techniques that shouldn't work but consistently do.

Key Insight: Model Merging as a Free LunchThe core breakthrough is deceptively simple: take two fine-tuned models, average their weights layer by layer, and often get better performance than either individual model. Maxime initially started writing an article to explain why this couldn't work, but his own experiments convinced him otherwise.

The magic lies in knowledge compression and regularization. When you train a model multiple times on similar data, each run creates slightly different weight configurations due to training noise. Averaging these weights creates a smoother optimization path that avoids local minima. You can literally run model merging on a CPU - no GPUs required.

In the podcast, we also touch on:

  • Obliteration: removing safety refusal mechanisms without retraining
  • Why synthetic data now comprises 90%+ of fine-tuning datasets
  • The evaluation crisis and automated benchmarks missing real-world performance
  • Chain of thought compression techniques for reasoning models

💡 Core Concepts

  • Model Merging: Averaging weights across layers from multiple fine-tuned models to create improved performance without additional training
  • Obliteration: Training-free method to remove refusal directions from models by computing activation differences
  • Linear Merging: The least opinionated merging technique that simply averages weights with optional scaling factors
  • Refusal Direction: The activation pattern that indicates when a model will output a safety refusal

📶 Connect with Maxime:

📶 Connect with Nicolay:

Important Moments

  • Model Merging Discovery Process: [00:00:30] Maxime explains how he started writing an article to debunk model merging
  • Two Main Merging Use Cases: [11:04] Clear distinction between merging checkpoints versus combining different task-specific capabilities
  • Linear Merging as Best Practice: [21:00] Why simple weight averaging consistently outperforms more complex techniques
  • Layer Importance Hierarchy: [21:18] First and last layers have the most influence on model behavior
  • Obliteration Technique Explained: [36:07] How to compute and subtract refusal directions from model activations
  • Synthetic Data Dominance: [50:00] Modern fine-tuning uses 90%+ synthetic data

🛠 Tools & Tech Mentioned

📚 Recommended Resources

  continue reading

63 episodes

Artwork
iconShare
 
Manage episode 497224954 series 3585930
Content provided by Nicolay Gerold. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Nicolay Gerold or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.

Nicolay here,most AI conversations focus on training bigger models with more compute. This one explores the counterintuitive world where averaging weights from different models creates better performance than expensive post-training.

Today I have the chance to talk to Maxime Labonne, who's a researcher at Liquid AI and the architect of some of the most popular open source models on Hugging Face.

He went from researching neural networks for cybersecurity to building "Frankenstein models" through techniques that shouldn't work but consistently do.

Key Insight: Model Merging as a Free LunchThe core breakthrough is deceptively simple: take two fine-tuned models, average their weights layer by layer, and often get better performance than either individual model. Maxime initially started writing an article to explain why this couldn't work, but his own experiments convinced him otherwise.

The magic lies in knowledge compression and regularization. When you train a model multiple times on similar data, each run creates slightly different weight configurations due to training noise. Averaging these weights creates a smoother optimization path that avoids local minima. You can literally run model merging on a CPU - no GPUs required.

In the podcast, we also touch on:

  • Obliteration: removing safety refusal mechanisms without retraining
  • Why synthetic data now comprises 90%+ of fine-tuning datasets
  • The evaluation crisis and automated benchmarks missing real-world performance
  • Chain of thought compression techniques for reasoning models

💡 Core Concepts

  • Model Merging: Averaging weights across layers from multiple fine-tuned models to create improved performance without additional training
  • Obliteration: Training-free method to remove refusal directions from models by computing activation differences
  • Linear Merging: The least opinionated merging technique that simply averages weights with optional scaling factors
  • Refusal Direction: The activation pattern that indicates when a model will output a safety refusal

📶 Connect with Maxime:

📶 Connect with Nicolay:

Important Moments

  • Model Merging Discovery Process: [00:00:30] Maxime explains how he started writing an article to debunk model merging
  • Two Main Merging Use Cases: [11:04] Clear distinction between merging checkpoints versus combining different task-specific capabilities
  • Linear Merging as Best Practice: [21:00] Why simple weight averaging consistently outperforms more complex techniques
  • Layer Importance Hierarchy: [21:18] First and last layers have the most influence on model behavior
  • Obliteration Technique Explained: [36:07] How to compute and subtract refusal directions from model activations
  • Synthetic Data Dominance: [50:00] Modern fine-tuning uses 90%+ synthetic data

🛠 Tools & Tech Mentioned

📚 Recommended Resources

  continue reading

63 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Quick Reference Guide

Copyright 2025 | Privacy Policy | Terms of Service | | Copyright
Listen to this show while you explore
Play