Case Study: Differentiable Adaptive Merging (DAM)

Ticket: GH-446 Module: aprender::online::dam

Overview

DAM learns per-model merge coefficients that minimize reconstruction loss on a calibration set. Unlike fixed-weight merging, DAM adapts weights per-tensor using Nelder-Mead simplex optimization.

Key Components

  • DamConfig — Learning rate, iterations, regularization, seed
  • DamLoss — MSE loss, L2 regularization, gradient step
  • softmax — Numerically stable logit-to-probability conversion
  • optimize_coefficients — Nelder-Mead simplex optimizer
  • DamReport — Final loss, convergence status, coefficients

Run

cargo run --example dam_merge

Falsification Tests

IDPropertyStatus
FALSIFY-DAM-001Softmax sums to 1Falsified (holds)
FALSIFY-DAM-002Optimization reduces lossFalsified (holds)
FALSIFY-DAM-003MSE loss is non-negativeFalsified (holds)