Skip to content

Neural Alignment

Prompt Browser

Note that both compressed models studied here, MKA and MKA with neural alignment, merge 11 layers, which corresponds to a compression ratio of 34.375 percent. Use the controls below to explore how the different models responded to each prompt.

Prompt



  

Ground truth



  

Model output


Neural alignment visualization

Accuracy vs. Compression Ratio (Neural Alignment Method)

The plot and table below compare our neural alignment merging strategy with the original MKA heuristic across different compression ratios. While MKA shows strong accuracy at low compression (approximately 3 percent), it suffers from a large performance collapse at 12.5 percent compression, dropping to 0.547. In contrast, our alignment method remains much more stable, achieving 0.633 at the same compression level, which is an improvement of 0.086 over MKA.

As compression increases, both methods converge to similar performance (approximately 0.63 at 40 percent), but neural alignment avoids the severe volatility exhibited by MKA. This indicates that correctly aligning neurons before merging leads to a more reliable and predictable degradation curve, especially in the moderate compression regime where real world deployments often operate.

Accuracy vs compression ratio

Layers Compression (%) Ours (Accuracy) MKA (Accuracy)
1 3.125 0.6478 0.6620
4 12.500 0.6334 0.5470
11 34.375 0.6124 0.6487
13 40.625 0.6392 0.6342
15 46.875 0.2631 0.3000