Research¶
Model Garage is validated through three peer-reviewed research papers demonstrating its capabilities on real problems. Each paper uses Model Garage modules directly and reports quantified, reproducible results.
Papers¶
-
Blades: Compositional Capability Enhancement Through Hidden State Injection
Hidden state injection between specialized models achieves +14.2% accuracy on medical reasoning tasks. Establishes 7 validated principles for capability transfer including the N-4 layer rule and same-domain synergy (+27.8%).
Modules used:
extract,inject,snapshot,core.hooks -
Learned Routers Don't Learn: Expert Miscalibration in MoE Models
Per-layer expert isolation reveals that learned MoE routers show Spearman rho = 0.069 between routing probability and expert quality. 482/896 expert-layer-domain combinations show statistically significant specialization, yet the router ignores it.
Modules used:
analyze,core.hooks,registry -
Sparse Pathways: Domain-Aware Neuron Routing for Efficient Inference
FFN neurons exhibit strong domain specialization (~50% in late layers), with r=0.999 correlation between model scale and specialization degree. Demonstrates 2-4x potential compute reduction via negative neuron selection.
Modules used:
analyze,snapshot,core.hooks
Citation¶
If you use these findings or Model Garage in your research:
@software{model_garage,
title = {Model Garage: Component-Level Neural Network Surgery},
author = {Model Garage Contributors},
year = {2026},
url = {https://github.com/Lumi-node/model-garage},
license = {Apache-2.0}
}
Pretrained Artifacts¶
The research directory includes pretrained components ready for use:
- Extracted Components — Pre-extracted attention heads and FFN modules
- Pretrained Blades — SAE-trained blades for capability injection
- Blade Principles — Validated rules for capability transfer