x
InputW₀
Frozen
175B Parameters
(Do Not Touch)
(Do Not Touch)
=
h
OutputInference Control
Total Params:175,000,000,000
Trainable Params:0
VRAM Usage:320 GB
> The model explains the concept of recursion._
Founder Note
Notice how the massive W₀ block never changes. LoRA injects trainable rank decomposition matrices into each layer of the Transformer architecture, allowing you to "steer" the model's behavior without retraining the original 175 billion parameters.