TII Researchers Introduce Learnable Multipliers for LLM Scaling
- •Researchers from the Technology Innovation Institute introduced learnable multipliers to optimize weight scaling in large language models.
- •The method replaces fixed parameterization with dynamic scalar scaling for matrices, rows, and columns during pretraining.
- •This approach improves model performance and efficiency, rivaling the gains seen in advanced optimizers like Muon.
Researchers at the Technology Innovation Institute (TII), a leading global research center, have developed a novel method called learnable multipliers to address a fundamental limitation in large language model training. Current training methods rely heavily on weight decay to stabilize weight norms, but this often results in a suboptimal equilibrium caused by stochastic gradient noise. This phenomenon, described as a Brownian-like expansion, restricts the expressive capacity of model layers by forcing them into rigid architectural constraints.
To bypass these limitations, the TII team introduced learnable scalar scaling at the matrix, row, and column levels, allowing the model to determine its own optimal weight scales dynamically. This technique serves as a more expressive generalization of Maximal Update Parametrization (muP), which typically requires extensive manual hyperparameter tuning. By enabling the model to adjust its scale during the pretraining phase, the researchers effectively free the weights from the negative artifacts of traditional regularization.
Empirical evaluations demonstrate that learnable multipliers yield significant performance improvements across various downstream benchmarks while reducing the computational overhead associated with hyperparameter selection. Interestingly, tests conducted with the standard Adam optimizer showed performance gains comparable to those achieved by switching to the more advanced Muon optimizer. This suggests that proper weight scaling is as critical to model optimization as the choice of the optimization algorithm itself.