DiaBlo: Diagonal Block Fine-Tuning Matches Full Model Performance With Lower Cost
Researchers introduce DiaBlo, a parameter-efficient fine-tuning method that updates only diagonal blocks of model weight matrices instead of full parameters. The approach matches full-model fine-tuning performance across reasoning, code generation, and safety tasks while maintaining comparable memory usage and training speed to LoRA.
DiaBlo: Diagonal Block Fine-Tuning Matches Full Model Performance With Lower Cost
A new parameter-efficient fine-tuning (PEFT) method eliminates the performance gap between efficient adaptation and full-model fine-tuning by selectively updating diagonal blocks in neural network weight matrices.
The technique, called DiaBlo (Diagonal Blocks), updates only the diagonal blocks of selected weight matrices rather than training full parameters or using low-rank approximations. This design avoids the auxiliary initialization schemes and customized optimizers required by LoRA and similar methods.
How DiaBlo Works
DiaBlo selects specific weight matrices in an LLM and partitions them into diagonal blocks. During fine-tuning, only these block-diagonal portions update, leaving the rest frozen. The approach requires no low-rank matrix products, simplifying the training process compared to LoRA variants that decompose weight updates into rank-constrained factors.
According to the researchers, this architectural choice enables more stable convergence. The method maintains LoRA-comparable memory usage and training speed while achieving stronger end-task performance.
Theoretical and Empirical Validation
The paper provides theoretical analysis showing that under mild low-rank conditions, DiaBlo is more expressive than LoRA for linear problems and converges to stationary points in general nonlinear optimization.
Experiments span multiple task categories:
- Commonsense reasoning: Performance competitive with full fine-tuning
- Arithmetic reasoning: Maintains accuracy across numerical tasks
- Code generation: Matches or exceeds LoRA baselines
- Safety alignment: Consistent improvements across alignment datasets
No specific benchmark scores or model sizes are disclosed in the abstract. The method preserves high memory efficiency—a critical requirement for production fine-tuning—while eliminating the need for task-specific hyperparameter tuning that LoRA variants often require.
What This Means
DiaBlo addresses a real constraint in LLM adaptation: existing PEFT methods trade performance for efficiency, requiring careful tuning to close the gap with full fine-tuning. By simplifying the update structure—using dense diagonal blocks instead of low-rank factors—DiaBlo reduces engineering complexity while maintaining practical efficiency.
The approach is particularly relevant for organizations fine-tuning on proprietary datasets where closing the full-fine-tuning performance gap matters. The code is available on GitHub, enabling immediate adoption and comparison against LoRA-based approaches in production pipelines.
The key practical question: whether diagonal block updates generalize as well across the diversity of domain-specific fine-tuning tasks remains to be validated at scale, though the initial results across reasoning, code, and safety domains are promising.