research

DiaBlo: Diagonal Block Finetuning Matches Full Model Performance With Lower Cost

Researchers propose DiaBlo, a parameter-efficient finetuning (PEFT) method that updates only diagonal blocks of model weight matrices, achieving comparable performance to full-model finetuning while maintaining LoRA-level efficiency. The approach eliminates low-rank matrix dependencies and provides theoretical guarantees of convergence.

2 min read

DiaBlo: Diagonal Block Finetuning Matches Full Model Performance With Lower Cost

A new parameter-efficient finetuning (PEFT) method called DiaBlo addresses a persistent gap in LLM adaptation: achieving full-model finetuning performance without the computational overhead.

The core innovation is simple: update only the diagonal blocks of selected weight matrices during finetuning, rather than adapting the entire model or using low-rank approximations. This eliminates dependencies on auxiliary initialization schemes and custom optimization strategies that plague existing low-rank methods like LoRA.

Key Technical Differences From LoRA

LoRA and variants achieve efficiency by injecting low-rank matrices (A×B products) into model layers. This requires careful initialization tuning and often needs specialized optimization approaches to converge reliably.

DiaBlo instead directly optimizes diagonal block submatrices—a structurally simpler approach that avoids low-rank factorization entirely. The method selects which weight matrices to modify (similar to LoRA), then updates only their diagonal block components.

Performance and Efficiency Claims

According to the research, DiaBlo delivers:

  • Competitive accuracy across commonsense reasoning, arithmetic reasoning, code generation, and safety alignment tasks compared to full finetuning
  • Comparable memory usage to LoRA during training
  • Similar training speed to existing PEFT methods
  • Stable convergence without requiring initialization tricks or custom optimization

The researchers provide theoretical guarantees showing that under mild low-rank conditions, DiaBlo is more expressive than LoRA for linear problems and converges to stationary points in general nonlinear finetuning scenarios.

Why This Matters

The PEFT problem has a fundamental tradeoff: methods that preserve performance (like full finetuning) are expensive; methods that are cheap (like LoRA) sacrifice accuracy. DiaBlo appears to push the frontier toward the Pareto frontier by combining structural simplicity with strong empirical results.

The elimination of low-rank auxiliary matrices also reduces implementation complexity—fewer hyperparameters to tune, fewer initialization strategies to choose between, and potentially easier integration into existing training pipelines.

Limitations and Context

The paper does not specify:

  • Which LLM sizes were tested (parameter counts)
  • Exact accuracy margins vs. full finetuning across benchmarks
  • Whether diagonal block selection is task-dependent or universal
  • Computational overhead of selecting which blocks to update

The code is available at https://github.com/ziyangjoy/DiaBlo, allowing independent verification of claims.

What This Means

If DiaBlo's theoretical guarantees and empirical results hold across diverse models and domains, it could become the default choice for efficient LLM adaptation. The method's simplicity—no low-rank products, no custom initialization—makes it likely to be more robust and easier to adopt than LoRA variants. However, the gains appear incremental rather than transformative; the real value is in reducing implementation complexity while maintaining efficiency-performance balance.