logo
today local_bar
Poster Session 5 · Friday, December 5, 2025 11:00 AM → 2:00 PM
#2613

CrossSpectra: Exploiting Cross-Layer Smoothness for Parameter-Efficient Fine-Tuning

NeurIPS Poster OpenReview

Abstract

Parameter-efficient fine-tuning (PEFT) is essential for adapting large foundation models without excessive storage cost. However, current approaches such as LoRA treat each layer’s adaptation independently, overlooking correlations across layers. This independence causes the number of trainable parameters to grow linearly with model depth.
We provide theoretical and empirical evidence that skip connections in transformers create smooth gradient propagation across layers. This smoothness leads to weight adaptations that concentrate most of their energy in low-frequency spectral components, especially along the layer dimension. Empirical analysis confirms this effect, showing that most of adaptation energy lies in low frequencies.
Building on this insight, we propose CrossSpectra, which parameterizes all attention-weight adaptations across layers as a single 3D tensor and represents them with sparse spectral coefficients (). Using non-zero coefficients within each layer’s frequency space and truncating to frequencies across layers, CrossSpectra requires parameters instead of LoRA’s , where is the number of layers and the rank.
Across natural-language and vision benchmarks, methodname matches or surpasses baseline performance while using fewer parameters than LoRA, achieving only of LoRA’s parameter count when fine-tuning LLaMA-7B on instruction-following tasks. These results show that exploiting the architectural smoothness of transformers through spectral analysis yields major efficiency gains in PEFT.
Poster