PhD student, Seoul National University
1 paper at NeurIPS 2025
FALQON accelerates LoRA fine-tuning by up to 3$\times$ through merging adapters into an FP8-quantized backbone, removing redundant quantization overhead from small matrices.