Tạp chí quốc tế khác

Parameter-efficient fine-tuning of small language models for code generation: a comparative study of Gemma, Qwen 2.5 and Llama 3.2

Tạp chí / Hội thảo: International Journal of Electrical and Computer Engineering Đơn vị: CNTT DOI / Link:

Tác giả

Van-Viet Nguyen ; Huu-Khanh Nguyen ; Duc-Quang Vu ; The-Vinh Nguyen

Tác giả liên hệ

Tóm tắt

Large language models (LLMs) have demonstrated impressive capabilities in code generation; however, their high computational demands, privacy limitations, and challenges in edge deployment restrict their practical use in domain-specific applications. This study explores the effectiveness of parameter efficient fine-tuning for small language models (SLMs) with fewer than 3 billion parameters. We adopt a hybrid approach that combines low-rank adaptation (LoRA) and 4-bit quantization (QLoRA) to reduce finetuning costs while preserving semantic consistency. Experiments on the CodeAlpaca-20k dataset reveal that SLMs fine-tuned with this method outperform larger baseline models, including Phi-3 Mini 4K base, in ROUGE-L. Notably, applying our approach to the LLaMA 3 3B and Qwen2. 5 3B models yielded performance improvements of 54% and 55%, respectively, over untuned counterparts. We evaluate …