|
|
Zapiski Nauchnykh Seminarov POMI, 2024, Volume 540, Pages 27–45
(Mi znsl7542)
|
|
|
|
Refining joint text and source code embeddings for retrieval task with parameter-efficient fine-tuning
K. Galliamov, L. Khaertdinova, K. Denisova Innopolis University, Innopolis, Russia
Abstract:
Latest developments in natural language processing demonstrate remarkable progress in the code-text retrieval problem. As Transformer-based models used for this task continue to increase in size, the computational costs and time required for end-to-end fine-tuning become substantial. This poses a significant challenge for adapting and utilizing these models when computational resources are limited. Motivated by these concerns, we propose a fine-tuning framework that leverages parameter-efficient fine-tuning (PEFT) techniques. Moreover, we adopt contrastive learning objectives to improve the quality of bimodal representations learned by Transformer-based models. Additionally, for PEFT methods we provide extensive benchmarking, the lack of which has been highlighted as a crucial problem in the literature. Based on extensive experiments with the CodeT5+ model conducted on two datasets, we demonstrate that the proposed fine-tuning framework has the potential to improve code-text retrieval performance by tuning only 0.4% parameters at the most.
Key words and phrases:
Code retrieval, PEFT, CodeT5+, contrastive learning, NLP.
Received: 15.11.2024
Citation:
K. Galliamov, L. Khaertdinova, K. Denisova, “Refining joint text and source code embeddings for retrieval task with parameter-efficient fine-tuning”, Investigations on applied mathematics and informatics. Part IV, Zap. Nauchn. Sem. POMI, 540, POMI, St. Petersburg, 2024, 27–45
Linking options:
https://www.mathnet.ru/eng/znsl7542 https://www.mathnet.ru/eng/znsl/v540/p27
|
| Statistics & downloads: |
| Abstract page: | 97 | | Full-text PDF : | 120 | | References: | 22 |
|