ORIGINAL RESEARCH article
Front. Artif. Intell.
Sec. Natural Language Processing
This article is part of the Research TopicThe Use of Large Language Models to Automate, Enhance, and Streamline Text Analysis Processes. Large Language Models Used to Analyze and Check Requirement Compliance.View all 9 articles
An Efficient Strategy for Fine-Tuning Large Language Models
Provisionally accepted- 1Marine Corps Tactical Systems Support Activity, US Marine Corps, Arlington, United States
- 2Massachusetts Institute of Technology Lincoln Laboratory, Lexington, United States
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
ABSTRACT Large Language Models (LLMs) are powerful tools for Natural Language Processing tasks; fine-tuning them for domain-specific applications is resource-intensive, requiring significant effort to develop datasets and considerable compute resources for fine-tuning. This work establishes a combination of techniques that enable users to rapidly fine-tune LLMs for domain-specific tasks using limited datasets and performs a comparative analysis across established fine-tuning methods using this strategy to determine the most efficient fine-tuning method to use in conjunction with these techniques. This strategy utilizes the Distillation Step-by-Step method for dataset development and model training applied to a language translation task as a case study using a teacher model to generate labels and rationales using Chain-of-Thought prompting. To determine the most efficient fine-tuning method, three approaches were compared with hyperparameter sweeps: full-precision, Low Rank Adaptation, and Quantized Low Rank Adaptation. Optimal results were obtained by combining Distilling Step-by-Step with full-precision fine-tuning. For resource-constrained environments, Distilling Step-by-Step coupled with Low Rank Adaptation for fine-tuning with Alpha to Rank ratio of 4:1 to balance performance and computation consumption. For environments where GPU compute is at a premium, utilize Distilling Step-by-Step with Quantized Low Rank Adaptation fine-tuning with Alpha to Rank ratio of 4:1. These methods were applied to three sizes of exemplar Text-to-Text Transfer Transformer (T5) fine-tuned Language Net LLM models; T5 models have broad utility in text classification, question answering, and language translation. This work provides a guide for efficiently fine-tuning LLMs for domain-specific tasks with limited data availability.
Keywords: deep learning, Distributed Computing, Fine-tuning, Large language models, neural networks, nlp
Received: 14 Jul 2025; Accepted: 26 Jan 2026.
Copyright: © 2026 Marsh, Michaleas, Ricke, Monera and Zembruski. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Benjamin Marsh
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
