Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Artif. Intell.

Sec. Natural Language Processing

This article is part of the Research TopicThe Use of Large Language Models to Automate, Enhance, and Streamline Text Analysis Processes. Large Language Models Used to Analyze and Check Requirement Compliance.View all 9 articles

An Efficient Strategy for Fine-Tuning Large Language Models

Provisionally accepted
Benjamin  MarshBenjamin Marsh1*Adam  MichaleasAdam Michaleas2Darrell  RickeDarrell Ricke2Shaun  MoneraShaun Monera1Shriya  ZembruskiShriya Zembruski1
  • 1Marine Corps Tactical Systems Support Activity, US Marine Corps, Arlington, United States
  • 2Massachusetts Institute of Technology Lincoln Laboratory, Lexington, United States

The final, formatted version of the article will be published soon.

ABSTRACT Large Language Models (LLMs) are powerful tools for Natural Language Processing tasks; fine-tuning them for domain-specific applications is resource-intensive, requiring significant effort to develop datasets and considerable compute resources for fine-tuning. This work establishes a combination of techniques that enable users to rapidly fine-tune LLMs for domain-specific tasks using limited datasets and performs a comparative analysis across established fine-tuning methods using this strategy to determine the most efficient fine-tuning method to use in conjunction with these techniques. This strategy utilizes the Distillation Step-by-Step method for dataset development and model training applied to a language translation task as a case study using a teacher model to generate labels and rationales using Chain-of-Thought prompting. To determine the most efficient fine-tuning method, three approaches were compared with hyperparameter sweeps: full-precision, Low Rank Adaptation, and Quantized Low Rank Adaptation. Optimal results were obtained by combining Distilling Step-by-Step with full-precision fine-tuning. For resource-constrained environments, Distilling Step-by-Step coupled with Low Rank Adaptation for fine-tuning with Alpha to Rank ratio of 4:1 to balance performance and computation consumption. For environments where GPU compute is at a premium, utilize Distilling Step-by-Step with Quantized Low Rank Adaptation fine-tuning with Alpha to Rank ratio of 4:1. These methods were applied to three sizes of exemplar Text-to-Text Transfer Transformer (T5) fine-tuned Language Net LLM models; T5 models have broad utility in text classification, question answering, and language translation. This work provides a guide for efficiently fine-tuning LLMs for domain-specific tasks with limited data availability.

Keywords: deep learning, Distributed Computing, Fine-tuning, Large language models, neural networks, nlp

Received: 14 Jul 2025; Accepted: 26 Jan 2026.

Copyright: © 2026 Marsh, Michaleas, Ricke, Monera and Zembruski. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Benjamin Marsh

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.