What is LoRA

LoRA, or Low-Rank Adaptation, is a technique designed to make fine-tuning Large Language Models (LLMs) much more efficient. In traditional fine-tuning, you have to update all the parameters of an LLM, which can be slow and resource-intensive. With LoRA, instead of changing the entire model, you add a few small, trainable components to the base LLM. These components are trained for your specific task, allowing the model to adapt quickly without needing to retrain everything from scratch.

When to use LoRA

Because LoRA only updates a small part of the model, it’s much faster than standard fine-tuning. As a general rule, if regular fine-tuning of an LLM on a large dataset would take around 30 minutes, LoRA can often get the job done in 10 minutes or less. This makes it a great choice when you want to adapt a model quickly and efficiently, without needing deep expertise in machine learning. Below is a table comparing when to use LoRA versus full fine-tuning:
ScenarioLoRAFull Fine-tuningReal-world Example
Quick adaptation πŸš€βœ…βŒCustomizing a chatbot for your company’s tone and style
Limited computational resources πŸ’»βœ…βŒSmall startup fine-tuning on a laptop or basic GPU
Small to medium datasets πŸ“Šβœ…βŒTraining on 1,000-10,000 customer support tickets
Fast experimentation πŸ§ͺβœ…βŒTesting different prompt styles for marketing content
Domain-specific tasks πŸŽ―βœ…βš οΈAdapting a model for legal document analysis
Massive datasets πŸ“ˆβŒβœ…Training on millions of medical research papers
Fundamental behavior change πŸ”„βŒβœ…Teaching a general model to code in a new programming language
Maximum performance πŸ†βš οΈβœ…Building a state-of-the-art translation system
Budget constraints πŸ’°βœ…βŒBootstrapped companies with limited cloud computing budget
Time-sensitive projects β°βœ…βŒLaunching a customer service bot in a week

Using LoRA in Prem Studio

1

Create a New Fine-Tuning Job

GIF of clicking to create a new FT JobTo get started, click the + Create Fine-Tuning Job button in the top right corner of the page. Fill in the name, select your dataset, and set the fine-tuning type to Non Reasoning. Note that LoRA is not supported for Reasoning Fine-Tuning. Once you’ve configured these settings, click the Create Fine-Tuning Job button.
2

Configure Your LoRA Settings

GIF of starting LoRA FT Job
  • Choose the model you want to fine-tune and toggle on the LoRA option.
  • Click the Start Experiments button.
  • A confirmation dialog will appear asking you to confirm starting the experiments.
Gemma models (specifically Gemma 3 1B and Gemma 3 4B) are not available for LoRA fine-tuning.
For complex tasks, you may need to increase the number of epochs to achieve the best results.
Once you’ve started the process, your LoRA fine-tuning job will complete in just a few minutes.