← All posts
AI / LLM Strategy

Fine-Tuning Is Back: When It Actually Beats Prompting in 2026

After two years of 'just prompt it,' fine-tuning is quietly winning back ground. Here's when distillation, LoRA, and small specialized models actually beat calling a frontier API.

Craig Hoffmeyer7 min read

For most of 2023 and 2024, the default answer to "should we fine-tune a model" at startups was no. The frontier models were improving so fast that anything you fine-tuned was obsolete before it was deployed. The cost of running inference on a fine-tuned model was usually higher than calling an API for the same task. The engineering overhead of maintaining a fine-tuned model — data pipelines, training infrastructure, evals, versioning — was prohibitive for a small team. "Just prompt it" was the right answer roughly 95% of the time.

That has shifted. Not dramatically, but meaningfully. In 2026, fine-tuning is back on the table for specific use cases, and the set of cases where it wins is larger than it was two years ago. This article is the honest read I give clients when the question comes up, including the specific conditions under which fine-tuning beats prompting, the approaches that work, and the traps that still catch founders.

Why fine-tuning is back

Three things changed.

First, distillation got practical. Distillation is the technique of training a smaller model to imitate the behavior of a larger one on a specific task. The workflow is: use a frontier model to generate a big training set for your task, then fine-tune a small model on that training set. The small model is much cheaper and faster at inference and does almost as well on the specific task. The frontier model does the hard work once; the small model does it forever after.

Second, LoRA and parameter-efficient fine-tuning matured. LoRA (Low-Rank Adaptation) lets you fine-tune a model by training a small additional set of parameters rather than modifying the whole model. The result: fine-tuning a large model is now possible on modest hardware in hours, not days, and with a training dataset in the thousands rather than the millions.

Third, small open-weight models got genuinely good. The gap between "frontier closed model" and "best small open-weight model on a specific task" has narrowed enough that for narrow tasks, a well-tuned small model can match or beat a prompted frontier model at 5–20% of the cost.

The combination means that for a specific shape of problem, fine-tuning is now the economically rational choice. The shape is not most problems, but it is more problems than it used to be.

When fine-tuning wins

Fine-tuning beats prompting when these conditions are all true:

1. The task is narrow and repetitive. Classification, extraction, specific transformations, specific format generation. If the model is doing the same kind of operation millions of times a day, fine-tuning compounds.

2. The volume is high. At 1,000 calls a month, the cost difference is rounding error. At 1 million calls a month, a 10x cost reduction is a budget item.

3. You have training data or can generate it. Either you have labeled examples from your product (golden outputs), or you can use a frontier model to generate synthetic training data (distillation).

4. The task is stable. If the requirements change every week, your fine-tuned model is obsolete every week. Fine-tuning rewards stable problems where the correct answer does not move.

5. Prompting has hit a wall. You have already tried prompting with frontier models and you are unable to hit the quality bar reliably, or the per-request cost is too high at the volume you need.

If all five are true, fine-tuning is worth investigating. If any are false, stay with prompting.

When prompting still wins

The inverse list. Prompting beats fine-tuning when:

The task is broad or exploratory. Features where the user asks for different things each time are a terrible fit for fine-tuning. The frontier model's breadth is the whole value proposition.

Volume is low. If you are calling the model a few thousand times a month total, the engineering cost of fine-tuning (data prep, training infrastructure, eval maintenance, deployment) is more than the inference cost savings.

The ground truth is subjective. For tasks where "correct" is a matter of taste or judgment, fine-tuning is hard because you do not have stable training labels.

You expect the task to evolve. Early-stage products learn what they should be doing by shipping. If your AI feature is going to change shape multiple times in the next few months, fine-tune later.

You lack the operational maturity. If you do not have eval infrastructure, data pipelines, and a deploy process you trust, adding fine-tuning on top of the existing complexity is usually a mistake.

A good heuristic: fine-tune a feature only after it has been running in production with prompting for at least 3 months and you are confident about both the task and the volume.

The approaches that work in 2026

Four specific approaches I see succeeding in real engagements:

1. Classifier distillation

Train a small model (usually a few billion parameters) to mimic a frontier model on a classification or routing task. Use the frontier model to label a large synthetic dataset, then fine-tune. The result is a classifier that runs at 1–5% of the frontier cost and matches it on accuracy for the specific task.

This is the easiest and most reliable fine-tuning project. If your product routes requests or classifies inputs at high volume, this should probably be on your roadmap.

2. Extraction specialization

Fine-tune a small model to extract structured data from unstructured inputs in your specific domain. Examples: pulling line items from invoices, extracting medical codes from clinical notes, pulling specific fields from forms. The structure is narrow, the volume is usually high, and the training data is often already in your system from manual processing.

3. Voice and style tuning

Fine-tune a model to write in a specific voice — your company's support tone, your specific legal language, the domain jargon of your customers. Prompting can get you most of the way, but fine-tuning locks in the style more reliably and cheaply at scale.

4. Domain-specific reasoning on small models

For tasks with domain-specific reasoning (medical, legal, financial), a small model fine-tuned on the domain can match a prompted frontier model at a fraction of the cost. This is the hardest of the four because evaluating domain reasoning is itself hard, and the training data is often sensitive.

The workflow

If you decide to fine-tune, the workflow I use is:

Week 1: Data collection. Pull real examples from production, label them (or use a frontier model to label them), and build a training set of 1,000–10,000 examples. This is where most of the real work lives.

Week 1 also: Build the eval. Never fine-tune without an eval that is independent of the training set. The eval tells you whether the fine-tuned model is actually better than the prompted baseline.

Week 2: Train. Use LoRA or full fine-tuning on an open-weight base model. The training itself is usually a few hours to a day on rented GPU time. Multiple runs with different hyperparameters.

Week 2 also: Evaluate. Run the eval against the fine-tuned candidates. Compare to the prompted baseline. The candidate has to beat the baseline on quality or match it at a meaningfully lower cost.

Week 3: Deploy. Ship the fine-tuned model to a small traffic slice behind a feature flag. Compare production metrics against the baseline. Roll out gradually.

Ongoing: Retrain. Task definitions drift. Data distributions shift. Plan to retrain every few months and re-evaluate every release. Fine-tuned models are not set-and-forget.

The costs and traps

Real costs founders underestimate:

Training data prep is the biggest line item. Not training compute. Not inference costs. Labeling, cleaning, and curating training data eats most of the engineering time. Budget for it.

Hosting the fine-tuned model is its own ops problem. You now need to run inference on a model you own. Managed services (Fireworks, Together, Modal, Replicate) handle most of this, but you still have decisions to make about GPU type, scaling, and cold starts.

The model gets stale. Frontier models improve constantly. Your fine-tuned model does not. The relative advantage shrinks over time. Plan for the comparison being worse next quarter than this quarter.

Eval drift sneaks up. As you iterate, the eval itself has to evolve. If you are not careful, the eval and the fine-tuning optimize against each other in weird ways.

The team ends up loving the model too much to replace it. Sunk cost is real. The team that built the fine-tuned model will defend it past the point where prompting has caught up. Be prepared to retire it when the math stops working.

Counterpoint: the default is still "just prompt it"

A warning. Despite the shift, the default at startups in 2026 is still prompting. Fine-tuning makes sense for a minority of features, and even for those, I usually recommend waiting until the feature is mature before investing. Do not fine-tune because it sounds sophisticated. Fine-tune because the numbers say it is the right move.

Your next step

If you have a feature in production that has been running stably for several months at high volume, calculate the current monthly inference cost. If it is a big enough number to matter, run the mental check: would a fine-tuned model plausibly save 70%+ on this feature? If yes, it is worth a real investigation. If no, stay the course.

Where I come in

Fine-tuning engagements are a small but growing part of my AI work in 2026. Usually a 3–5 week investigation for a specific feature, with a clear go/no-go decision based on measured comparison to the prompted baseline. Book a call if you have a high-volume AI feature and want an honest read on whether fine-tuning is worth pursuing.


Related reading: Model Selection Guide · The Hidden Cost Curve of LLM Features · Why Your Eval Suite Matters More Than Your Prompt

Wondering if fine-tuning fits your feature? Book a call.

Get in touch →