LLMs vs Fine-Tuned Models: When Zero-Shot Works - and When Specialized Models Win
Instruction-tuned LLMs can often do classification and generation without retraining — called zero-shot (no examples) or few-shot (a handful of examples). In many recent tests, these decoder-only models match or even beat older, task-specific BERT-like models on things like sentiment, topic, and genre classification, including experiments in South Slavic languages. That sounds powerful, but it’s not the whole story: LLMs are slower, more expensive to run, and their outputs are less predictable. For high-volume, repeatable pipelines, fine-tuned BERT-style models still make more sense in practice.
What those comparisons actually showed (plain language):
- LLMs often reach strong zero-shot performance on classification tasks; sometimes they match or exceed fine-tuned BERT-like models.
- Zero-shot performance of LLMs was similar across South Slavic languages and English, meaning LLMs can generalize across languages better than expected in some settings.
- Big downsides of LLMs: unpredictable outputs (harder to control or debug), much slower inference (each request takes longer), and higher computational cost (cloud bills or GPUs add up).
- When you need massive, repeated automatic annotation (e.g., labeling millions of documents), fine-tuned BERT-like models are still the practical winner.
Parliamentary speech generation — a special-case example: generating realistic political speeches isn’t just about fluent language. It must sound politically authentic and stick to an ideological line. To measure that, researchers built ParliaBench (a UK Parliament speech dataset) and an evaluation setup that mixes standard metrics with LLM-as-a-judge evaluations. They also created two new embedding-based measures — Political Spectrum Alignment and Party Alignment — to quantify how closely generated speeches match party or ideological positions.
Key findings there:
- Fine-tuning LLMs on parliamentary data produced clear improvements across linguistic, semantic, and political-authenticity metrics.
- The new embedding-based metrics proved useful for detecting ideological alignment — they compare a speech’s vector (embedding) to reference vectors representing parties or the ideological axis.
- Using an LLM to judge political authenticity is practical, but it should be combined with automatic scores and human checks because LLM evaluators have their own biases.
Model architecture choices matter — but not always how people expect: positional encodings (how a model represents word order) were tested in monolingual models using absolute, relative, or no positional encodings across typologically diverse languages. A common theory said languages with complex word morphology can get away with flexible word order (and so might tolerate different positional encodings). The experiments didn’t find a clear, consistent interaction: the effect of positional encodings depends a lot on the language, tasks, and metrics. In short, there’s no single encoding strategy that wins everywhere.
When domain structure is complex, specialized models still win — example: Sanskrit poetry → prose conversion: converting Sanskrit verse to canonical prose requires many language-specific steps (breaking compounds, resolving dependencies, re-ordering words according to grammar and meter). Instruction-tuned LLMs and in-context prompting helped, but a fully fine-tuned, task-specific Seq2Seq model built for Sanskrit (ByT5-Sanskrit) outperformed all LLM approaches. Human evaluations agreed with automatic metrics: domain-specific fine-tuning matter a lot for tasks requiring structured, multi-step reasoning.
Practical decision checklist — which path to pick?
- If you need quick prototypes or have no labeled data: try LLM zero-shot/few-shot prompts. Fast to try, good baseline — but verify outputs carefully.
- If you run high-volume, repeatable labeling: fine-tune an encoder (BERT-like) or a small Seq2Seq model. Faster and cheaper at scale, and more predictable.
- If the task needs deep domain knowledge or multi-step reasoning (legal language, poetry, parliamentary style): collect task data and fine-tune a specialized model. It will usually outperform prompts alone.
- If compute cost and latency matter: avoid repeated calls to big LLMs; consider distillation, quantization, or fine-tuning a smaller model.
- If you care about ideological authenticity or alignment: use domain-aware metrics (embedding alignment, party-spectrum checks) and include human judges.
Evaluation tips that help you avoid surprises
- Don’t rely only on standard NLP scores (accuracy, F1, BLEU). For political or stylistic tasks, add domain-specific measures like Political Spectrum Alignment and Party Alignment.
- Use a mix of automatic metrics, LLM-as-judge heuristics, and human evaluation. LLM judges are fast but biased; human checks calibrate them.
- Measure inference time and cost per query during evaluation — a model that’s slightly more accurate but 10x slower could be impractical in production.
- For low-resource languages, test both: LLM zero-shot (good for quick coverage) and a fine-tuned small model (better for robust, repeatable systems).
Bottom line: LLMs are powerful and can cover many tasks without retraining, but they are costly and less predictable. For reliable, large-scale, or deeply structured tasks — especially in specific languages or political domains — targeted fine-tuning of compact models still gives the best practical results.
Related Papers
- arXiv Query: search_query=&id_list=2511.07989v1&start=0&max_results=10
Until recently, fine-tuned BERT-like models provided state-of-the-art performance on text classification tasks. With the rise of instruction-tuned decoder-only models, commonly known as large language…
- arXiv Query: search_query=&id_list=2511.08247v1&start=0&max_results=10
Parliamentary speech generation presents specific challenges for large language models beyond standard text generation tasks. Unlike general text generation, parliamentary speeches require not only li…
- arXiv Query: search_query=&id_list=2511.08139v1&start=0&max_results=10
Language model architectures are predominantly first created for English and subsequently applied to other languages. It is an open question whether this architectural bias leads to degraded performan…
- arXiv Query: search_query=&id_list=2511.08145v1&start=0&max_results=10
Large Language Models (LLMs) are increasingly treated as universal, general-purpose solutions across NLP tasks, particularly in English. But does this assumption hold for low-resource, morphologically…
Hiring AI researchers or engineers?
Your job post reaches ML engineers, PhD researchers, and AI leads who read arXiv daily. Transparent pricing, real impression data, no middlemen.
