When Should You Actually Train an AI Model? (Rarely — But Sometimes It Makes Sense)

Published on 13 Dec 2025 by New Media Aid — bespoke SME app development since the year 2000

AI development Machine Learning Software Engineering AI decision making Applied AI SME systems Model training LLMs


In a previous article, I argued that most senior software engineers should not be trying to train AI models — especially when building systems for SMEs. That position often raises a fair follow-up question:

“So when should you train a model?”

The honest answer is: rarely — but sometimes it does make sense.

This article explores those exceptions, and, just as importantly, explains how to recognise when training a model is the wrong tool for the job.


The Default Assumption Should Be: Don’t Train

Before we look at the exceptions, it is important to set the baseline.

Training an AI model should never be the default choice. It introduces significant overhead:

  • Data collection and cleaning
  • Labelling and validation
  • Infrastructure costs
  • Ongoing maintenance and monitoring

For most software projects — even AI-powered ones — existing models combined with prompts, guardrails and retrieval will solve the problem more effectively and with far less risk.

If an off-the-shelf LLM with domain context can do the job, training your own model is usually unnecessary engineering.


The First Valid Reason: You Have a Truly Unique Dataset

One of the few legitimate reasons to train a model is when you possess data that:

  • Is proprietary
  • Is unavailable publicly
  • Cannot be effectively retrieved at runtime
  • Represents patterns rather than simple facts

Examples might include:

  • Years of operational sensor data
  • Large volumes of annotated inspections or reports
  • Historical behaviour data that reveals trends, not text answers

In these cases, knowledge is embedded in patterns — not documents a model can simply read.

Even then, the bar should remain high. Having data is not enough; it needs to be useful, clean and representative.


The Second Reason: You Need Deterministic Classification

Another scenario where model training can make sense is constrained classification problems.

Examples include:

  • Sorting inbound messages into fixed categories
  • Flagging documents that meet specific criteria
  • Detecting anomalies within structured data

In these cases:

  • The input format is consistent
  • The output categories are known
  • Accuracy can be measured objectively

Small, focused models — or even traditional ML techniques — can outperform large language models here.

Crucially, these are narrow problems, not general reasoning tasks.


The Third Reason: Latency, Cost or Isolation Requirements

There are practical considerations that sometimes justify training:

  • Strict latency constraints
  • Offline or air-gapped environments
  • Regulatory requirements around data residency

In these situations, running a small local model can be more predictable than calling external APIs.

However, this is a trade-off decision — not an argument that training models is inherently better.

It also assumes you are prepared to own and maintain that model long-term.


What Usually Is Not a Good Reason

Some motivations sound compelling but rarely stand up to scrutiny:

  • “We want control” — control without expertise becomes liability
  • “Our data is special” — often it isn’t
  • “This will differentiate us” — customers buy outcomes, not models
  • “We don’t want vendor lock-in” — self-maintained AI is its own lock-in

These reasons often mask discomfort with dependency rather than a genuine technical need.


RAG vs Training: A Crucial Distinction

One of the most common mistakes is choosing training when retrieval would do.

If your problem is:

  • Answering questions about policies
  • Explaining regulations
  • Summarising internal documentation

Then retrieval-augmented generation is almost always superior.

RAG offers:

  • Up-to-date information
  • Auditability
  • Simple updates
  • Clear data ownership

Training a model to “remember” facts is expensive and brittle compared to letting it read them when needed.


The Senior Engineer’s Role: Saying No Early

One of the most valuable contributions a senior engineer can make is preventing unnecessary complexity.

That often means asking hard questions:

  • What failure looks like
  • How correctness will be measured
  • Who maintains this in 12 months
  • What happens when requirements change

Model training dramatically increases the cost of getting these answers wrong.


A Practical Decision Framework

Before training a model, a pragmatic checklist helps:

  • Can an existing LLM solve this with context?
  • Is RAG sufficient?
  • Is the problem deterministic?
  • Do we have enough quality data?
  • Can success be measured objectively?
  • Are we prepared to maintain the model?

If several of these answers are “no”, training is almost certainly the wrong choice.


Final Thoughts

Training AI models is not inherently bad — but it is a specialised tool for specific situations.

Most senior software engineers add far more value by:

  • Using models wisely
  • Constraining behaviour
  • Designing robust systems
  • Focusing on business outcomes

Knowing when not to train a model is often the clearest sign of experience.

Previous blog

Why Senior Software Engineers Should Stop Trying to Train AI Models (and Focus on Making AI Actually Useful)

A pragmatic look at why most senior software engineers — especially solo developers and SME specialists — should not waste time training AI models, and should instead focus on using existing LLMs effectively with prompts, guardrails, RAG and system design.