Skip to main content RJChow's Notes

M in LLM Stands for Model

A follow up to my previous LLM explainer. In that post, I discussed the recent evolution of LLMs and why the increase to their context window length improved their usefulness significantly.

In this post I will discuss one of the limitations of LLMs, and that is that they are a simplified representation of … their training data.

The M in LLM stands for ‘Model’.

The OED definition of model:

A simplified or idealized description or conception of a particular system, situation, or process, often in mathematical terms, that is put forward as a basis for theoretical or empirical understanding, or for calculations, predictions, etc.; a conceptual or mental representation of something.

All models are wrong

The phrase “all models are wrong” was attributed[1] to George Box who used the phrase in a 1976 paper to refer to the limitations of models, arguing that while no model is ever completely accurate, simpler models can still provide valuable insights if applied judiciously.

As much as LLMs have grown in size recently, they are still smaller than their training data set. This inherently means that some information has been lost, since it is not a lossless compression algorithm.

Due to the probabilistic nature of LLM training, it’s the fine details of reality that are lost. The overall structure of reality is probably quite well represented, within everything that humans have written down and published.

Reality has a surprising amount of detail

A short excerpt from an article by John Salvatier, pre-dating the era of LLMs.

Consider building some basement stairs for a moment. Stairs seem pretty simple at first, and at a high level they are simple, just two long, wide parallel boards (2” x 12” x 16’), some boards for the stairs and an angle bracket on each side to hold up each stair. But as you actually start building you’ll find there’s a surprising amount of nuance.

The first thing you’ll notice is that there are actually quite a few subtasks. Even at a high level, you have to cut both ends of the 2x12s at the correct angles; then screw in some u-brackets to the main floor to hold the stairs in place; then screw in the 2x12s into the u-brackets; then attach the angle brackets for the stairs; then screw in the stairs.

[… condensed for brevity]

Next you’ll notice that each of those steps above decomposes into several steps, some of which have some tricky details to them due to the properties of the materials and task and the limitations of yourself and your tools.

[…]

At every step and every level there’s an abundance of detail with material consequences.

It’s tempting to think ‘So what?’ and dismiss these details as incidental or specific to stair carpentry. And they are specific to stair carpentry; that’s what makes them details. But the existence of a surprising number of meaningful details is not specific to stairs. Surprising detail is a near universal property of getting up close and personal with reality.

Anyone that has ever built a product for users (even something for yourself) would know that building something is never as easy as it sounds. This is because the closer you look at real world systems, or real world users, the more complex they get. Like fractals, every part reveals smaller moving parts, each with its own context, dependencies, and gotchas.

In a product with heritage and legacy, you also need to understand how any change fits into the context of legacy decisions, constraints, team workflows, user expectations, security concerns, and business goals.

Context and Reasoning

The finest examples of LLM productivity improvements show up when the LLM is properly prompted with all the (sidenote: for this reason, I believe that remote-first asynchronous companies have an edge with LLM usage. If all decisions and context around them are written down, it can be easily retrieved for LLMs.) around the task, and is being driven by someone who has this context in their head.

While it is debatable about whether LLMs can truly reason, it can be quite a magical experience watching the chain-of-thoughts internal monologue on a reasoning model.

Some of these reasoning LLMs also have the ability to make web searches. They can retrieve information from the Internet to populate their context, and to fill in some of the fine details that have been lost.

Does that replace humans?

No.

Doing something new - building a novel system, inventing a new algorithm, exploring new science - requires attention to the fine-grained, often invisible layers of reality that don’t follow precedent.

LLMs operate on compressed patterns, and those patterns flatten out the very details that innovation depends on. When a model hasn’t seen something before, it guesses based on statistical gravity, not understanding.

The difference is critical: scientists and engineers handle the unknown by interrogating it. Models interpolate, extrapolate or simulate.

This is why low-effort AI written articles just read like slop. They have a very low signal to noise ratio, and do not contain any original insight.

They’re Tools, Not Replacements

This doesn’t mean LLMs are useless. Far from it. They can accelerate tedious tasks, help explore ideas, and even spot bugs.

But they’re force multipliers - not substitutes. Like any tool, they require someone who knows what they’re doing to wield them safely and effectively.

Since they are a abstract map of concepts, in varying levels of detail, they are also quite good at surfacing related concepts that you may have missed. I don’t think anyone’s denied that LLMs are quite good at taking you from 0 to a strong junior!