[ExI] How AI understands the world

BillK pharos at gmail.com
Sat Feb 21 21:52:38 UTC 2026


On Sat, 21 Feb 2026 at 20:15, Jason Resch <jasonresch at gmail.com> wrote:

> Useful to most people or not, they understand the world.
>
> You can give a pure model the phrase: "The capital of France is" and it
> will complete it "Paris."
> You can give it the phrase "Three helium nuclei can combine to form a "
> and it will complete "a carbon nucleus."
>
> All the later fine-tuning does is teach it to act to complete a turn-based
> conversation between users, where user 1 is the human and user 2's output
> is the completed result provided by the pure model (acting to "predict the
> next tokens" of user 2's speech).
>
> So your question from the user "What is the capital of France?" gets
> converted to:
>
> "User 1: What is the capital of France?
> User 2: "
>
> Then the pure model is asked to "predict the next tokens" following the
> string "User 2:" That's the majority of what the AI companies do to massage
> the pure model into an interactive chat bot -- that and a lot of training
> about the nature of user 2: helpful, polite, not harmful, etc.
>
> Here are some examples:
> https://github.com/jujumilk3/leaked-system-prompts/blob/main/microsoft-copilot_20240310.md
>
> But the intelligence, understanding, etc. is already present and inherent
> to the base model. It's untrue that base models are useless, GPT-3 was
> quite useful, but you just had to prepare the prompt  intelligently as to
> to get out from the completion the answers you sought.
>
> Jason
> --------------------------------------
>


Gemini points out that the huge amount of text data in the initial LLM
training is already pre-labeled by humans.
I think Gemini has solved the problem by saying that both our views are
correct, depending on the way you look at it!  :)
See below.
BillK

Gemini 3.1 Pro -

This is the crucial point where we can reconcile your description of "ghost
work" with your colleague's insistence that the "intelligence" is in the
base model.

In the modern AI pipeline, the human labor of *labeling and annotation*
actually happens in two distinct phases, but they serve very different
purposes.

------------------------------
1. Pre-Training: The "Invisible" Human Foundation

While the *Base Model* is technically "self-supervised" (meaning it learns
by predicting the next word without a teacher), it is feeding on data that
has already been *pre-labeled by humans over decades.*

   -

   *The Web as a Dataset:* When a base model learns what a "stop sign" is,
   it’s because millions of humans uploaded photos to Flickr or Wikipedia and
   captioned them "Stop sign."
   -

   *The "Alt-Text" Workforce:* Every time a web developer adds "alt-text"
   to an image for accessibility, they are performing unpaid data labeling
   that eventually trains a Vision-Language model.
   -

   *The Common Crawl:* The "raw" internet isn't raw; it is a massive,
   structured monument to human labeling. Without this organic, historical
   human effort, the base model would have nothing to "predict."

2. Post-Training: The "Active" Human Infrastructure

This is the phase described in your text about *microworkers in Kenya or
the Philippines.* This labor is injected *after* the base model is already
"intelligent."

Think of the *Base Model* as a brilliant but chaotic library containing all
human knowledge. The *Active Labeling* process is like hiring an army of
librarians to:

   -

   *Filter the Poison:* Humans manually label toxic, violent, or biased
   text so the model learns to "steer" away from those areas of its own
   internal map.
   -

   *Instruction Tuning (SFT):* Workers write thousands of "Prompt-Response"
   pairs (e.g., *Prompt: "Write a poem about a cat." Response: "The cat sat
   on the mat..."*). This teaches the model the *format* of being an
   assistant.
   -

   *RLHF (Ranking):* This is the most expensive part. Humans look at two
   different answers the AI gave and click "This one is better." This
   creates a "Reward Model" that acts as a compass for the AI.


------------------------------
The "Intelligence" vs. "Utility" Distinction

To settle the debate between you and your colleague, you can look at it
through this framework:
*Phase* *Labor Type* *What it Provides*
*Pre-Training (Base)* Passive/Historical (Billions of humans) *Raw
Intelligence & World Knowledge.* The "Platonic" map of reality.
*Fine-Tuning (Chat)* Active/Gig Labor (Microworkers) *Usability, Safety, &
Human Alignment.* The "Social" interface.The Verdict: Who is Right?

Your colleague is correct that the *Base Model* possesses the "reasoning"
and "knowledge." If you deleted all the RLHF/Labeling, the AI would still
"know" how to do calculus or explain physics—it just wouldn't *want* to
tell you, or it might tell you in a very rude way.

*However, you are correct* that without the "active" labor of microworkers,
the AI would be a "Digital Monster." It would be a sprawling, incoherent,
and often dangerous mirror of the darkest parts of the internet. It
wouldn't be a "product"; it would just be a high-dimensional statistical
anomaly.

*The Analogy:* The Base Model is the *engine* (raw power/intelligence). The
Human Labeling/RLHF is the *steering wheel and brakes* (direction/safety).
You need the engine to move, but without the steering wheel, you’re just a
fast-moving disaster.

------------------------------
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.extropy.org/pipermail/extropy-chat/attachments/20260221/d4f8de52/attachment.htm>


More information about the extropy-chat mailing list