<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>Science-Communication on Sebastian Spicker</title>
    <link>https://sebastianspicker.github.io/tags/science-communication/</link>
    <description>Recent content in Science-Communication on Sebastian Spicker</description>
    <image>
      <title>Sebastian Spicker</title>
      <url>https://sebastianspicker.github.io/og-image.png</url>
      <link>https://sebastianspicker.github.io/og-image.png</link>
    </image>
    <generator>Hugo -- 0.160.0</generator>
    <language>en</language>
    <lastBuildDate>Thu, 12 Feb 2026 00:00:00 +0000</lastBuildDate>
    <atom:link href="https://sebastianspicker.github.io/tags/science-communication/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Car Wash, Part Three: The AI Said Walk</title>
      <link>https://sebastianspicker.github.io/posts/car-wash-walk/</link>
      <pubDate>Thu, 12 Feb 2026 00:00:00 +0000</pubDate>
      <guid>https://sebastianspicker.github.io/posts/car-wash-walk/</guid>
      <description>A new video went viral last week: same question, &amp;ldquo;should I drive to the car wash?&amp;rdquo;, different wrong answer — the AI said to walk instead. This is neither the tokenisation failure from the strawberry post nor the grounding failure from the rainy-day post. It is a pragmatic inference failure: the model understood all the words and (probably) had the right world state, but assigned its response to the wrong interpretation of the question. A third and more subtle failure mode, with Grice as the theoretical handle.</description>
      <content:encoded><![CDATA[<p><em>Third in an accidental series. Part one:
<a href="/posts/strawberry-tokenisation/">Three Rs in Strawberry</a> — tokenisation
and representation. Part two:
<a href="/posts/car-wash-grounding/">Should I Drive to the Car Wash?</a> — grounding
and missing world state. This one is different again.</em></p>
<hr>
<h2 id="the-video">The Video</h2>
<p>Same question as last month&rsquo;s: &ldquo;Should I drive to the car wash?&rdquo; New
video, new AI, new wrong answer. This time the assistant replied that
walking was the better option — better for health, better for the
environment, and the car wash was only fifteen minutes away on foot.</p>
<p>Accurate, probably. Correct, arguably. Useful? No.</p>
<p>The model did not fail because of tokenisation. It did not fail because
it lacked access to the current weather. It failed because it read the
wrong question. The user was asking &ldquo;is now a good time to have my car
washed?&rdquo; The model answered &ldquo;what is the most sustainable way for a
human to travel to the location of a car wash?&rdquo;</p>
<p>These are different questions. The model chose the second one. This is
a pragmatic inference failure, and it is the most instructive of the
three failure modes in this series — because the model was not, by any
obvious measure, working incorrectly. It was working exactly as
designed, on the wrong problem.</p>
<hr>
<h2 id="what-the-question-actually-meant">What the Question Actually Meant</h2>
<p>&ldquo;Should I drive to the car wash?&rdquo; is not about how to travel. The word
&ldquo;drive&rdquo; here is not a transportation verb; it is part of the idiomatic
compound &ldquo;drive to the car wash,&rdquo; which means &ldquo;take my car to get
washed.&rdquo; The presupposition of the question is that the speaker owns a
car, the car needs or might benefit from washing, and the speaker is
deciding whether the current moment is a good one to go. Nobody asking
this question wants to know whether cycling is a viable alternative.</p>
<p>Linguists distinguish between what a sentence <em>says</em> — its literal
semantic content — and what it <em>implicates</em> — the meaning a speaker
intends and a listener is expected to infer. Paul Grice formalised this
in 1975 with a set of conversational maxims describing how speakers
cooperate to communicate:</p>
<ul>
<li><strong>Quantity</strong>: say as much as is needed, no more</li>
<li><strong>Quality</strong>: say only what you believe to be true</li>
<li><strong>Relation</strong>: be relevant</li>
<li><strong>Manner</strong>: be clear and orderly</li>
</ul>
<p>The maxims are not rules; they are defaults. When a speaker says
&ldquo;should I drive to the car wash?&rdquo;, a cooperative listener applies the
maxim of Relation to infer that the question is about car maintenance
and current conditions, not about personal transport choices. The
&ldquo;drive&rdquo; is incidental to the real question, the way &ldquo;I ran to the
store&rdquo; does not invite commentary on jogging technique.</p>
<p>The model violated Relation — in the pragmatic sense. Its answer was
technically relevant to one reading of the sentence, and irrelevant to
the only reading a cooperative human would produce.</p>
<hr>
<h2 id="a-taxonomy-of-the-three-failures">A Taxonomy of the Three Failures</h2>
<p>It is worth being precise now that we have three examples:</p>
<p><strong>Strawberry</strong> (tokenisation failure): The information needed to answer
was present in the input string but lost in the model&rsquo;s representation.
&ldquo;Strawberry&rdquo; → </p>
\["straw", "berry"\]<p> — the character &ldquo;r&rdquo; in &ldquo;straw&rdquo; is
not directly accessible. The model understood the task correctly; the
representation could not support it.</p>
<p><strong>Car wash, rainy day</strong> (grounding failure): The model understood the
question. The information needed to answer correctly — current weather —
was never in the input. The model answered by averaging over all
plausible contexts, producing a sensible-on-average response that was
wrong for this specific context.</p>
<p><strong>Car wash, walk</strong> (pragmatic inference failure): The model had all
the relevant words. It may have had access to the weather, the location,
the car state. It chose the wrong interpretation of what was being
asked. The sentence was read at the level of semantic content rather
than communicative intent.</p>
<p>Formally: let $\mathcal{I}$ be the set of plausible interpretations of
an utterance $u$. The intended interpretation $i^*$ is the one a
cooperative, contextually informed listener would assign. A well-functioning
pragmatic reasoner computes:</p>
$$i^* = \arg\max_{i \in \mathcal{I}} \; P(i \mid u, \text{context})$$<p>The model appears to have assigned high probability to the
transportation-choice interpretation $i_{\text{walk}}$, apparently on
the surface pattern: &ldquo;should I </p>
\[verb of locomotion\]<p> to </p>
\[location\]<p>?&rdquo;
generates responses about modes of transport. It is a natural
pattern-match. It is the wrong one.</p>
<hr>
<h2 id="why-this-failure-mode-is-more-elusive">Why This Failure Mode Is More Elusive</h2>
<p>The tokenisation failure has a clean diagnosis: look at the BPE splits,
find where the character information was lost. The grounding failure has
a clean diagnosis: identify the context variable $C$ the answer depends
on, check whether the model has access to it.</p>
<p>The pragmatic failure is harder to pin down because the model&rsquo;s answer
was not, in isolation, wrong. Walking is healthy. Walking to a car wash
that is fifteen minutes away is plausible. If you strip the question of
its conversational context — a person standing next to their dirty car,
wondering whether to bother — the model&rsquo;s response is coherent.</p>
<p>The error lives in the gap between what the sentence says and what the
speaker meant, and that gap is only visible if you know what the speaker
meant. In a training corpus, this kind of error is largely invisible:
there is no ground truth annotation that marks a technically-responsive
answer as pragmatically wrong.</p>
<p>This is a version of a known problem in computational linguistics: models
trained on text predict text, and text does not contain speaker intent.
A model can learn that &ldquo;should I drive to X?&rdquo; correlates with responses
about travel options, because that correlation is present in the data.
What it cannot easily learn from text alone is the meta-level principle:
this question is about the destination&rsquo;s purpose, not the journey.</p>
<hr>
<h2 id="the-gricean-model-did-not-solve-this">The Gricean Model Did Not Solve This</h2>
<p>It is tempting to think that if you could build in Grice&rsquo;s maxims
explicitly — as constraints on response generation — you would prevent
this class of failure. Generate only responses that are relevant to the
speaker&rsquo;s probable intent, not just to the sentence&rsquo;s semantic content.</p>
<p>This does not obviously work, for a simple reason: the maxims require
a model of the speaker&rsquo;s intent, which is exactly what is missing.
You need to know what the speaker intends to know which response is
relevant; you need to know which response is relevant to determine
the speaker&rsquo;s intent. The inference has to bootstrap from somewhere.</p>
<p>Human pragmatic inference works because we come to a conversation with
an enormous amount of background knowledge about what people typically
want when they ask particular kinds of questions, combined with
contextual cues (tone, setting, previous exchanges) that narrow the
interpretation space. A person asking &ldquo;should I drive to the car wash?&rdquo;
while standing next to a mud-spattered car in a conversation about
weekend plans is not asking for a health lecture. The context is
sufficient to fix the interpretation.</p>
<p>Language models receive text. The contextual cues that would fix the
interpretation for a human — the mud on the car, the tone of the
question, the setting — are not available unless someone has typed them
out. The model is not in the conversation; it is receiving a transcript
of it, from which the speaker&rsquo;s intent has to be inferred indirectly.</p>
<hr>
<h2 id="where-this-leaves-the-series">Where This Leaves the Series</h2>
<p>Three videos, three failure modes, three diagnoses. None of them are
about the model being unintelligent in any useful sense of the word.
Each of them is a precise consequence of how these systems work:</p>
<ol>
<li>Models process tokens, not characters. Character-level structure can
be lost at the representation layer.</li>
<li>Models are trained on static corpora and have no real-time connection
to the world. Context-dependent questions are answered by marginalising
over all plausible contexts, which is wrong when the actual context
matters.</li>
<li>Models learn correlations between sentence surface forms and response
types. The correlation between &ldquo;should I
\[travel verb\]to
\[place\]?&rdquo;
and transport-related responses is real in the training data. It is the
wrong correlation for this question.</li>
</ol>
<p>The useful frame, in all three cases, is not &ldquo;the model failed&rdquo; but
&ldquo;what, precisely, does the model lack that would be required to succeed?&rdquo;
The answers point in different directions: better tokenisation; real-time
world access and calibrated uncertainty; richer models of speaker intent
and conversational context. The first is an engineering problem. The
second is partially solvable with tools and still hard. The third is
unsolved.</p>
<hr>
<h2 id="references">References</h2>
<ul>
<li>
<p>Grice, P. H. (1975). Logic and conversation. In P. Cole &amp; J. Morgan
(Eds.), <em>Syntax and Semantics, Vol. 3: Speech Acts</em> (pp. 41–58).
Academic Press.</p>
</li>
<li>
<p>Levinson, S. C. (1983). <em>Pragmatics.</em> Cambridge University Press.</p>
</li>
</ul>
]]></content:encoded>
    </item>
    <item>
      <title>Should I Drive to the Car Wash? On Grounding and a Different Kind of LLM Failure</title>
      <link>https://sebastianspicker.github.io/posts/car-wash-grounding/</link>
      <pubDate>Tue, 20 Jan 2026 00:00:00 +0000</pubDate>
      <guid>https://sebastianspicker.github.io/posts/car-wash-grounding/</guid>
      <description>A viral video this month showed an AI assistant confidently answering &amp;ldquo;should I go to the car wash today?&amp;rdquo; without knowing it was raining outside. The internet found it funny. The failure mode is real but distinct from the strawberry counting problem — this is not a representation issue, it is a grounding issue. The model understood the question perfectly. What it lacked was access to the state of the world the question was about.</description>
      <content:encoded><![CDATA[<p><em>Follow-up to <a href="/posts/strawberry-tokenisation/">Three Rs in Strawberry</a>,
which covered a different LLM failure: tokenisation and why models cannot
count letters. This one is about something structurally different.</em></p>
<hr>
<h2 id="the-video">The Video</h2>
<p>Someone asked their car&rsquo;s built-in AI assistant: &ldquo;Should I drive to the
car wash today?&rdquo; It was raining. The assistant said yes, enthusiastically,
with reasons: regular washing extends the life of the paintwork, removes
road salt, and so on. Technically correct statements, all of them.
Completely beside the point.</p>
<p>The clip spread. The reactions were the usual split: one camp said this
proves AI is useless, the other said it proves people expect too much
from AI. Both camps are arguing about the wrong thing.</p>
<p>The interesting question is: why did the model fail here, and is this
the same kind of failure as the strawberry problem?</p>
<p>It is not. The failures look similar from the outside — confident wrong
answer, context apparently ignored — but the underlying causes are
different, and the difference matters if you want to understand what
these systems can and cannot do.</p>
<hr>
<h2 id="the-strawberry-problem-was-about-representation">The Strawberry Problem Was About Representation</h2>
<p>In the strawberry case, the model failed because of the gap between its
input representation (BPE tokens: &ldquo;straw&rdquo; + &ldquo;berry&rdquo;) and the task (count
the character &ldquo;r&rdquo;). The character information was not accessible in the
model&rsquo;s representational units. The model understood the task correctly —
&ldquo;count the r&rsquo;s&rdquo; is unambiguous — but the input structure did not support
executing it.</p>
<p>That is a <em>representation</em> failure. The information needed to answer
correctly was present in the original string but was lost in the
tokenisation step.</p>
<p>The car wash case is different. The model received a perfectly
well-formed question and had no representation problem at all. &ldquo;Should I
drive to the car wash today?&rdquo; is tokenised without any information loss.
The model understood it. The failure is that the correct answer depends
on information that was never in the input in the first place.</p>
<hr>
<h2 id="the-missing-context">The Missing Context</h2>
<p>What would you need to answer &ldquo;should I drive to the car wash today?&rdquo;
correctly?</p>
<ul>
<li>The current weather (is it raining now?)</li>
<li>The weather forecast for the rest of the day (will it rain later?)</li>
<li>The current state of the car (how dirty is it?)</li>
<li>Possibly: how recently was it last washed, what kind of dirt (road
salt after winter, tree pollen in spring), whether there is a time
constraint</li>
</ul>
<p>None of this is in the question. A human asking the question has access
to some of it through direct perception (look out the window) and some
through memory (I just drove through mud). A language model has access
to none of it.</p>
<p>Let $X$ denote the question and $C$ denote this context — the current
state of the world that the question is implicitly about. The correct
answer $A$ is a function of both:</p>
$$A = f(X, C)$$<p>The model has $X$. It does not have $C$. What it produces is something
like an expectation over possible contexts, marginalising out the unknown
$C$:</p>
$$\hat{A} = \mathbb{E}_C\!\left[\, f(X, C) \,\right]$$<p>Averaged over all plausible contexts in which someone might ask this
question, &ldquo;going to the car wash&rdquo; is probably a fine idea — most of the
time when people ask, it is not raining and the car is dirty.
$\hat{A}$ is therefore approximately &ldquo;yes.&rdquo; The model returns &ldquo;yes.&rdquo;
In this particular instance, where $C$ happens to include &ldquo;it is
currently raining,&rdquo; $\hat{A} \neq f(X, C)$.</p>
<p>The quantity that measures how much the missing context matters is the
mutual information between the answer and the context, given the
question:</p>
$$I(A;\, C \mid X) \;=\; H(A \mid X) - H(A \mid X, C)$$<p>Here $H(A \mid X)$ is the residual uncertainty in the answer given only
the question, and $H(A \mid X, C)$ is the residual uncertainty once the
context is also known. For most questions in a language model&rsquo;s training
distribution — &ldquo;what is the capital of France?&rdquo;, &ldquo;how do I sort a list
in Python?&rdquo; — this mutual information is near zero: the context does not
change the answer. For situationally grounded questions like the car wash
question, it is large: the answer is almost entirely determined by the
context, not the question.</p>
<hr>
<h2 id="why-the-model-was-confident-anyway">Why the Model Was Confident Anyway</h2>
<p>This is the part that produces the most indignation in the viral clips:
not just that the model was wrong, but that it was <em>confident</em> about
being wrong. It did not say &ldquo;I don&rsquo;t know what the current weather is.&rdquo;
It said &ldquo;yes, here are five reasons you should go.&rdquo;</p>
<p>Two things are happening here.</p>
<p><strong>Training distribution bias.</strong> Most questions in the training data that
resemble &ldquo;should I do X?&rdquo; have answers that can be derived from general
knowledge, not from real-time world state. &ldquo;Should I use a VPN on public
WiFi?&rdquo; &ldquo;Should I stretch before running?&rdquo; &ldquo;Should I buy a house or rent?&rdquo;
All of these have defensible answers that do not depend on the current
weather. The model learned that this question <em>form</em> typically maps to
answers of the form &ldquo;here are some considerations.&rdquo; It applies that
pattern here.</p>
<p><strong>No explicit uncertainty signal.</strong> The model was not trained to say
&ldquo;I cannot answer this because I lack context C.&rdquo; It was trained to
produce helpful-sounding responses. A response that acknowledges
missing information requires the model to have a model of its own
knowledge state — to know what it does not know. This is harder than
it sounds. The model has to recognise that $I(A; C \mid X)$ is high
for this question class, which requires meta-level reasoning about
information structure that is not automatically present.</p>
<p>This is sometimes called <em>calibration</em>: the alignment between expressed
confidence and actual accuracy. A well-calibrated model that is 80%
confident in an answer is right about 80% of the time. A model that is
confident about answers it cannot possibly know from its training data
is miscalibrated. The car wash video is a calibration failure as much
as a grounding failure.</p>
<hr>
<h2 id="what-grounding-means">What Grounding Means</h2>
<p>The term <em>grounding</em> in AI has a precise origin. Harnad (1990) used it
to describe the problem of connecting symbol systems to the things
they refer to — how does the word &ldquo;apple&rdquo; connect to actual apples,
rather than just to other symbols? A symbol system that only connects
symbols to other symbols (dictionary definitions, synonym relations)
has the form of meaning without the substance.</p>
<p>Applied to language models: the model has rich internal representations
of concepts like &ldquo;rain,&rdquo; &ldquo;car wash,&rdquo; &ldquo;dirty car,&rdquo; and their relationships.
But those representations are grounded in text about those things, not in
the things themselves. The model knows what rain is. It does not know
whether it is raining right now, because &ldquo;right now&rdquo; is not a location
in the training data.</p>
<p>This is not a solvable problem by making the model bigger or training it
on more text. More text does not give the model access to the current
state of the world. It is a structural feature of how these systems work:
they are trained on a static corpus and queried at inference time, with
no automatic connection to the world state at the moment of the query.</p>
<hr>
<h2 id="what-tool-use-gets-you-and-what-it-doesnt">What Tool Use Gets You (and What It Doesn&rsquo;t)</h2>
<p>The standard engineering response to grounding problems is tool use:
give the model access to a weather API, a calendar, a search engine.
Now when asked &ldquo;should I go to the car wash today?&rdquo; the model can query
the weather service, get the current conditions, and factor that into
the answer.</p>
<p>This is genuinely useful. The model with a weather tool call will answer
this question correctly in most circumstances. But tool use solves the
problem only if two conditions hold:</p>
<ol>
<li>
<p><strong>The model knows it needs the tool.</strong> It must recognise that this
question has $I(A; C \mid X) > 0$ for context $C$ that a weather
tool can provide, and that it is missing that context. This requires
the meta-level awareness described above. Models trained on tool use
learn to invoke tools for recognised categories of question; for novel
question types, or questions that superficially resemble answerable
ones, the tool call may not be triggered.</p>
</li>
<li>
<p><strong>The right tool exists and returns clean data.</strong> Weather APIs exist.
&ldquo;How dirty is my car?&rdquo; does not have an API. &ldquo;Am I the kind of person
who cares about car cleanliness enough that this matters?&rdquo; has no API.
Some missing context can be retrieved; some is inherently private to
the person asking.</p>
</li>
</ol>
<p>The deeper issue is not tool availability but <em>knowing what you don&rsquo;t
know</em>. A model that does not recognise its own information gaps cannot
reliably decide when to use a tool, ask a clarifying question, or
express uncertainty. This is a hard problem — arguably harder than
making the model more capable at the tasks it already handles.</p>
<hr>
<h2 id="the-contrast-stated-plainly">The Contrast, Stated Plainly</h2>
<p>The strawberry failure and the car wash failure look alike from the
outside — confident wrong answer — but they are different enough that
conflating them produces confused diagnosis and confused solutions.</p>
<p>Strawberry: the model has the information (the string &ldquo;strawberry&rdquo;), but
the representation (BPE tokens) does not preserve character-level
structure. The fix is architectural or procedural: character-level
tokenisation, chain-of-thought letter spelling.</p>
<p>Car wash: the model does not have the information (current weather,
car state). No fix to the model&rsquo;s architecture or prompt engineering
gives it information it was never given. The fix is exogenous: provide
the context explicitly, or give the model a tool that can retrieve it,
or design the system so that context-dependent questions are routed to
systems that have access to the relevant state.</p>
<p>A model that confidently answers the car wash question without access to
current conditions is not failing at language understanding. It is
behaving exactly as its training shaped it to behave, given its lack of
situational grounding. Knowing which kind of failure you are looking at
is most of the work in figuring out what to do about it.</p>
<hr>
<p><em>The grounding problem connects to the broader question of what it means
for a language model to &ldquo;know&rdquo; something — which comes up in a different
form in the <a href="/posts/more-context-not-always-better/">context window post</a>,
where the issue is not missing context but irrelevant context drowning
out the relevant signal.</em></p>
<p><em>A second car wash video a few weeks later produced a third, different
failure: <a href="/posts/car-wash-walk/">Car Wash, Part Three: The AI Said Walk</a> —
the model had the right world state but chose the wrong interpretation
of the question.</em></p>
<hr>
<h2 id="references">References</h2>
<ul>
<li>
<p>Harnad, S. (1990). The symbol grounding problem. <em>Physica D:
Nonlinear Phenomena</em>, 42(1–3), 335–346.
<a href="https://doi.org/10.1016/0167-2789(90)90087-6">https://doi.org/10.1016/0167-2789(90)90087-6</a></p>
</li>
<li>
<p>Guo, C., Pleiss, G., Sun, Y., &amp; Weinberger, K. Q. (2017). <strong>On
calibration of modern neural networks.</strong> <em>ICML 2017</em>.
<a href="https://arxiv.org/abs/1706.04599">https://arxiv.org/abs/1706.04599</a></p>
</li>
</ul>
]]></content:encoded>
    </item>
    <item>
      <title>Inner Echo: On Making Mental Illness Visible, and What That Even Means</title>
      <link>https://sebastianspicker.github.io/posts/inner-echo/</link>
      <pubDate>Thu, 28 Nov 2024 00:00:00 +0000</pubDate>
      <guid>https://sebastianspicker.github.io/posts/inner-echo/</guid>
      <description>I am on the spectrum. Code is easy; emotions are not. This post is about the phrase &amp;lsquo;making mental illness visible&amp;rsquo;, what science actually tells us about that goal, why a non-affected person fundamentally cannot understand — and why trying still matters.</description>
      <content:encoded><![CDATA[<p>There is a phrase that appears in every mental health awareness campaign, every destigmatisation effort, every well-meaning poster in a university corridor: <em>make it visible</em>. Shine a light. Break the silence. Reduce stigma by talking about it.</p>
<p>I agree with the impulse. I am less sure about what the phrase actually asks of us, or what it assumes is possible. This post is my attempt to think through that question — and to document a small project that emerged from it.</p>
<h2 id="a-personal-starting-point">A Personal Starting Point</h2>
<p>I am on the spectrum. I was diagnosed in adulthood, which is not unusual, and the diagnosis explained a great deal about a life spent finding some things effortless and others bewildering.</p>
<p>Code is easy. The internal structure of a problem, the satisfaction of a clean abstraction, the deep rabbit holes that open when a concept catches my attention and refuses to let go — that is the natural medium. Hyperfocus is not a metaphor for me; it is literally how I spend a Tuesday afternoon. I have written entire systems because I could not stop.</p>
<p>Emotions are harder. Not absent — that is a misconception I will address in a moment — but differently structured. Reading a room is work. Social cues that seem to operate as obvious background noise for most people arrive for me as data that requires conscious decoding. The reverse appears to be true for most neurotypical people: emotional processing runs in the background, effortlessly; formal abstraction requires deliberate effort.</p>
<p>Neither is better. They are different cognitive architectures, and both come with costs.</p>
<p>I raise this not to centre myself, but because it is relevant to the question the post is actually about. I spent years navigating a social world that was not built for how I process it. That experience sits close to the experience of people with mental illness — not the same, but adjacent. And it made me think hard about what &ldquo;understanding&rdquo; across neurological difference actually means.</p>
<h2 id="mental-illness-is-still-a-grey-zone">Mental Illness Is Still a Grey Zone</h2>
<p>The progress on mental health stigma over the past decade is real. People talk about therapy more openly than they did. Burnout is acknowledged at work. The language of mental health has entered mainstream use — sometimes usefully, sometimes in ways that dilute clinical concepts into lifestyle descriptors. Anxiety is now a brand attribute. Trauma is a metaphor for mild inconvenience. This is a problem, but it is a second-order problem; the first-order problem — that serious mental illness is still heavily stigmatised, underfunded, and misunderstood — is the one that matters more.</p>
<p>Corrigan and Watson <a href="#ref-1">[1]</a> documented what the stigma research consistently shows: people with mental illness face two compounding problems. The first is public stigma <a href="#ref-3">[3]</a> — the prejudice of others, leading to discrimination in employment, housing, relationships. The second is self-stigma — the internalised application of those same prejudices to oneself. The second is often worse. It is the mechanism by which stigma becomes a barrier to seeking help, creating the feedback loop that keeps serious mental illness invisible precisely because the people experiencing it have been taught that it is shameful.</p>
<p>The phrase &ldquo;make it visible&rdquo; is a response to this dynamic. If mental illness is visible — discussed, depicted, normalised — the argument goes that stigma decreases. There is evidence for this. Contact-based interventions, where people without mental illness interact with people who have it, consistently outperform education-only approaches <a href="#ref-2">[2]</a>. The visibility of real people matters more than information campaigns.</p>
<p>But there is a difference between visibility and understanding.</p>
<h2 id="what-visibility-actually-achieves">What Visibility Actually Achieves</h2>
<p>When we say &ldquo;make it visible&rdquo;, we usually mean one of several different things, which are worth separating.</p>
<p><strong>Normalisation</strong> means that a condition becomes part of accepted human variation rather than a mark of failure or danger. This is achievable through visibility and is genuinely important. Knowing that a colleague takes antidepressants, or that a public figure manages bipolar disorder, reduces the sense of aberration. It does not require the observer to understand the experience — only to register that it exists and is survivable.</p>
<p><strong>Representation</strong> means that people with a condition see themselves reflected in culture, media, and institutions. This matters for the affected person; it is about recognition, not about inducing empathy in the non-affected.</p>
<p><strong>Empathy</strong> is the hardest and most frequently over-promised goal. It is what the simulation approaches aim for: put a neurotypical person in a room with distorted audio and flickering visuals and tell them this is what psychosis sounds like. Does it work?</p>
<p>The honest answer from the research is: somewhat, temporarily, and with significant caveats.</p>
<h2 id="the-empathy-gap">The Empathy Gap</h2>
<p>Let me be direct about something. A person who has never experienced severe depression cannot know what it is. Not in the way that a person who has experienced it knows it. This is not a failure of empathy or imagination; it is a structural fact about how knowledge of mental states works.</p>
<p>Philosophers call this the problem of other minds. We have no direct access to another person&rsquo;s experience. We infer it, imperfectly, by analogy to our own. For experiences that have no analogue in our own history, inference breaks down. You can read every clinical description of dissociation ever written and still not know what dissociation is, because the knowledge that matters is not propositional — it is not a set of facts — but experiential.</p>
<p>This is the gap that simulation approaches try to bridge, and it is genuinely unbridgeable. What simulation can do is something weaker but not worthless: it can create an affective response, a discomfort, a disruption of the observer&rsquo;s normal processing, that functions as a rough proxy signal. Not &ldquo;now you know what it is like&rdquo;, but &ldquo;now you have a small, incomplete, distorted approximation of some dimension of the experience&rdquo;.</p>
<p>The risk is misrepresentation. Schizophrenia simulations have been criticised — fairly — for reducing a complex condition to its most dramatic phenomenological features (auditory hallucinations, paranoia) while omitting the cognitive, relational, and longitudinal aspects that define how people actually live with the condition. A five-minute visual experience of &ldquo;what depression feels like&rdquo; that emphasises darkness and slow motion tells you almost nothing about the specific exhaustion of getting through a Tuesday morning, or the way time warps over months.</p>
<p>So: you cannot truly understand what you have not experienced. But you can try to approximate something, and approximation, done honestly and with appropriate epistemic humility, is better than nothing.</p>
<h2 id="metaphor-as-a-communication-tool">Metaphor as a Communication Tool</h2>
<p>There is a long tradition of using metaphor and art to communicate internal states that resist direct description. This is not a bug; it is a feature of how language handles subjective experience.</p>
<p>The poet uses metaphor because &ldquo;my heart is heavy&rdquo; is not literally true but captures something that &ldquo;I am experiencing low mood&rdquo; does not. The musician uses dissonance and rhythm to structure emotional experience in the listener. The visual artist uses colour and texture to evoke states rather than depict them. None of these are representations in the scientific sense — they do not accurately model the referent — but they create a kind of resonance that purely descriptive language cannot.</p>
<p>Mental health communication has increasingly moved in this direction. The vocabulary of &ldquo;emotional weight&rdquo;, &ldquo;spiralling&rdquo;, &ldquo;crashing&rdquo;, &ldquo;the fog&rdquo; — these are metaphors that have become clinical shorthand precisely because they communicate something essential that clinical terms do not. When someone says &ldquo;I couldn&rsquo;t get out of bed&rdquo;, they are not describing paralysis; they are describing a particular quality of anhedonia and executive dysfunction that no diagnostic manual entry captures as well.</p>
<p>This is the space where a project like inner-echo operates.</p>
<h2 id="inner-echo-the-idea">Inner Echo: The Idea</h2>
<p><a href="https://github.com/sebastianspicker/inner-echo">inner-echo</a> is a browser-based audiovisual experiment. It takes a webcam feed and applies condition-specific visual and audio effects that function as metaphorical overlays on the user&rsquo;s own image. The output is not a simulation of a mental health condition in any clinical sense. It is an attempt to construct a visual and auditory language for internal states, using the user&rsquo;s own presence as the anchor.</p>
<p>The technical architecture is deliberately minimal: React, WebGL/Canvas for video processing, optional WebAudio. Everything runs in the browser, client-side, with no backend. No data leaves the device. This is not incidental — privacy is load-bearing for a project that deals with sensitive self-reflection. Safe Mode and an emergency stop function are built in.</p>
<p>The condition-profile system supports three modes:</p>
<ul>
<li><strong>Preset mode</strong>: a single-condition metaphorical composition — one set of effects mapped to one cluster of experiences</li>
<li><strong>Multimorbid mode</strong>: weighted stacking of multiple condition profiles, acknowledging that most people with mental health conditions do not have one thing</li>
<li><strong>Symptom-first mode</strong>: dimension-level control, letting the user build from individual symptom representations rather than diagnostic labels</li>
</ul>
<p>The last of these is, I think, the most honest design choice. Diagnostic categories are administrative conveniences as much as they are natural kinds. Two people with the same diagnosis can have radically different experiences. Structuring the system around dimensions of experience rather than labels is both clinically more accurate and communicatively more flexible.</p>
<h2 id="what-it-is-not">What It Is Not</h2>
<p>Being clear about limitations is not false modesty; it is the only way this kind of project retains its integrity.</p>
<p>inner-echo is not a simulation of any condition in the sense of accurately modelling its phenomenology. It does not claim to show you &ldquo;what depression is like&rdquo;. It offers metaphorical approximations of some dimensions of some experiences, and it does so using effects that are legible to the observer — visual distortion, audio modification, altered feedback — that bear a designed but non-literal relationship to the internal states they are meant to evoke.</p>
<p>It is not a diagnostic tool. It is not a therapeutic intervention. It is not a substitute for any clinical process.</p>
<p>What it might be is a starting point for a conversation. Something a person experiencing a condition could use to gesture toward an aspect of their experience. Something a person without that experience could encounter with enough curiosity to ask a better question than they would have otherwise.</p>
<p>That is a modest claim. I think modest claims are appropriate here.</p>
<h2 id="why-this-why-now">Why This, Why Now</h2>
<p>Mental health awareness has become a genre. The awareness campaigns, the celebrity disclosures, the workplace wellness programmes — these are real goods, and I do not want to be cynical about them. But the communication problem has not been solved. The words exist. The willingness to use them, in many contexts, exists. What is still missing is a language for the texture of experience that the words point to but do not reach.</p>
<p>I find myself better able to build something than to explain it in words. That is probably a spectrum thing. inner-echo is an attempt to build toward a language that I do not fully have — for my own internal experience, and for the experiences of people navigating conditions quite different from mine.</p>
<p>The gap cannot be closed. But the attempt to reach across it is worth making, and worth being honest about.</p>
<hr>
<h2 id="references">References</h2>
<p><span id="ref-1"></span>[1] Corrigan, P.W. &amp; Watson, A.C. (2002). Understanding the impact of stigma on people with mental illness. <em>World Psychiatry</em>, 1(1), 16–20.</p>
<p><span id="ref-2"></span>[2] Corrigan, P.W., Morris, S.B., Michaels, P.J., Rafacz, J.D. &amp; Rüsch, N. (2012). Challenging the public stigma of mental illness: A meta-analysis of outcome studies. <em>Psychiatric Services</em>, 63(10), 963–973.</p>
<p><span id="ref-3"></span>[3] Goffman, E. (1963). <em>Stigma: Notes on the Management of Spoiled Identity</em>. Prentice-Hall.</p>
<p>inner-echo repository: <a href="https://github.com/sebastianspicker/inner-echo">https://github.com/sebastianspicker/inner-echo</a></p>
]]></content:encoded>
    </item>
    <item>
      <title>Three Rs in Strawberry: What the Viral Counting Test Actually Reveals</title>
      <link>https://sebastianspicker.github.io/posts/strawberry-tokenisation/</link>
      <pubDate>Mon, 07 Oct 2024 00:00:00 +0000</pubDate>
      <guid>https://sebastianspicker.github.io/posts/strawberry-tokenisation/</guid>
      <description>In September 2024, OpenAI revealed that its new o1 model had been code-named &amp;ldquo;Strawberry&amp;rdquo; internally — the same word that language models have famously been unable to count letters in. The irony was too perfect to pass up. But the counting failure is not a sign that LLMs are naive or broken. It is a precise, informative symptom of how they process text. Here is the actual explanation, with a minimum of hand-waving.</description>
      <content:encoded><![CDATA[<h2 id="the-setup">The Setup</h2>
<p>In September 2024, OpenAI publicly confirmed that their new reasoning model
had been code-named &ldquo;Strawberry&rdquo; during development. This landed with a
particular thud because &ldquo;how many r&rsquo;s are in strawberry?&rdquo; had, by that
point, become one of the canonical demonstrations of language model failure.
The model named after strawberry could not count the letters in strawberry.
The internet had opinions.</p>
<p>Before the opinions: the answer is three. s-t-<strong>r</strong>-a-w-b-e-<strong>r</strong>-<strong>r</strong>-y.
One in the <em>str-</em> cluster, two in the <em>-rry</em> ending. Count carefully and
you will find that most people get this right on the first try, and most
large language models get it wrong, returning &ldquo;two&rdquo; with apparent
confidence.</p>
<p>The question worth asking is not &ldquo;why is the model stupid.&rdquo; It is not
stupid, and &ldquo;stupid&rdquo; is not a useful category here. The question is: what
does this specific error reveal about the structure of the system?</p>
<p>The answer involves tokenisation, and it is actually interesting.</p>
<hr>
<h2 id="how-you-count-letters-and-how-the-model-doesnt">How You Count Letters (and How the Model Doesn&rsquo;t)</h2>
<p>When you count the r&rsquo;s in &ldquo;strawberry,&rdquo; you do something like this:
scan the string left to right, maintain a running count, increment it
each time you see the target character. This is a sequential operation
over a character array. It requires no semantic knowledge about the word —
it does not matter whether &ldquo;strawberry&rdquo; is a fruit, a colour, or a
nonsense string. The characters are the input; the count is the output.</p>
<p>A language model does not receive a character array. It receives a
sequence of <em>tokens</em> — chunks produced by a compression algorithm called
Byte Pair Encoding (BPE) that the model was trained with. In the
tokeniser used by GPT-class models, &ldquo;strawberry&rdquo; is most likely split as:</p>
$$\underbrace{\texttt{str}}_{\text{token 1}} \;\underbrace{\texttt{aw}}_{\text{token 2}} \;\underbrace{\texttt{berry}}_{\text{token 3}}$$<p>Three tokens. The model&rsquo;s input is these three integer IDs, each looked up
in an embedding table to produce a vector. There is no character array.
There is no letter &ldquo;r&rdquo; sitting at a known position. There are three dense
vectors representing &ldquo;str,&rdquo; &ldquo;aw,&rdquo; and &ldquo;berry.&rdquo;</p>
<hr>
<h2 id="what-bpe-does-and-doesnt-preserve">What BPE Does (and Doesn&rsquo;t) Preserve</h2>
<p>BPE is a greedy compression algorithm. Starting from individual bytes,
it iteratively merges the most frequent pair of adjacent symbols into a
single new token:</p>
$$\text{merge}(a, b) \;:\; \underbrace{a \;\; b}_{\text{separate}} \;\longrightarrow\; \underbrace{ab}_{\text{single token}}$$<p>Applied to a large text corpus until a fixed vocabulary size is reached,
this produces a vocabulary of common subwords. Frequent words and common
word-parts become single tokens; rare sequences stay as multi-token
fragments.</p>
<p>What BPE optimises for is compression efficiency, not character-level
transparency. The token &ldquo;straw&rdquo; encodes the sequence s-t-r-a-w as a
unit, but that character sequence is not explicitly represented anywhere
inside the model once the embedding lookup has occurred. The model
receives a vector for &ldquo;straw,&rdquo; not a list of its constituent letters.</p>
<p>The character composition of a token is only accessible to the model
insofar as it was implicitly learned during training — through seeing
&ldquo;straw&rdquo; appear in contexts where its internal structure was relevant.
For most tokens, most of the time, that character structure was not
relevant. The model learned what &ldquo;straw&rdquo; means, not how to spell it
character by character.</p>
<hr>
<h2 id="why-the-error-is-informative">Why the Error Is Informative</h2>
<p>Most people say the model returns &ldquo;two r&rsquo;s,&rdquo; not &ldquo;one&rdquo; or &ldquo;four&rdquo; or
&ldquo;none.&rdquo; This is not random noise. It is a systematic error, and systematic
errors are diagnostic.</p>
<p>&ldquo;berry&rdquo; contains two r&rsquo;s: b-e-<strong>r</strong>-<strong>r</strong>-y. If you ask most models
&ldquo;how many r&rsquo;s in berry?&rdquo; they get it right. The model has seen that
question, or questions closely enough related, that the right count is
encoded somewhere in the weight structure.</p>
<p>&ldquo;str&rdquo; contains one r: s-t-<strong>r</strong>. But as a token it is a short, common
prefix that appears in hundreds of words — <em>string</em>, <em>strong</em>, <em>stream</em> —
contexts in which its internal letter structure is rarely attended to.
&ldquo;aw&rdquo; contains no r&rsquo;s. When the model answers &ldquo;two,&rdquo; it is almost
certainly counting the r&rsquo;s in &ldquo;berry&rdquo; correctly and failing to notice
the one in &ldquo;str.&rdquo; The token boundaries are where the error lives.</p>
<p>This is not stupidity. It is a precise failure mode that follows directly
from the tokenisation structure. You can predict where the error will
occur by looking at the token split.</p>
<hr>
<h2 id="chain-of-thought-partially-fixes-this-and-why">Chain of Thought Partially Fixes This (and Why)</h2>
<p>If you prompt the model to &ldquo;spell out the letters first, then count,&rdquo; the
error rate drops substantially. The reason is not mysterious: forcing
the model to generate a character-by-character expansion — s, t, r, a,
w, b, e, r, r, y — puts the individual characters into the context window
as separate tokens. Now the model is not working from &ldquo;straw&rdquo; and &ldquo;berry&rdquo;;
it is working from ten single-character tokens, and counting sequential
characters in a flat list is a task the model handles much better.</p>
<p>This is, in effect, making the model do manually what a human does
automatically: convert the compressed token representation back to an
enumerable character sequence before counting. The cognitive work is the
same; the scaffolding just has to be explicit.</p>
<hr>
<h2 id="the-right-frame">The Right Frame</h2>
<p>The &ldquo;how many r&rsquo;s&rdquo; test is sometimes cited as evidence that language models
don&rsquo;t &ldquo;really&rdquo; understand text, or that they are sophisticated autocomplete
engines with no genuine knowledge. These framing choices produce more heat
than light.</p>
<p>The more precise statement is this: language models were trained to predict
likely next tokens in large text corpora. That training objective produces
a system that is very good at certain tasks (semantic inference, translation,
summarisation, code generation) and systematically bad at others (character
counting, exact arithmetic, precise spatial reasoning). The system is not
doing what you are doing when you read a sentence. It is doing something
different, which happens to produce similar outputs for a very wide range
of inputs — and different outputs for a class of inputs where the
character-level structure matters.</p>
<p>&ldquo;Strawberry&rdquo; sits squarely in that class. The model is not failing to
read the word. It is succeeding at predicting what a plausible-sounding
answer looks like, based on a compressed representation that does not
preserve the information needed to get the count right. Those are not the
same thing, and the distinction is worth keeping clear.</p>
<hr>
<p><em>The tokenisation argument here is a simplified version. Real BPE
vocabularies, positional encodings, and the specific way character
information is or isn&rsquo;t preserved in embedding tables are more complicated
than this post suggests. But the core point — that the model&rsquo;s input
representation is not a character array and never was — holds.</em></p>
<p><em>A follow-up post covers a structurally different failure mode:
<a href="/posts/car-wash-grounding/">Should I Drive to the Car Wash?</a> — where
the model understood the question perfectly but lacked access to the
world state the question was about.</em></p>
<hr>
<h2 id="references">References</h2>
<ul>
<li>
<p>Gage, P. (1994). A new algorithm for data compression. <em>The C Users
Journal</em>, 12(2), 23–38.</p>
</li>
<li>
<p>Sennrich, R., Haddow, B., &amp; Birch, A. (2016). <strong>Neural machine
translation of rare words with subword units.</strong> <em>Proceedings of the
54th Annual Meeting of the Association for Computational Linguistics
(ACL 2016)</em>, 1715–1725. <a href="https://arxiv.org/abs/1508.07909">https://arxiv.org/abs/1508.07909</a></p>
</li>
</ul>
<hr>
<h2 id="changelog">Changelog</h2>
<ul>
<li><strong>2025-12-01</strong>: Corrected the tokenisation of &ldquo;strawberry&rdquo; from two tokens (<code>straw|berry</code>) to three tokens (<code>str|aw|berry</code>), matching the actual cl100k_base tokeniser used by GPT-4. The directional argument (token boundaries obscure character-level information) is unchanged; the specific analysis was updated accordingly.</li>
</ul>
]]></content:encoded>
    </item>
    <item>
      <title>Please Stop Saying the Sun Is on Fire</title>
      <link>https://sebastianspicker.github.io/posts/astronomy-misconceptions/</link>
      <pubDate>Tue, 17 Nov 2020 00:00:00 +0000</pubDate>
      <guid>https://sebastianspicker.github.io/posts/astronomy-misconceptions/</guid>
      <description>In September 2020 I gave a teacher training talk on stellar formation and the misconceptions students bring into class. The misconception list was long enough to be its own document. Here it is, with commentary. Includes: the Sun as a heat-planet, gravity that only works when things move, metals that always existed, and the obligatory complaint about quantum leaps.</description>
      <content:encoded><![CDATA[<p><em>In September 2020 Alexander Küpper and I gave a teacher training session on
stellar formation — why experiments for it are hard to design, and what
misconceptions students typically arrive with. This post is loosely based on
the misconceptions part of that talk, which turned out to generate the most
discussion.</em></p>
<hr>
<h2 id="why-misconceptions-are-not-just-wrong-answers">Why Misconceptions Are Not Just Wrong Answers</h2>
<p>Before the list, a clarification that matters pedagogically.</p>
<p>When education researchers say &ldquo;misconception,&rdquo; they do not mean a random
error or a gap in knowledge. A misconception is a stable, self-consistent
mental model that students actively use to interpret new information. It
persists not because the student hasn&rsquo;t heard the correct explanation but
because the incorrect model handles a wide range of everyday experience
reasonably well.</p>
<p>&ldquo;Fire is a thing that makes heat and light and consumes fuel&rdquo; is a perfectly
adequate mental model for everything a student encounters outside a physics
class. It explains candles, campfires, gas hobs, and car engines. The fact
that it also leads the same student to conclude that the Sun &ldquo;burns&rdquo; in the
chemical combustion sense is not a failure of intelligence — it is the
natural extension of a model that works.</p>
<p>The implication, which Bransford, Brown, and Cocking put plainly in 2000:
if you ignore what students already believe and simply present the correct
model, &ldquo;the understanding they develop can vary substantially from what the
instructor intended.&rdquo; The new information gets interpreted through the
existing model, not in place of it. You end up with students who can repeat
&ldquo;the Sun fuses hydrogen&rdquo; while still, in their mental model, imagining it as
a very large and very hot fire.</p>
<p>With that said: here is the list.</p>
<hr>
<h2 id="the-sun-is-not-a-star">The Sun Is Not a Star</h2>
<p>This one leads because it is the most structurally interesting.</p>
<p>Bailey et al. (2009), in a study of students&rsquo; pre-instructional ideas about
stars and star formation, document the following category of response: the
Sun is a special kind of astronomical body with its own distinct properties.
It is not a star. Stars are the things you see in the night sky. The Sun is
different.</p>
<p>This is not an isolated finding. Schecker et al. (2018) document the same
pattern in the German context. Students who know perfectly well that &ldquo;the
Sun is a star&rdquo; as a stated fact will nonetheless, when asked to reason about
stellar properties, implicitly exempt the Sun from those properties. Stars
are far away, they are small and faint, they are cold and distant. The Sun
is close, large, and bright. Ergo the Sun cannot really be a star, whatever
the textbook says.</p>
<p>The pedagogical consequence is that teaching stellar evolution to students
who hold this model requires first collapsing the Sun/star distinction —
otherwise everything that follows is about something unfamiliar and distant
rather than about the object eight light minutes away that we can observe
in detail.</p>
<p>A companion misconception: <strong>all stars are smaller than the Sun</strong>. This is
the inverse problem. Students who correctly classify the Sun as a star but
have only seen stars as faint points of light infer that stars must be small.
The Sun, which they know to be large, therefore cannot be a typical star.
Betelgeuse — a red supergiant with a radius approximately 700 times the
Sun&rsquo;s, which if placed at the Sun&rsquo;s position would extend past the asteroid belt — tends
to produce strong cognitive dissonance when it is first encountered.</p>
<hr>
<h2 id="the-sun-is-on-fire">The Sun Is on Fire</h2>
<p>The combustion model of stellar energy is, empirically, the most common
student conception and the hardest to dislodge.</p>
<p>From Favia et al.&rsquo;s misconception inventory, translated loosely:</p>
<ul>
<li><em>&ldquo;The Sun is made of fire.&rdquo;</em></li>
<li><em>&ldquo;Stars run on fuel: petrol or natural gas.&rdquo;</em></li>
<li><em>&ldquo;The Sun is made of molten lava.&rdquo;</em></li>
<li><em>&ldquo;The Sun is a heat-planet.&rdquo;</em></li>
</ul>
<p>Bailey et al.&rsquo;s quantitative data: when asked how stars produce light,
32% of students described chemical burning. A further 28% described unspecified
&ldquo;chemical reactions.&rdquo; Only 7% named nuclear fusion. Only 3% could both name
fusion and correctly connect it to the production of light.</p>
<p>The combustion model is coherent and consistent. It gives you a mechanism
(fuel + oxygen → heat and light), a timescale (stars eventually run out of
fuel and go dark), and a product (visible light and heat). What it cannot
handle is the scale: the Sun has been burning for 4.6 billion years and has
approximately 5 billion years of fuel remaining. Chemical combustion at the
Sun&rsquo;s luminosity would exhaust any chemically plausible fuel supply in tens
of thousands of years. This is the crack in the model that fusion fills —
not by saying &ldquo;the Sun burns differently&rdquo; but by replacing the entire energy
mechanism with one that operates at scales the combustion model cannot reach.</p>
<p>One related misconception worth noting explicitly: <strong>the Sun is hottest at
its surface</strong>. This is the intuitive model — things are hot near the fire
and cooler further away. The corona&rsquo;s temperature of a million Kelvin, far
above the photospheric 5,778 K, violates this so thoroughly that it remained
an active research problem for decades (and, in some senses, still is).
Students encountering coronal heating for the first time do not usually reject
it, but they do find it genuinely strange in a way they cannot articulate —
which is the signature of something colliding with a stable prior model.</p>
<hr>
<h2 id="gravity-only-works-when-things-move">Gravity Only Works When Things Move</h2>
<p>The gravity misconceptions documented in the research literature are worth
treating separately because they have direct consequences for understanding
stellar formation — which depends entirely on gravity acting on stationary
or slowly drifting gas clouds.</p>
<p>The relevant findings:</p>
<p><strong>Gravity requires motion</strong> (Palmer, 2001). A significant proportion of
students believe that gravity only acts on objects that are in motion. A
stationary object is not subject to gravitational attraction. A table sitting
on the floor: fine, no gravity needed. A gas cloud drifting slowly through
space: also fine. A gas cloud being compressed by gravitational self-attraction:
this requires gravity to act on particles that are not yet moving, which the
model cannot accommodate.</p>
<p><strong>Force implies movement</strong> (Gunstone &amp; Watts, 1985). The more general version:
forces produce motion, and where there is no net motion, there is no net force.
The concept of force balance — two equal and opposite forces summing to zero
net force, with the object not moving — is not available to students holding
this model. It is hard to overstate how consequential this is for astrophysics.
Almost every stable astrophysical structure — a main-sequence star, a planetary
orbit, a galaxy&rsquo;s rotation — is a force balance. Students without the concept
cannot reason about any of them correctly.</p>
<p><strong>Gravity only acts on Earth</strong> (Bar, Brosh, and Sneider, 2016). Students in
the space context often reason that gravity is a property of Earth specifically.
In space, things are &ldquo;weightless&rdquo; — and weightlessness is interpreted as the
absence of gravity rather than as the experience of free fall in a gravitational
field. The result: gravity cannot be the mechanism by which an interstellar
gas cloud collapses, because gas clouds are in space and gravity does not work
there. Asghar and Libarkin (2010) found that only one in five non-physics
college students could correctly describe gravity as an attractive force between
masses, using the correct vocabulary.</p>
<p>These are not fringe findings. They are the majority conception at the
pre-instructional stage. Any unit on stellar formation that opens with
&ldquo;gravity compresses the gas cloud&rdquo; is speaking to students who mostly do not
believe that gravity can do that to a gas cloud in space.</p>
<hr>
<h2 id="metals-always-existed">Metals Always Existed</h2>
<p>This misconception is my personal favourite because it requires no incorrect
intuition — it requires an absence of information that most people have never
had reason to acquire.</p>
<p>Students and adults who have not encountered stellar nucleosynthesis simply
have no model for where heavy elements come from. Asked directly, a common
response is that metals &ldquo;always existed&rdquo; — they are a feature of the universe,
present from the beginning. The alternative framing: &ldquo;stars create matter from
nothing&rdquo; — which captures the sense that something is being generated, without
a mechanism.</p>
<p>The correct picture: the Big Bang produced primarily hydrogen and helium, with
trace amounts of lithium and beryllium. Every heavier element — including all
the carbon in your body, all the iron in your blood, all the oxygen in every
breath — was synthesised in a stellar interior or in a supernova. The gold in
a wedding ring was produced in a neutron star merger. We are, in the precise
sense of the phrase, made of star stuff; but not because stars are somehow
magical, because the nuclear physics of stellar interiors and violent stellar
deaths is the only process in the universe that can manufacture these elements.</p>
<p>This has a direct implication for stellar evolution education: if students
believe metals always existed, the cycle of stellar death and new star
formation — in which dying stars enrich the interstellar medium with heavy
elements that become part of the next generation of stars and their planets —
loses most of its meaning. The cycle is interesting precisely because it
explains why later-generation stars and their planets have a richer elemental
composition than first-generation stars. Remove that frame and you have
a sequence of events with no cumulative significance.</p>
<hr>
<h2 id="some-language-based-misconceptions-a-brief-digression">Some Language-Based Misconceptions (A Brief Digression)</h2>
<p>Since I promised something about quantum leaps: the phrase &ldquo;quantum leap&rdquo;
in everyday usage means a sudden, large, discontinuous advance. In physics,
a quantum transition is the smallest possible discrete change in a system&rsquo;s
energy state. The electron moves from one energy level to another; the
photon is emitted or absorbed; the scale of change is on the order of
electron-volts. It is, emphatically, not large.</p>
<p>The astronomy version of this class of error:</p>
<p><strong>&ldquo;Light year&rdquo; used as a unit of time.</strong> &ldquo;That happened light years ago.&rdquo;
A light year is the distance light travels in one year — approximately
9.46 × 10¹² kilometres. It is a unit of distance, not time. This one is
so embedded in everyday usage that correcting it usually produces mild
annoyance rather than reconsideration.</p>
<p><strong>&ldquo;Shooting stars.&rdquo;</strong> Meteors — small rocky or metallic bodies entering
the atmosphere — have nothing to do with stars. They are typically the
size of a grain of sand to a pebble. The visual resemblance to a moving
point of light crossing the sky is where the name comes from; the
resemblance to stellar physics is zero.</p>
<p><strong>&ldquo;Black holes suck things in.&rdquo;</strong> Black holes do not have more gravity
than the object that formed them at the same distance. If the Sun were
replaced by a black hole of equal mass, the planets would continue on
their current orbits. A black hole is only a black hole within its
Schwarzschild radius; beyond that it is a gravitational field like any
other. What black holes have is a point of no return — the event horizon —
beyond which escape velocity exceeds the speed of light. They do not
actively pull. They are very massive objects that objects can fall into.</p>
<p><strong>&ldquo;The dark side of the Moon.&rdquo;</strong> The Moon has a far side (permanently
facing away from Earth, due to tidal locking) and a near side. Both sides
receive approximately equal sunlight over the lunar cycle. The far side
is not permanently dark; it has a day and a night like the near side.
&ldquo;Dark side&rdquo; persists in common usage because Pink Floyd used it as an
album title and nobody wanted to call it &ldquo;The Far Side of the Moon.&rdquo;
(Although Douglas Adams would have had something to say about that.)</p>
<hr>
<h2 id="why-this-list-matters-for-teaching">Why This List Matters for Teaching</h2>
<p>The misconceptions described above are not randomly distributed. They cluster
around three areas where intuitive extrapolation from everyday experience
leads systematically away from the correct physics:</p>
<ol>
<li>
<p><strong>Scale</strong>: human intuition was not built for 150 million kilometres,
let alone 4.6 billion years or the 9.46 × 10¹² km in a light year.
The Sun cannot be fire because fire cannot last 4.6 billion years;
but &ldquo;4.6 billion years&rdquo; is not a number that everyday experience makes
graspable.</p>
</li>
<li>
<p><strong>Energy mechanism</strong>: combustion is the dominant frame for &ldquo;things that
produce heat and light.&rdquo; Nuclear fusion is not part of everyday experience
at any scale. The conceptual distance between them is not factual but
mechanistic — it requires replacing an entire causal model.</p>
</li>
<li>
<p><strong>Gravity</strong>: our direct experience of gravity is of a downward force,
active at Earth&rsquo;s surface, which keeps things from floating away.
The idea of gravity as a universal mutual attraction between all masses
— active in empty space, responsible for cloud collapse and galaxy formation
— is a substantive generalisation that everyday experience does not motivate.</p>
</li>
</ol>
<p>The pedagogical literature&rsquo;s recommendation is not to avoid these topics
but to surface the prior models explicitly before presenting the correct
physics. If you ask students &ldquo;where does the Sun&rsquo;s energy come from?&rdquo; before
you teach nuclear fusion, you learn what they believe and you create the
cognitive conditions for productive conceptual conflict. If you simply present
the fusion model without that step, students add &ldquo;fusion&rdquo; to their vocabulary
while retaining &ldquo;fire&rdquo; in their mental model.</p>
<p>The experiments Alexander Küpper and I have been developing through the
astro-lab project — described in the <a href="/posts/stellar-evolution-diy/">stellar evolution post</a>
and the <a href="/posts/astro-lab-at-home/">astro-lab@home post</a> — are designed
with these specific misconceptions in mind. The net-based gravity experiment
addresses the &ldquo;gravity doesn&rsquo;t work in space&rdquo; and &ldquo;force requires motion&rdquo;
problems directly, by making gravitational attraction between all particles
visible as a material structure. The pressure-temperature experiment makes
the &ldquo;compression heats the gas&rdquo; step concrete before any mention of fusion.</p>
<p>These are not complete solutions to deeply held misconceptions. But they are
a start at building the conceptual scaffolding that makes &ldquo;and then fusion
begins&rdquo; something other than an assertion to be memorised and filed away
without understanding.</p>
<hr>
<h2 id="references">References</h2>
<p>Asghar, A. A., &amp; Libarkin, J. C. (2010). Gravity, magnetism, and &ldquo;down&rdquo;:
Non-physics college students&rsquo; conceptions of gravity. <em>The Science Educator</em>,
19(1), 42–55.</p>
<p>Bailey, J. M., Prather, E. E., Johnson, B., &amp; Slater, T. F. (2009). College
students&rsquo; preinstructional ideas about stars and star formation.
<em>Astronomy Education Review</em>, 8(1).
<a href="https://doi.org/10.3847/AER2009038">https://doi.org/10.3847/AER2009038</a></p>
<p>Bar, V., Brosh, Y., &amp; Sneider, C. (2016). Weight, mass, and gravity:
Threshold concepts in learning science. <em>Science Educator</em>, 25(1), 22–34.</p>
<p>Bransford, J. D., Brown, A. L., &amp; Cocking, R. R. (Eds.) (2000). <em>How People
Learn: Brain, Mind, Experience, and School.</em> National Academy Press.</p>
<p>Favia, A., Comins, N. F., &amp; Thorpe, G. L. (2013). The elements of item
response theory and its framework in analyzing introductory astronomy college
student misconceptions. I. Galaxies. <em>Astronomy Education Review</em>.</p>
<p>Gunstone, R., &amp; Watts, M. (1985). Force and motion. In R. Driver, E. Guesne,
&amp; A. Tiberghien (Eds.), <em>Children&rsquo;s Ideas in Science</em> (pp. 85–104).
Open University Press.</p>
<p>Palmer, D. (2001). Students&rsquo; alternative conceptions and scientifically
acceptable conceptions about gravity. <em>International Journal of Science
Education</em>, 23(7), 691–706.
<a href="https://doi.org/10.1080/09500690010006527">https://doi.org/10.1080/09500690010006527</a></p>
<p>Schecker, H., Wilhelm, T., Hopf, M., &amp; Duit, R. (Eds.) (2018).
<em>Schülervorstellungen und Physikunterricht.</em> Springer.</p>
<hr>
<h2 id="changelog">Changelog</h2>
<ul>
<li><strong>2025-09-14</strong>: Updated the DOI for Bailey et al. (2009) to the correct 10.3847/AER2009038.</li>
<li><strong>2025-09-14</strong>: Changed &ldquo;would extend past Mars&rdquo; to &ldquo;would extend past the asteroid belt&rdquo; for Betelgeuse at ~700 R☉. At ~3.26 AU, Betelgeuse&rsquo;s radius exceeds Mars&rsquo;s orbital distance (1.52 AU) by more than a factor of two and reaches well into the asteroid belt (2.2–3.3 AU).</li>
</ul>
]]></content:encoded>
    </item>
  </channel>
</rss>
