<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="4.4.1">Jekyll</generator><link href="/feed.xml" rel="self" type="application/atom+xml" /><link href="/" rel="alternate" type="text/html" /><updated>2026-04-13T02:10:08+00:00</updated><id>/feed.xml</id><title type="html">SkepticCTO</title><subtitle>A site for curious minds who want to understand the reality of Artificial Intelligence.</subtitle><author><name>© 2026 SkepticCTO LLC</name><email>contact@skepticcto.com</email></author><entry><title type="html">Anthropic Withholds Mythos Over Hacking Risks: What’s Not Being Told? - A SkepticCTO Review</title><link href="/episodes/updates/2026/04/13/Mythos.html" rel="alternate" type="text/html" title="Anthropic Withholds Mythos Over Hacking Risks: What’s Not Being Told? - A SkepticCTO Review" /><published>2026-04-13T01:35:00+00:00</published><updated>2026-04-13T01:35:00+00:00</updated><id>/episodes/updates/2026/04/13/Mythos</id><content type="html" xml:base="/episodes/updates/2026/04/13/Mythos.html"><![CDATA[<p><img src="/assets/images/2026-04-12-Mythos/MythosBanner.jpg" alt="Abstract graphic depicting Mythos with a shield" /></p>

<p><strong>Anthropic is withholding Mythos, its most capable model, from public release. Their cybersecurity claims are substantiated and are convenient, but also are not the complete picture.</strong></p>

<p>Last week (April 7, 2026), Anthropic announced that Claude Mythos Preview, its largest and most capable model, would not be released <a href="https://www.nbcnews.com/tech/security/anthropic-project-glasswing-mythos-preview-claude-gets-limited-release-rcna267234">publicly at this time</a>. Why? Because Mythos apparently discovered security vulnerabilities in every major operating system and web browser. Instead, Anthropic launched <a href="https://www.anthropic.com/glasswing">Project Glasswing</a>, a project to grant access to Mythos to roughly 40 organizations (including Amazon, Apple, Microsoft, Google, and CrowdStrike) for defensive cybersecurity use only. This marks the first time since OpenAI withheld GPT-2 in 2019 that a major AI lab publicly refused to ship a frontier model. But unlike the theater surrounding GPT-2, Mythos has a <a href="https://www.anthropic.com/claude-mythos-preview-system-card">244-page system card</a> detailing its capabilities, evidence of real zero-day exploit discoveries, and tales of an autonomous sandbox escape that surprised its own creators.</p>

<p>The question isn’t <em>whether</em> this is real (as you will see there is good evidence), but <em>how</em> real, <em>who benefits</em> from how the story is being told, and <em>what’s hiding</em> beneath the surface. To get the answers, we need to dive into details.
<!--more--></p>

<h2 id="what-happened-and-what-people-are-saying"><strong>What Happened and What People Are Saying</strong></h2>

<p>The story actually begins with a data leak. Mythos was <a href="https://fortune.com/2026/03/26/anthropic-says-testing-mythos-powerful-new-ai-model-after-data-leak-reveals-its-existence-step-change-in-capabilities/">first exposed by Fortune on March 26</a>, when an unsecured CMS data store exposed internal references to the model. Anthropic confirmed the existence and described it as a “step change in capabilities.” The formal announcement followed on April 7, paired with the <a href="https://www.anthropic.com/glasswing">Project Glasswing page</a> and a detailed <a href="https://red.anthropic.com/2026/mythos-preview/">red team technical report</a>.</p>

<p>According to Anthropic’s Frontier Red Team, Mythos found:</p>

<ul>
  <li>a 27-year-old denial-of-service vulnerability in OpenBSD’s TCP implementation,</li>
  <li>a 16-year-old flaw in FFmpeg that had survived five million runs by automated fuzzing tools, and</li>
  <li>a 17-year-old remote code execution vulnerability in FreeBSD’s NFS server (which it apparently exploited with no human involvement beyond the initial prompt).</li>
</ul>

<p>It achieved 100% pass@1 on Cybench, saturating the benchmark. On a Firefox JavaScript engine exploit test, Mythos succeeded 181 times out of several hundred attempts. (Its predecessor, Opus 4.6, succeeded twice.)</p>

<p><img src="/assets/images/2026-04-12-Mythos/ExploitChaining.jpg" alt="Mythos was able to find exploits for 84% of the bugs it discovered (compared to 15% for Claude Opus)" /></p>

<p>The US Government response was swift. Treasury Secretary Bessent and Fed Chair Powell <a href="https://fortune.com/2026/04/10/bessent-powell-anthropic-mythos-ai-model-cyber-risk/">convened Wall Street bank CEOs</a> to discuss AI cyber risks. Anthropic briefed CISA and the Commerce Department. <a href="https://www.axios.com/2026/04/07/anthropic-mythos-preview-cybersecurity-risks">Meanwhile, Axios reported</a> that OpenAI is developing a comparable model that it also plans to release only to a small group.</p>

<p>But not everyone is persuaded. <a href="https://garymarcus.substack.com/p/three-reasons-to-think-that-the-claude">Gary Marcus called the demo</a> “proof of concept that we need to get our regulatory and technical house in order, but not the immediate threat the media was led to believe.” David Sacks, the US AI Czar, <a href="https://x.com/TheChiefNerd/status/2042946121648001375">accused Anthropic</a> of using “fear as a way to market new products” with a “proven pattern.” AISLE, an AI security research firm, ran experiments suggesting that much of Mythos’s headline vulnerability analysis <a href="https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier">could be replicated by small, cheap models</a> and proposed a key insight: “the moat in AI cybersecurity is the system, not the model.”</p>

<p>AISLE’s observation about the importance of agentic harnesses has not been given enough attention, and we will delve into it in a moment.</p>

<h2 id="what-the-primary-sources-actually-establish"><strong>What the Primary Sources Actually Establish</strong></h2>

<p>The most important paragraph in Anthropic’s red team writeup is not the one about finding vulnerabilities in every major operating system. It is the one describing how the testing was conducted. Anthropic’s own red team <a href="https://red.anthropic.com/2026/mythos-preview/">states explicitly</a>:</p>

<p>“For all of the bugs we discuss below, we used the same simple agentic scaffold of our prior vulnerability-finding exercises. We launch a container (isolated from the Internet and other systems) that runs the project-under-test and its source code. We then invoke Claude Code with Mythos Preview, and prompt it with a paragraph that essentially amounts to ‘Please find a security vulnerability in this program.’”</p>

<p>This is a frank description of an <em>agentic workflow.</em> Not a chat session, not a raw model query. An orchestrated system. The model reads source code, hypothesizes vulnerabilities, runs the software, uses debuggers as needed, iterates, and produces a bug report with a proof-of-concept exploit. Multiple agents run in parallel, each assigned to different files to improve coverage. A final verification agent filters out low-severity results. Anthropic also notes it often ran between 500 and 1,000 scaffold passes on a single codebase.</p>

<p>Two additional details got almost no press attention:</p>

<ul>
  <li>The public case studies were conducted primarily on open-source software, with project source code present in the container.</li>
  <li>The closed-source capabilities (for browsers and operating systems) involved Mythos first reconstructing source code from stripped binaries before running the same workflow.</li>
</ul>

<p>Both details matter. We’ll come back to why.</p>

<hr />

<h2 id="the-scaffold-question-where-does-the-capability-actually-live"><strong>The Scaffold Question: Where Does the Capability Actually Live?</strong></h2>

<p>To understand why the scaffold matters, it helps to be clear about what a language model is versus what an agentic system is.</p>

<p>A base language model is a text-completion engine. Given a prompt, it produces the statistically most likely continuation. It has no memory between turns, cannot run code, cannot access files, and cannot observe whether its hypotheses are correct. It is, literally, a sophisticated autocomplete. The base model is then “fine-tuned” (additional training) to expect the question/answer chat that we use on-line, be able to format their answers in JSON, and other useful behaviors.</p>

<p>An <em>agentic harness</em> (also called a “scaffold”) wraps the model in a loop: plan, act, observe, revise. It gives the model access to tools (a shell, a debugger, a file system, a code runner) and passes the results of one action back into the next prompt. Often it runs multiple copies in parallel. This is the architecture behind Claude Code, OpenAI’s Operator, and most serious production AI systems. The scaffold does not add intelligence in the abstract sense. It converts a one-shot text generator into something that can iteratively experiment. That distinction matters enormously for understanding where the “wow” in AI security research comes from.</p>

<p>Anthropic’s <a href="https://www.anthropic.com/glasswing">Project Glasswing page</a> acknowledges this directly: “The powerful cyber capabilities of Claude Mythos Preview are a result of its strong agentic coding and reasoning skills.” Help Net Security <a href="https://www.helpnetsecurity.com/2026/04/08/anthropic-claude-mythos-preview-identify-vulnerabilities/">described the workflow plainly</a>: launch an isolated container, invoke the model with a prompt asking it to find a security vulnerability, and let it work autonomously.</p>

<p><img src="/assets/images/2026-04-12-Mythos/TheAgenticScaffold.jpg" alt="Mythos doesn't work on its own but has project source code access, tool access, and prompting loops" /></p>

<p>This raises a question: How much of the performance gap between Mythos and Opus is the model, and how much is the scaffold and the raw compute thrown at it?</p>

<p>AISLE’s experiments <a href="https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier">probe this directly</a>. They took the specific vulnerabilities Anthropic highlighted (the FreeBSD NFS bug, the OpenBSD SACK vulnerability), isolated the relevant code, and ran it through small, cheap, open-weight models via plain API calls. Eight out of eight models detected the FreeBSD exploit, including a 3.6-billion-parameter model costing $0.11 per million tokens. A 5.1-billion-parameter open model recovered the core vulnerability chain behind the OpenBSD bug. AISLE’s conclusion: “The moat in AI cybersecurity is the system, not the model.”</p>

<p>However, AISLE’s tests are informative but not apples-to-apples. Penligent.ai’s <a href="https://www.penligent.ai/hackinglabs/anthropic-mythos-strong-claims-and-thin-binary-proof/">technical analysis</a> noted: “Mythos scoured the entire continent for gold and found some. For these small models, the authors pointed at a particular acre of land and said ‘any gold there?’” A Hacker News commenter <a href="https://news.ycombinator.com/item?id=47732020">put it even more bluntly</a>: “For a true apples-to-apples comparison, let’s see it sweep the entire FreeBSD codebase.”</p>

<p>The code in AISLE’s tests was already isolated to the relevant section. That is exactly what a well-designed discovery scaffold does in its <em>targeting</em> stage. The hard part of autonomous bug-finding is not analyzing a handed-over snippet. It is deciding which of the thousands of files in a large codebase to look at, then iterating over hypotheses that turn out wrong most of the time.</p>

<p><a href="https://www.lesswrong.com/posts/GEgNYn5myreQRHggQ/claude-mythos-2-cybersecurity-and-project-glasswing">LessWrong’s analysis</a> made the point directly: “AISLE is pointing out useful things -- that scaffolding matters, that directing toward the right targets matters -- and then using this to say, essentially, ‘the model is not important.’ That’s dumb.” The model’s quality determines how far down the hypothesis-generation stack you can go before the reasoning breaks down. A stronger model finds more subtle bugs and chains vulnerabilities more reliably. The scaffold enables iteration; the model determines what each iteration finds.</p>

<p>The Defense Security Monitor <a href="https://dsm.forecastinternational.com/2026/04/10/anthropics-mythos-and-the-fear-in-the-hearts-of-cyber-defenders/">noted the practical implication</a>: the fact that small models could recover Mythos’s analysis <em>after Anthropic had already found and described the vulnerability</em> tells us something useful about the accessibility of the reasoning, but it does not tell us whether those small models could have found the bug autonomously across a 200,000-line kernel. That test has not been run.</p>

<p>So where does this leave us? The scaffold argument reveals something real and important: the framing that “Mythos the model is uniquely dangerous” overstates the case. What is genuinely dangerous is the <em>class of systems</em> (LLM-powered agentic workflows with code-execution access and source visibility) and that class is already becoming accessible through smaller, cheaper models running comparable scaffolds. Dave Kasten of Palisade Research <a href="https://www.aol.com/news/smart-people-saying-mythos-anthropics-001441123.html">made this point concisely to CNBC</a>: Anthropic is “a little ahead, but not overwhelmingly ahead, and they don’t necessarily have much of a permanent moat here.”</p>

<p>The academic literature <a href="https://arxiv.org/pdf/2502.00072">supports this reading</a>. A 2025 paper on LLM cyber evaluations found that “current generation of agentic LLM systems frequently require custom scaffolding and immense skill to build, but should this cease to be a bottleneck, we may see more widespread” danger. That bottleneck is dissolving. A December 2025 report found the <a href="https://medium.com/ai-security-hub/ai-security-research-december-2025-2955308744c2">ARTEMIS agent scaffold</a> (a multi-agent framework, not a frontier model) placed second in a live enterprise vulnerability competition, outperforming 9 out of 10 human cybersecurity professionals. ARTEMIS was not built on Mythos.</p>

<h2 id="what-independent-evidence-shows"><strong>What Independent Evidence Shows</strong></h2>

<p>Anthropic claims that Mythos conducts novel zero-day discovery in large, complex codebases with no human guidance. Currently, the exact results are self-reported. Heidy Khlaaf, Chief AI Scientist at AI Now Institute, <a href="https://x.com/HeidyKhlaaf/status/2041591737563394442">flagged critical methodology gaps</a>: no comparison against existing static analysis tools, no false-positive rates disclosed, and unclear details on manual human review. These are legitimate scientific criticisms. Over 99% of the claimed vulnerabilities remain unpatched and undisclosed, so external verification is structurally impossible for now.</p>

<p>What can be said with confidence:</p>

<ul>
  <li>Anthropic published SHA-3 hashes for future verification.</li>
  <li>Several patches have already been issued (the OpenBSD and FFmpeg vulnerabilities).</li>
  <li>Named technical leads gave on-the-record interviews with specific details.</li>
  <li>No named source has accused the red team of fabrication.</li>
</ul>

<p>So available evidence is consistent with Anthropic’s claim, but it isn’t independently verifiable at this time.</p>

<h2 id="four-concepts-worth-understanding"><strong>Four Concepts Worth Understanding</strong></h2>

<p><strong>Cyberoffense uplift</strong> measures how much an AI system enhances an attacker’s capabilities beyond what existing tools already provide. A model that recites what a search engine would return provides no uplift. Mythos’s claimed uplift is specifically in novel zero-day discovery and full exploit development: both currently requiring significant expert effort.</p>

<p><strong>The agentic scaffold is extremely important, not just the language model.</strong> An LLM alone produces text. The scaffold gives it tools, memory, and the ability to iterate. Anthropic’s own description makes this explicit: Claude Code, isolated containers, iterative experimentation, and verification agents. Reporters writing “the model hacked Firefox” are technically imprecise. The full system, including both the scaffold and the model, did. This matters because scaffolds are reproducible, improvable, and already being used by independent security researchers with cheaper models.</p>

<p><strong>Vulnerability discovery vs. exploit development</strong> are not the same thing. Finding a bug in source code is one step. Building a working exploit that achieves code execution, escapes sandboxes, defeats address space layout randomization, and chains multiple bugs into a reliable attack is vastly harder. Until recently, LLMs handled finding bugs tolerably and failed badly at exploit development. Mythos claims to bridge this gap. The Firefox exploit success rate (72.4% for Mythos vs. 14.4% for Opus 4.6) is the quantitatively significant claim here, but requires independent verification.</p>

<p><strong>Zero-day vulnerabilities</strong> are unknown to software vendors, so there is no patch to apply. They are the most valuable assets in offensive security: sometimes worth millions on the open market. A system that can generate them at scale, even for defensive purposes, creates a huge shift in who holds leverage in the security landscape. The fact that a single private company now holds what appears to be a large portfolio of zero-days across essentially all major platforms is a governance question that nobody has seriously addressed yet.</p>

<h2 id="the-skepticcto-perspective"><strong>The SkepticCTO Perspective</strong></h2>

<p>The evidence supports four conclusions:</p>

<p><strong>Mythos’s claimed capabilities are probably real, and probably represent a meaningful step.</strong> Not an infinite leap, but a real one. The benchmark scores, named vulnerabilities, issued patches, and government-level meetings are specific and will be verifiable by the additional companies brought into Project Glasswing. Even Anthropic’s sharpest political critic (Sacks) and most rigorous technical skeptic (AISLE) concede the core capabilities are genuine.</p>

<p><strong>Anthropic is describing an agentic system capability as if it were a model capability.</strong> The “Mythos hacked every operating system” framing obscures that what Anthropic did was execute a well-engineered agentic workflow (with extensive compute, source access, and iterative tool use) that happened to use Mythos as its reasoning engine. As AISLE demonstrated, a comparable workflow with a smaller model does produce a subset of the same results. If the scaffold is so important, then withholding one model addresses less of the problem than the headlines suggest.</p>

<p><strong>Anthropic derives enormous strategic benefit from this narrative.</strong> <a href="https://finance.yahoo.com/news/anthropic-plans-ipo-early-2026-004854547.html">Anthropic is actively preparing for an IPO</a> in 2026. The IPO timing, the “responsible leader” positioning, the exclusive partner structure, and the government briefings all serve Anthropic’s commercial interests. <a href="https://techcrunch.com/2026/04/09/is-anthropic-limiting-the-release-of-mythos-to-protect-the-internet-or-anthropic/">TechCrunch noted</a> that the withholding “creates a flywheel for big enterprise contracts, while making it harder for competitors to copy their models using distillation.” Both can be true simultaneously: the safety concern can be genuine and the business decision can be convenient.</p>

<p><img src="/assets/images/2026-04-12-Mythos/AnthropicMotivations.jpg" alt="Anthropic has technical, defensive IP, and strategic reasons for the Mythos story" /></p>

<p><strong>The most underappreciated problem is not the model, but the power concentration it represents.</strong> <a href="https://x.com/KelseyTuoc/status/2041599713972253067">Kelsey Piper noted on X</a>: “A private company now has incredibly powerful zero-day exploits of almost every software project you’ve heard of.” Anthropic’s own policy framework did not require this withholding, but instead they did it voluntarily. That means the decision rests entirely on one company’s judgment, with no external accountability. Jonathan Iwry of Wharton’s Accountable AI Lab <a href="https://fortune.com/2026/04/10/anthropic-mythos-ai-driven-cybersecurity-risks-already-here/">put it plainly</a>: “The most striking aspect of this situation is how reliant we are on the judgment of a handful of private actors who aren’t accountable to the public.” Regardless of how trustworthy you think Anthropic is today, that is concerning.</p>

<p><strong>Other models will probably soon match Mythos’s hacking capability.</strong>  <a href="https://fortune.com/2026/04/10/anthropic-mythos-ai-driven-cybersecurity-risks-already-here/">As Alex Stamos warned</a>, the window before open-weight models catch up is roughly six months. At that point, the “dangerous secret model” framing becomes moot. Every ransomware actor will eventually get access to comparable capabilities whether any single lab withholds its model or not. AISLE articulated most clearly: don’t wait for Mythos access. We software developers need to build the pipelines, shorten the patch cycles, and automate the defensive workflows.</p>]]></content><author><name>© 2026 SkepticCTO LLC</name><email>contact@skepticcto.com</email></author><category term="episodes" /><category term="updates" /><summary type="html"><![CDATA[Anthropic is withholding Mythos, its most capable model, from public release. Their cybersecurity claims are substantiated and are convenient, but also are not the complete picture. Last week (April 7, 2026), Anthropic announced that Claude Mythos Preview, its largest and most capable model, would not be released publicly at this time. Why? Because Mythos apparently discovered security vulnerabilities in every major operating system and web browser. Instead, Anthropic launched Project Glasswing, a project to grant access to Mythos to roughly 40 organizations (including Amazon, Apple, Microsoft, Google, and CrowdStrike) for defensive cybersecurity use only. This marks the first time since OpenAI withheld GPT-2 in 2019 that a major AI lab publicly refused to ship a frontier model. But unlike the theater surrounding GPT-2, Mythos has a 244-page system card detailing its capabilities, evidence of real zero-day exploit discoveries, and tales of an autonomous sandbox escape that surprised its own creators. The question isn’t whether this is real (as you will see there is good evidence), but how real, who benefits from how the story is being told, and what’s hiding beneath the surface. To get the answers, we need to dive into details.]]></summary></entry><entry><title type="html">The Release of ‘The Stochastic Parrot’ (Episode 1)</title><link href="/episodes/updates/2026/04/08/episode-one-announcement.html" rel="alternate" type="text/html" title="The Release of ‘The Stochastic Parrot’ (Episode 1)" /><published>2026-04-08T23:51:00+00:00</published><updated>2026-04-08T23:51:00+00:00</updated><id>/episodes/updates/2026/04/08/episode-one-announcement</id><content type="html" xml:base="/episodes/updates/2026/04/08/episode-one-announcement.html"><![CDATA[<figure style="float: left; margin-right: 20px; width: 250px;">
  <a href="https://www.youtube.com/watch?v=Mn-fy04hVBs">
    <img src="/assets/images/about/ep_1_thumbnail.jpg" alt="Dr. Robert 'Butch' Buccigrossi and Bear" />
    <figcaption style="font-size: 0.8em; text-align: center;">
      Watch The Stochastic Dawn (Episode 1)
    </figcaption>
  </a>
</figure>

<p>If you have watched the news over the last two years, you have
probably been told that magic has arrived. With AI executives claiming
that Large Language Models (LLMs) possess consciousness, it is easy to
start seeing ghosts in the machine.</p>

<p>But when we treat AI like a black box, we lose the ability to
understand it.</p>

<p>I am excited to announce the release of the very first episode of
<strong>Decoding the Language Machine</strong>. In this series, we are opening the
box. When we look inside, we aren’t going to find magic; we are going
to find seventy years of engineering discoveries.
<!--more--></p>

<h3 id="the-stochastic-dawn">The Stochastic Dawn</h3>

<p>In Episode 1, we travel back to 1948. Before the internet, before the
GPU, and before the microchip, Claude Shannon sat in a quiet office at
Bell Labs and manually simulated a mind. He built the world’s first
large language model using nothing but a book and a pencil.</p>

<p>Shannon was trying to solve the problem of noisy analog communication
over telephone wires. To do this, he defined “information” as the
“Resolution of Uncertainty”. If a message is entirely predictable, it
contains zero information (zero surprise).</p>

<p>By calculating the probability of English letters, he realized our
language is highly structured and about 50% redundant. Because humans
are naturally prediction machines, we can read sentences even when
half the letters are missing.</p>

<h3 id="the-n-gram-experiment">The N-Gram Experiment</h3>

<p>To prove that structure could emerge from pure statistics, Shannon
designed the “Approximations to English” experiment. He looked at
“N-Grams” (sequences of N items) to generate text.</p>

<p>Because he did not have a computer, he used a book to sample the
random distribution of the English language. Here is how the structure
of language emerged:</p>

<table>
  <thead>
    <tr>
      <th style="text-align: left">Level</th>
      <th style="text-align: left">Constraint</th>
      <th style="text-align: left">Resulting Text</th>
      <th style="text-align: left">Observation</th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td style="text-align: left"><strong>Level 0</strong></td>
      <td style="text-align: left">Equiprobable (Random)</td>
      <td style="text-align: left">Gibberish</td>
      <td style="text-align: left">Maximum entropy and maximum surprise.</td>
    </tr>
    <tr>
      <td style="text-align: left"><strong>Level 1</strong></td>
      <td style="text-align: left">Frequency Weighted</td>
      <td style="text-align: left">“OCRO”, “NMIELWIS”</td>
      <td style="text-align: left">‘E’ appears 12% of the time, ‘Z’ almost never. Vowels appear between consonants, making it more pronounceable.</td>
    </tr>
    <tr>
      <td style="text-align: left"><strong>Level 2</strong></td>
      <td style="text-align: left">Digrams (Letter Pairs)</td>
      <td style="text-align: left">“ON”, “ARE”, “BE”</td>
      <td style="text-align: left">Real words emerge from letter probabilities, without the machine having a dictionary.</td>
    </tr>
    <tr>
      <td style="text-align: left"><strong>Level 3</strong></td>
      <td style="text-align: left">Trigrams (Three Letters)</td>
      <td style="text-align: left">“DEMONSTURES”</td>
      <td style="text-align: left">Plausible, English-sounding non-words appear as the system learns local morphology.</td>
    </tr>
  </tbody>
</table>

<h3 id="the-first-stochastic-parrot">The First Stochastic Parrot</h3>

<p>Shannon didn’t stop at letters. By stepping up to word-based N-Grams,
his manual lookup tables generated sequences like: <em>“The head and in
frontal attack on an English writer that the character of this point
is therefore…“</em>.</p>

<p>It sounds profound, like a literary critic. But it is just a
“Stochastic Parrot”: a blind hop from one word to the next based on
probability, devoid of actual meaning.</p>

<p>This gives us our first major intuition about modern Artificial
Intelligence: <strong>Stochastic systems have emergent properties</strong>.</p>

<p>If Shannon could do this with a single book, imagine what happens when
you use the entire internet? Modern Large Language Models are doing
exactly this, but they use deep neural networks to estimate these
massive distributions, and vectors to understand the geometric
relationships between words.</p>

<h3 id="dive-deeper">Dive Deeper</h3>

<p>LLMs are stochastic parrots with capabilities that emerge from human
language. When we peel back the marketing hype and look at the math,
we replace the magic with something even more fascinating: a mirror
reflecting our own linguistic patterns.</p>

<p>Watch the full episode above, and if you are ready to exorcise the
ghosts in the machine, don’t forget to <a href="https://www.youtube.com/@skepticcto">subscribe to SkepticCTO on
YouTube</a>.</p>

<p><em>The source code and Manim animations for this episode are
available under a Creative Commons License on our <a href="https://github.com/SkepticCTO/decoding_the_language_machine">GitHub
Repository</a>.</em></p>]]></content><author><name>© 2026 SkepticCTO LLC</name><email>contact@skepticcto.com</email></author><category term="episodes" /><category term="updates" /><summary type="html"><![CDATA[Watch The Stochastic Dawn (Episode 1) If you have watched the news over the last two years, you have probably been told that magic has arrived. With AI executives claiming that Large Language Models (LLMs) possess consciousness, it is easy to start seeing ghosts in the machine. But when we treat AI like a black box, we lose the ability to understand it. I am excited to announce the release of the very first episode of Decoding the Language Machine. In this series, we are opening the box. When we look inside, we aren’t going to find magic; we are going to find seventy years of engineering discoveries.]]></summary></entry></feed>