<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>AI on Antonio Space</title>
    <link>/tags/ai/</link>
    <description>Recent content in AI on Antonio Space</description>
    <generator>Hugo -- gohugo.io</generator>
    <language>en</language>
    <lastBuildDate>Sat, 28 Mar 2026 00:00:00 +0000</lastBuildDate><atom:link href="/tags/ai/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>AI, Dopamine, and the Night I Rewired My Home Lab</title>
      <link>/posts/aidiction/</link>
      <pubDate>Sat, 28 Mar 2026 00:00:00 +0000</pubDate>
      
      <guid>/posts/aidiction/</guid>
      <description>AI, Dopamine, and the Night I Rewired My Home Lab A few nights ago, I opened my real backlog.
Not Jira. Not a curated roadmap.
The actual list:
stabilize my k3s cluster running across Raspberry Pis and an old desktop build a network sentinel agent on my OpenWrt router to detect malicious traffic improve observability (Prometheus, exporters, Grafana dashboards) monitor smart plugs for power anomalies orchestrate domotic scenarios based on real signals (not just timers) This is the kind of list that usually grows.</description>
      <content>&lt;h1 id=&#34;ai-dopamine-and-the-night-i-rewired-my-home-lab&#34;&gt;AI, Dopamine, and the Night I Rewired My Home Lab&lt;/h1&gt;
&lt;p&gt;A few nights ago, I opened my real backlog.&lt;/p&gt;
&lt;p&gt;Not Jira. Not a curated roadmap.&lt;/p&gt;
&lt;p&gt;The actual list:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;stabilize my k3s cluster running across Raspberry Pis and an old desktop&lt;/li&gt;
&lt;li&gt;build a network sentinel agent on my OpenWrt router to detect malicious traffic&lt;/li&gt;
&lt;li&gt;improve observability (Prometheus, exporters, Grafana dashboards)&lt;/li&gt;
&lt;li&gt;monitor smart plugs for power anomalies&lt;/li&gt;
&lt;li&gt;orchestrate domotic scenarios based on real signals (not just timers)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is the kind of list that usually grows.&lt;/p&gt;
&lt;p&gt;Not shrinks.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;then-i-turned-on-ai&#34;&gt;Then I turned on AI&lt;/h2&gt;
&lt;p&gt;I had AI wired directly into my environment.&lt;/p&gt;
&lt;p&gt;And instead of prioritizing…&lt;/p&gt;
&lt;p&gt;I just started executing.&lt;/p&gt;
&lt;p&gt;One task.&lt;br&gt;
Then another.&lt;br&gt;
Then another.&lt;/p&gt;
&lt;p&gt;&lt;strong&gt;Bang. Bang. Bang.&lt;/strong&gt;&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Exporters deployed&lt;/li&gt;
&lt;li&gt;Metrics flowing&lt;/li&gt;
&lt;li&gt;Alerts firing&lt;/li&gt;
&lt;li&gt;Dashboards shaping up&lt;/li&gt;
&lt;li&gt;Agents scanning traffic&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;At some point, I wasn’t implementing ideas anymore.&lt;/p&gt;
&lt;p&gt;I was &lt;em&gt;discovering systems&lt;/em&gt;.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;security-changes-when-exploration-becomes-cheap&#34;&gt;Security changes when exploration becomes cheap&lt;/h2&gt;
&lt;p&gt;Here’s what surprised me the most.&lt;/p&gt;
&lt;p&gt;As I instrumented the system:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;smart plugs started exposing power patterns&lt;/li&gt;
&lt;li&gt;network flows revealed unexpected behaviors&lt;/li&gt;
&lt;li&gt;devices that “looked fine” showed anomalies&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Not vulnerabilities in the traditional sense.&lt;/p&gt;
&lt;p&gt;But &lt;strong&gt;weak signals&lt;/strong&gt;:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;unusual consumption patterns&lt;/li&gt;
&lt;li&gt;unexpected network chatter&lt;/li&gt;
&lt;li&gt;implicit dependencies between devices&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Things I would have never had time to explore manually.&lt;/p&gt;
&lt;p&gt;And that’s when it clicked:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;With AI, security is no longer just about known threats —&lt;br&gt;
it’s about &lt;strong&gt;exploring unknown states at scale&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h2 id=&#34;the-dopamine-spike&#34;&gt;The dopamine spike&lt;/h2&gt;
&lt;p&gt;Somewhere in the middle of all this, I paused.&lt;/p&gt;
&lt;p&gt;Because I felt it.&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;hyper-focused&lt;/li&gt;
&lt;li&gt;highly motivated&lt;/li&gt;
&lt;li&gt;unable to stop&lt;/li&gt;
&lt;li&gt;constantly jumping to the next improvement&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;It felt like flow.&lt;/p&gt;
&lt;p&gt;But more intense.&lt;/p&gt;
&lt;p&gt;And I asked myself:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;Is this just productivity… or something else?&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h2 id=&#34;engineers--ai--infinite-execution-loop&#34;&gt;Engineers + AI = infinite execution loop&lt;/h2&gt;
&lt;p&gt;If you’re in infra or security, you already have the mindset:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;curiosity-driven&lt;/li&gt;
&lt;li&gt;system-oriented&lt;/li&gt;
&lt;li&gt;always seeing improvements&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Normally, execution is the bottleneck.&lt;/p&gt;
&lt;p&gt;AI removes that.&lt;/p&gt;
&lt;p&gt;Now:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;writing exporters takes minutes&lt;/li&gt;
&lt;li&gt;building agents is trivial&lt;/li&gt;
&lt;li&gt;testing hypotheses is cheap&lt;/li&gt;
&lt;li&gt;iterating is almost free&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;So what happens?&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;You stop choosing carefully — and start executing continuously.&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h2 id=&#34;the-loop-i-fell-into&#34;&gt;The loop I fell into&lt;/h2&gt;
&lt;ol&gt;
&lt;li&gt;Identify a gap (monitoring, security, automation)&lt;/li&gt;
&lt;li&gt;Build a quick solution with AI&lt;/li&gt;
&lt;li&gt;Discover new signals or anomalies&lt;/li&gt;
&lt;li&gt;Expand scope&lt;/li&gt;
&lt;li&gt;Move to the next idea&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;Repeat.&lt;/p&gt;
&lt;p&gt;This feels like peak productivity.&lt;/p&gt;
&lt;p&gt;But it’s actually:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;A &lt;strong&gt;dopamine-driven infra loop&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h2 id=&#34;why-this-is-different-from-other-dopamine-traps&#34;&gt;Why this is different from other “dopamine traps”&lt;/h2&gt;
&lt;p&gt;We’ve seen loops like this before:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;social media&lt;/li&gt;
&lt;li&gt;gaming&lt;/li&gt;
&lt;li&gt;dashboards for the sake of dashboards&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;But this is not consumption.&lt;/p&gt;
&lt;p&gt;This is:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;dopamine attached to building systems&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And that’s why it’s dangerous.&lt;/p&gt;
&lt;p&gt;Because it &lt;em&gt;looks like progress&lt;/em&gt;.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;the-real-risk-in-infra--security&#34;&gt;The real risk in infra &amp;amp; security&lt;/h2&gt;
&lt;p&gt;The risk is not addiction.&lt;/p&gt;
&lt;p&gt;The risk is:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;building systems that are wide… but not deep&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;You end up with:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;10 dashboards, none fully reliable&lt;/li&gt;
&lt;li&gt;multiple agents, none production-ready&lt;/li&gt;
&lt;li&gt;partial observability&lt;/li&gt;
&lt;li&gt;fragmented automation&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;In security, this is worse than doing nothing.&lt;/p&gt;
&lt;p&gt;Because:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;false confidence is more dangerous than no visibility&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h2 id=&#34;what-actually-worked-lesson-learned&#34;&gt;What actually worked (lesson learned)&lt;/h2&gt;
&lt;p&gt;After that session, I changed one rule:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;&lt;strong&gt;Build one thing. Then make it real.&lt;/strong&gt;&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Not:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;“good enough”&lt;/li&gt;
&lt;li&gt;not “it works on my machine”&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;But:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;observable&lt;/li&gt;
&lt;li&gt;reliable&lt;/li&gt;
&lt;li&gt;documented&lt;/li&gt;
&lt;li&gt;explainable&lt;/li&gt;
&lt;/ul&gt;
&lt;hr&gt;
&lt;h2 id=&#34;example-from-toy-to-system&#34;&gt;Example: from toy to system&lt;/h2&gt;
&lt;h3 id=&#34;-before-dopamine-mode&#34;&gt;❌ Before (dopamine mode)&lt;/h3&gt;
&lt;ul&gt;
&lt;li&gt;exporter deployed&lt;/li&gt;
&lt;li&gt;metrics visible&lt;/li&gt;
&lt;li&gt;quick dashboard&lt;/li&gt;
&lt;li&gt;move on&lt;/li&gt;
&lt;/ul&gt;
&lt;hr&gt;
&lt;h3 id=&#34;-after-controlled-mode&#34;&gt;✅ After (controlled mode)&lt;/h3&gt;
&lt;ul&gt;
&lt;li&gt;exporter deployed&lt;/li&gt;
&lt;li&gt;metrics validated (correctness &amp;gt; existence)&lt;/li&gt;
&lt;li&gt;alerts defined with real thresholds&lt;/li&gt;
&lt;li&gt;failure modes tested&lt;/li&gt;
&lt;li&gt;dashboard tied to action&lt;/li&gt;
&lt;li&gt;integrated into a workflow (not just visualization)&lt;/li&gt;
&lt;/ul&gt;
&lt;hr&gt;
&lt;h2 id=&#34;smart-plugs-a-concrete-case&#34;&gt;Smart plugs: a concrete case&lt;/h2&gt;
&lt;p&gt;Monitoring smart plugs started as a small idea.&lt;/p&gt;
&lt;p&gt;It became a system.&lt;/p&gt;
&lt;p&gt;From:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;“see power consumption in Grafana”&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;To:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;detect abnormal usage patterns&lt;/li&gt;
&lt;li&gt;correlate with device state&lt;/li&gt;
&lt;li&gt;trigger domotic actions:
&lt;ul&gt;
&lt;li&gt;shut down unstable devices&lt;/li&gt;
&lt;li&gt;alert on unexpected consumption&lt;/li&gt;
&lt;li&gt;adapt behavior based on load&lt;/li&gt;
&lt;/ul&gt;
&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;This is where infra meets automation.&lt;/p&gt;
&lt;p&gt;And where:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;observability becomes control&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h2 id=&#34;openwrt-agent-another-example&#34;&gt;OpenWrt agent: another example&lt;/h2&gt;
&lt;p&gt;The idea:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;detect malicious traffic on my home network&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;With AI, it was easy to:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;parse flows&lt;/li&gt;
&lt;li&gt;classify patterns&lt;/li&gt;
&lt;li&gt;generate alerts&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;But the real work was:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;defining what “malicious” actually means&lt;/li&gt;
&lt;li&gt;reducing false positives&lt;/li&gt;
&lt;li&gt;integrating with existing signals&lt;/li&gt;
&lt;li&gt;deciding what action to take&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;That’s the difference between:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;a demo&lt;br&gt;
and&lt;br&gt;
a security system&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h2 id=&#34;ai-is-a-cognitive-amplifier-and-a-stimulant&#34;&gt;AI is a cognitive amplifier (and a stimulant)&lt;/h2&gt;
&lt;p&gt;The closest analogy I’ve found:&lt;/p&gt;
&lt;p&gt;AI behaves like a &lt;strong&gt;cognitive stimulant&lt;/strong&gt;.&lt;/p&gt;
&lt;p&gt;Not chemically.&lt;/p&gt;
&lt;p&gt;But functionally:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;faster feedback&lt;/li&gt;
&lt;li&gt;higher engagement&lt;/li&gt;
&lt;li&gt;reduced friction&lt;/li&gt;
&lt;li&gt;continuous reward&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Which means:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;you need discipline at a different layer&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h2 id=&#34;a-better-way-to-use-ai-in-infra--security&#34;&gt;A better way to use AI in infra &amp;amp; security&lt;/h2&gt;
&lt;p&gt;Here’s the model I’m converging on.&lt;/p&gt;
&lt;hr&gt;
&lt;h3 id=&#34;1-exploration-is-allowed--but-bounded&#34;&gt;1. Exploration is allowed — but bounded&lt;/h3&gt;
&lt;p&gt;Use AI to:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;explore ideas&lt;/li&gt;
&lt;li&gt;generate prototypes&lt;/li&gt;
&lt;li&gt;map solution space&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;But set limits:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;time-box it&lt;/li&gt;
&lt;li&gt;define exit criteria&lt;/li&gt;
&lt;/ul&gt;
&lt;hr&gt;
&lt;h3 id=&#34;2-execution-is-sacred&#34;&gt;2. Execution is sacred&lt;/h3&gt;
&lt;p&gt;For anything that matters:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;validate signals&lt;/li&gt;
&lt;li&gt;test failure modes&lt;/li&gt;
&lt;li&gt;define ownership&lt;/li&gt;
&lt;li&gt;connect to action&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;If it doesn’t change behavior:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;it’s noise&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h3 id=&#34;3-observability--decision--action&#34;&gt;3. Observability → Decision → Action&lt;/h3&gt;
&lt;p&gt;Every system you build should follow:&lt;/p&gt;
&lt;ol&gt;
&lt;li&gt;Observe (metrics, logs, signals)&lt;/li&gt;
&lt;li&gt;Decide (rules, models, thresholds)&lt;/li&gt;
&lt;li&gt;Act (automation, alerts, orchestration)&lt;/li&gt;
&lt;/ol&gt;
&lt;p&gt;If you stop at step 1:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;you built a dashboard, not a system&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;h3 id=&#34;4-build-less-but-build-deeper&#34;&gt;4. Build less, but build deeper&lt;/h3&gt;
&lt;p&gt;AI makes building easy.&lt;/p&gt;
&lt;p&gt;Value still comes from:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;reliability&lt;/li&gt;
&lt;li&gt;clarity&lt;/li&gt;
&lt;li&gt;integration&lt;/li&gt;
&lt;/ul&gt;
&lt;hr&gt;
&lt;h2 id=&#34;final-thought&#34;&gt;Final thought&lt;/h2&gt;
&lt;p&gt;That night felt like unlocking a new capability.&lt;/p&gt;
&lt;p&gt;Not because I became better.&lt;/p&gt;
&lt;p&gt;But because:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;the cost of turning ideas into systems collapsed&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;And when that happens, a new problem appears:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;choosing what &lt;em&gt;not&lt;/em&gt; to build&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;If you’re in infra or security, this matters.&lt;/p&gt;
&lt;p&gt;Because the goal is not to build more systems.&lt;/p&gt;
&lt;p&gt;It’s to build:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;fewer systems that you can actually trust&lt;/p&gt;
&lt;/blockquote&gt;
&lt;hr&gt;
&lt;p&gt;&lt;strong&gt;Build fast.&lt;br&gt;
But finish one.&lt;/strong&gt;&lt;/p&gt;
</content>
    </item>
    
    <item>
      <title>LLMs Are Math</title>
      <link>/posts/llm-math/</link>
      <pubDate>Sun, 22 Feb 2026 14:00:00 +0800</pubDate>
      
      <guid>/posts/llm-math/</guid>
      <description>“AI feels magical, until you realize it’s mostly linear algebra.”
I learnt that when people interact LLMs, it can feel like intelligence: understanding, creativity, reasoning.
But under the hood?
It’s math!
Not magic. Not consciousness. Not a digital brain.
Just math — and beautiful math at that.
1) Everything starts with vectors LLMs don’t “understand” words the way humans do. They convert text into vectors — lists of numbers.</description>
      <content>&lt;blockquote&gt;
&lt;p&gt;“AI feels magical, until you realize it’s mostly linear algebra.”&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;I learnt that when people interact LLMs, it can feel like intelligence:
understanding, creativity, reasoning.&lt;/p&gt;
&lt;p&gt;But under the hood?&lt;/p&gt;
&lt;p&gt;It’s math!&lt;/p&gt;
&lt;p&gt;Not magic. Not consciousness. Not a digital brain.&lt;/p&gt;
&lt;p&gt;Just math — and beautiful math at that.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;1-everything-starts-with-vectors&#34;&gt;1) Everything starts with vectors&lt;/h2&gt;
&lt;p&gt;LLMs don’t “understand” words the way humans do. They convert text into &lt;strong&gt;vectors&lt;/strong&gt; — lists of numbers.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;/images/embedding-space.png&#34; alt=&#34;Toy embedding space (2D projection)&#34;&gt;&lt;/p&gt;
&lt;p&gt;For example:&lt;/p&gt;
&lt;div class=&#34;highlight&#34;&gt;&lt;pre tabindex=&#34;0&#34; style=&#34;color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;&#34;&gt;&lt;code class=&#34;language-text&#34; data-lang=&#34;text&#34;&gt;&lt;span style=&#34;display:flex;&#34;&gt;&lt;span&gt;&amp;#34;king&amp;#34;  -&amp;gt; [0.21, -0.84, 1.33, ..., 0.02]
&lt;/span&gt;&lt;/span&gt;&lt;span style=&#34;display:flex;&#34;&gt;&lt;span&gt;&amp;#34;queen&amp;#34; -&amp;gt; [0.25, -0.79, 1.40, ..., 0.04]
&lt;/span&gt;&lt;/span&gt;&lt;/code&gt;&lt;/pre&gt;&lt;/div&gt;&lt;p&gt;Each token becomes a point in a high-dimensional space (often hundreds or thousands of dimensions).
The wild part is that meaning becomes geometry. Relationships show up as vector arithmetic:&lt;/p&gt;
&lt;p&gt;$$
\text{king} - \text{man} + \text{woman} \approx \text{queen}
$$&lt;/p&gt;
&lt;h2 id=&#34;thats-linear-algebra-working-in-semantic-space&#34;&gt;That’s linear algebra working in semantic space.&lt;/h2&gt;
&lt;h2 id=&#34;2-matrices-are-the-real-workhorses&#34;&gt;2) Matrices are the real workhorses&lt;/h2&gt;
&lt;p&gt;If vectors are points, matrices are transformations.&lt;/p&gt;
&lt;p&gt;Here’s a visual intuition: a matrix transforms a grid.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;/images/grid-original.png&#34; alt=&#34;Original grid&#34;&gt;&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;/images/grid-transformed.png&#34; alt=&#34;Grid after linear transform (matrix W)&#34;&gt;&lt;/p&gt;
&lt;p&gt;A neural network layer is often described as:&lt;/p&gt;
&lt;p&gt;$$
y = xW + b
$$&lt;/p&gt;
&lt;p&gt;Where:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;$x$ is an input vector&lt;/li&gt;
&lt;li&gt;$W$ is a weight matrix (millions or billions of learned numbers)&lt;/li&gt;
&lt;li&gt;$b$ is a bias vector&lt;/li&gt;
&lt;li&gt;$y$ is the transformed output&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;When people say:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;“This model has 70 billion parameters.”&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;They mean:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;“There are 70 billion numbers in matrices (and vectors) inside the model.”&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;Training is “just” learning those numbers.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;3-attention-is-still-just-math-dot-products--softmax&#34;&gt;3) Attention is still just math (dot-products + softmax)&lt;/h2&gt;
&lt;p&gt;Modern LLMs are based on the Transformer architecture (introduced in 2017). The key idea is &lt;strong&gt;attention&lt;/strong&gt;:
the model computes how strongly each token should relate to every other token.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;/images/attention-heatmap.png&#34; alt=&#34;Toy attention heatmap&#34;&gt;&lt;/p&gt;
&lt;p&gt;The core formula (scaled dot-product attention) is:&lt;/p&gt;
&lt;p&gt;$$
\text{Attention}(Q, K, V) = \text{softmax}\left(\frac{QK^T}{\sqrt{d}}\right)V
$$&lt;/p&gt;
&lt;p&gt;What that means in plain English:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;compute similarities via dot-products ($QK^T$)&lt;/li&gt;
&lt;li&gt;scale them ($\sqrt{d}$) so things don’t blow up&lt;/li&gt;
&lt;li&gt;normalize into probabilities with softmax&lt;/li&gt;
&lt;li&gt;mix the values ($V$) using those probabilities&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;It’s matrix multiplication, normalization, and more multiplication.&lt;/p&gt;
&lt;p&gt;Math all the way down.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;4-llms-predict-the-next-token-probability-not-certainty&#34;&gt;4) LLMs predict the next token (probability, not certainty)&lt;/h2&gt;
&lt;p&gt;At its core, an LLM is a probability machine.&lt;/p&gt;
&lt;p&gt;Given a prompt like:&lt;/p&gt;
&lt;blockquote&gt;
&lt;p&gt;“The sky is”&lt;/p&gt;
&lt;/blockquote&gt;
&lt;p&gt;it produces a probability distribution over the next token.&lt;/p&gt;
&lt;p&gt;&lt;img src=&#34;/images/softmax-probabilities.png&#34; alt=&#34;Softmax probabilities&#34;&gt;&lt;/p&gt;
&lt;p&gt;The final layer uses &lt;strong&gt;softmax&lt;/strong&gt; to convert scores (“logits”) into probabilities:&lt;/p&gt;
&lt;p&gt;$$
p_i = \frac{e^{z_i}}{\sum_j e^{z_j}}
$$&lt;/p&gt;
&lt;p&gt;Training minimizes &lt;strong&gt;cross-entropy loss&lt;/strong&gt; — a way to measure how wrong the predicted distribution is compared to the true next token:&lt;/p&gt;
&lt;p&gt;$$
\mathcal{L} = -\sum_i y_i \log(p_i)
$$&lt;/p&gt;
&lt;p&gt;Then optimization (gradient descent) adjusts parameters to reduce that loss.&lt;/p&gt;
&lt;p&gt;Which is… calculus and optimization.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;5-so-where-does-intelligence-come-from&#34;&gt;5) So where does “intelligence” come from?&lt;/h2&gt;
&lt;p&gt;There is no single place in the model that contains:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;grammar rules&lt;/li&gt;
&lt;li&gt;facts about Spain&lt;/li&gt;
&lt;li&gt;knowledge about GPUs&lt;/li&gt;
&lt;li&gt;a hard-coded reasoning engine&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;Instead, those behaviors &lt;strong&gt;emerge&lt;/strong&gt; from:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;linear algebra (vectors + matrices)&lt;/li&gt;
&lt;li&gt;non-linear functions&lt;/li&gt;
&lt;li&gt;probability distributions&lt;/li&gt;
&lt;li&gt;gradient-based optimization&lt;/li&gt;
&lt;li&gt;scale (lots of data + lots of parameters)&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;That’s the surprising part:&lt;/p&gt;
&lt;p&gt;Not that it “thinks” like us — but that math at scale can produce behavior that &lt;em&gt;feels like thinking&lt;/em&gt;.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;6-why-this-matters-if-youre-learning-ai&#34;&gt;6) Why this matters if you’re learning AI&lt;/h2&gt;
&lt;p&gt;It’s easy to feel overwhelmed by buzzwords:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Transformers&lt;/li&gt;
&lt;li&gt;RLHF&lt;/li&gt;
&lt;li&gt;fine-tuning&lt;/li&gt;
&lt;li&gt;agents&lt;/li&gt;
&lt;li&gt;multimodal models&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;But the foundation is compact:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Linear algebra&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Probability&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Calculus&lt;/strong&gt;&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Optimization&lt;/strong&gt;&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;If you understand:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;what vectors represent&lt;/li&gt;
&lt;li&gt;what matrix multiplication does&lt;/li&gt;
&lt;li&gt;what a derivative tells you&lt;/li&gt;
&lt;li&gt;what a probability distribution means&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;you understand a huge chunk of modern AI.&lt;/p&gt;
&lt;p&gt;The rest is mostly engineering choices and scale.&lt;/p&gt;
&lt;hr&gt;
&lt;h2 id=&#34;7-llms-are-math&#34;&gt;7) LLMs are math&lt;/h2&gt;
&lt;p&gt;There’s something empowering about this:&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;AI isn’t mystical.&lt;/li&gt;
&lt;li&gt;It isn’t unreachable.&lt;/li&gt;
&lt;li&gt;It isn’t reserved for a “priesthood”.&lt;/li&gt;
&lt;/ul&gt;
&lt;p&gt;It’s math.&lt;/p&gt;
&lt;p&gt;And math is learnable.&lt;/p&gt;
&lt;p&gt;The next time you see a model produce a surprisingly elegant answer, remember:&lt;/p&gt;
&lt;p&gt;Behind those words is a giant pile of matrices multiplying vectors at insane speed.&lt;/p&gt;
&lt;p&gt;And somehow… that’s enough.&lt;/p&gt;
</content>
    </item>
    
  </channel>
</rss>
