<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>Fundamentals on Probably Aligned</title>
    <link>https://probablyaligned.ai/fundamentals/</link>
    <description>Recent content in Fundamentals on Probably Aligned</description>
    <generator>Hugo</generator>
    <language>en-us</language>
    <lastBuildDate>Wed, 09 Apr 2025 00:00:00 +0000</lastBuildDate>
    <atom:link href="https://probablyaligned.ai/fundamentals/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Decision Theory: What an Agent Will Do and Why</title>
      <link>https://probablyaligned.ai/fundamentals/decision-theory-basics/</link>
      <pubDate>Wed, 09 Apr 2025 00:00:00 +0000</pubDate>
      <guid>https://probablyaligned.ai/fundamentals/decision-theory-basics/</guid>
      <description>Expected utility, the VNM axioms, risk sensitivity, Newcomb&amp;#39;s problem, and instrumental convergence — the decision theory that determines whether your AI is safe.</description>
    </item>
    <item>
      <title>Game Theory: Why Coordination on Safety Is So Hard</title>
      <link>https://probablyaligned.ai/fundamentals/game-theory-basics/</link>
      <pubDate>Wed, 26 Mar 2025 00:00:00 +0000</pubDate>
      <guid>https://probablyaligned.ai/fundamentals/game-theory-basics/</guid>
      <description>Nash equilibria, dominant strategies, and mechanism design — building the game theory needed to understand why AI safety suffers from coordination failures, and what it would take to fix them.</description>
    </item>
    <item>
      <title>Why Sparsity, and How Do We Get It?</title>
      <link>https://probablyaligned.ai/fundamentals/sparsity/</link>
      <pubDate>Wed, 12 Mar 2025 00:00:00 +0000</pubDate>
      <guid>https://probablyaligned.ai/fundamentals/sparsity/</guid>
      <description>Why sparse representations matter, how training encourages them, and why they&amp;#39;re useful for safety.</description>
    </item>
    <item>
      <title>What Is RL, Why Is It So Hard, and How Does It Go Wrong?</title>
      <link>https://probablyaligned.ai/fundamentals/what-is-rl/</link>
      <pubDate>Wed, 26 Feb 2025 00:00:00 +0000</pubDate>
      <guid>https://probablyaligned.ai/fundamentals/what-is-rl/</guid>
      <description>Where the neural nets are, what they do for us, why RL is fundamentally harder than supervised learning, and reward hacking vs. misalignment.</description>
    </item>
    <item>
      <title>Loss Functions, Decision Boundaries, Activation Spaces, and Why MSE</title>
      <link>https://probablyaligned.ai/fundamentals/loss-functions-and-spaces/</link>
      <pubDate>Wed, 12 Feb 2025 00:00:00 +0000</pubDate>
      <guid>https://probablyaligned.ai/fundamentals/loss-functions-and-spaces/</guid>
      <description>Three views of what a model does, how they differ, which are data-dependent, why mini-batch works, and an intuition for MSE.</description>
    </item>
    <item>
      <title>Linear Algebra Proofs of Optimality for Gradient Descent</title>
      <link>https://probablyaligned.ai/fundamentals/linear-algebra-optimality/</link>
      <pubDate>Tue, 28 Jan 2025 00:00:00 +0000</pubDate>
      <guid>https://probablyaligned.ai/fundamentals/linear-algebra-optimality/</guid>
      <description>When and why gradient descent provably finds the best solution — the linear algebra behind it.</description>
    </item>
    <item>
      <title>Gradient Descent, Backpropagation, and the Misconceptions That Tripped Me Up</title>
      <link>https://probablyaligned.ai/fundamentals/gradient-descent-and-backprop/</link>
      <pubDate>Wed, 15 Jan 2025 00:00:00 +0000</pubDate>
      <guid>https://probablyaligned.ai/fundamentals/gradient-descent-and-backprop/</guid>
      <description>Building the intuition for why the gradient is steepest descent, what it actually lives in, and how it connects to training neural networks — including the wrong mental models I had to unlearn.</description>
    </item>
  </channel>
</rss>
