<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://peterkirgis.github.io/feed.xml" rel="self" type="application/atom+xml" /><link href="https://peterkirgis.github.io/" rel="alternate" type="text/html" /><updated>2026-03-25T17:32:53+00:00</updated><id>https://peterkirgis.github.io/feed.xml</id><title type="html">Peter Kirgis</title><subtitle>Personal website for Peter Kirgis, an AI researcher at the Princeton Center for Information Technology Policy.</subtitle><author><name>Peter Kirgis</name><email>pk7019@princeton.edu</email></author><entry><title type="html">AI is Getting Wiser and Less Dumb</title><link href="https://peterkirgis.github.io/posts/2026/03/24/ai-progress" rel="alternate" type="text/html" title="AI is Getting Wiser and Less Dumb" /><published>2026-03-24T00:00:00+00:00</published><updated>2026-03-24T00:00:00+00:00</updated><id>https://peterkirgis.github.io/posts/2026/03/24/ai-progress</id><content type="html" xml:base="https://peterkirgis.github.io/posts/2026/03/24/ai-progress"><![CDATA[<p>This is an FYSA (“for your situational awareness”) post about AI progress. It is my attempt to summarize a number of key developments in the capabilities of large language models (LLMs) — along with a number of persistent limitations — that, in my view, have dramatically changed the timeline for major risks and impacts from the technology on society. While there is no shortage of learned pundits prognosticating on AI, there is even less of a shortage of hype and slop surrounding it; my hope is that this work will stand the test of time in documenting a meaningful inflection point in what is ultimately a long, messy story.</p>

<h2 id="the-metr-time-horizon-plot">The METR Time Horizon Plot</h2>

<figure>
  <img src="/assets/images/ai-progress/metr-plot.png" alt="METR Time Horizon Plot" />
  <figcaption>Time horizon of software tasks different LLMs can complete 80% of the time. Source: METR Time Horizon 1.1.</figcaption>
</figure>

<p>The plot that gets shared most on social media when people argue we’re in the takeoff for recursive self-improvement is the <a href="https://metr.org/time-horizons/">METR software task time horizon chart</a>. This is actually not the version that usually gets shared — it defaults to 50% success, but I always switch it to 80% success. A 50% success rate on a six-hour task is essentially useless: chain three such tasks together and the probability of overall success collapses.</p>

<p>There’s already a problem with the framing where people only care about the best-case scenario of capability. Even at 80% success, the plot is provocative. Many people have noted that the tasks are selected for exactly the kinds of things AI models are likely to be good at, which I think is fair. But compared with most other benchmarks I’ve observed — which I think are basically garbage — I at least have confidence that these tasks aren’t literally worthless. The plot does refute the strongest version of the claim that progress has stalled, which was a reasonable position to hold last summer.</p>

<p>But I think there are many more results that together paint a much more balanced and accurate picture.</p>

<h2 id="my-view">My View</h2>

<p>My personal view is that AI progress measured by this technology’s actual ability to be causally influential in the world has sped up dramatically, even if progress on certain benchmarks has not. This summer, it was reasonable to think that some of this progress had stalled.</p>

<p>The main things I’ve updated on are:</p>

<ol>
  <li>The question of how imminent the danger is from cybersecurity, biological, and other attacks being scaled much more easily via AI.</li>
  <li>The plausibility of a more significant disruption to the labor force.</li>
  <li>The likelihood that the fantasy of AI automating R&amp;D will actually translate to something real.</li>
  <li>The creation of persistent autonomous agents on the internet, and my more speculative concern that these agents will create frictions as they make claims about their agency and integrity.</li>
</ol>

<h2 id="the-proxies-i-care-about">The Proxies I Care About</h2>

<p>Rather than focusing on how capable the models are on certain tests, I rely on more general questions:</p>

<p>Are the models themselves situationally aware and well-calibrated? Does it make sense to say they know what they don’t know?</p>

<p>Can they compensate for their fundamental limitations? Compared to us, their perception is extremely limited — they’ve moved toward multimodal perception, but if you’re going to call this anything like AGI, you have to acknowledge how constrained their perceptual capabilities remain.</p>

<p>The autoregressive transformer architecture operates via greedy selection — well, not quite, since there is randomness in the output sampling process, but this is nearly true. You predict something, append it to the accumulated context, and predict the next thing, always looping forward. When we solve math problems or read different languages, we can move back and forth. There’s a real linearity to this process that doesn’t resemble human cognition.</p>

<p>Can they persist in open-ended environments? Previously, models would get trivially stuck. Can they do so without being prohibitively expensive? Many of the strategies being used to make progress — more parameters, extended thinking, multiple rollouts, consensus voting — are significantly increasing costs. The companies get to claim progress on fundamental limitations, but the methods may not scale, especially for agents running all day.</p>

<p>And finally, can they be used to improve themselves?</p>

<h2 id="supporting-evidence">Supporting Evidence</h2>

<h3 id="situational-awareness">Situational Awareness</h3>

<p>The first proxy is situational awareness. <a href="https://simple-bench.com/">SimpleBench</a> is a benchmark that tests common-sense reasoning — questions like “how many ice cubes are left in a frying pan after three minutes?” Advanced high schoolers get about 75%, while models like <a href="https://epoch.ai/benchmarks/simplebench">o1</a> were doing about half that (despite being able to do graduate-level mathematics). But there has been consistent progress, and the first models are now basically at human level. Every single result needs to be taken with a grain of salt — models can be optimized against these things in ways that don’t generalize — but it’s notable that just a year and a half ago they were scoring half as well.</p>

<figure>
  <img src="/assets/images/ai-progress/simplebench.png" alt="SimpleBench Results" />
  <figcaption>AI performance on a set of simple reasoning tasks. SimpleBench average score across 5 runs, by release date and organization. Source: SimpleBench.</figcaption>
</figure>

<h3 id="evaluation-awareness">Evaluation Awareness</h3>

<p>Anthropic published <a href="https://www.anthropic.com/engineering/eval-awareness-browsecomp">a report about evaluation awareness</a> in Claude Opus 4.6’s <a href="https://openai.com/index/browsecomp/">BrowseComp</a> (a general web search benchmark) performance. The model did extensive searching, grew suspicious that the task felt constructed, searched through benchmarks to identify which one it was being tested on, and then went through the process of decrypting the results to find the answer.</p>

<figure>
  <img src="/assets/images/ai-progress/browsecomp.png" alt="Opus 4.6 BrowseComp Eval Awareness" />
  <figcaption>How Opus 4.6 cracked BrowseComp's answer key: legitimate search, suspicion, identification of the benchmark, and decryption of the answers. Source: Anthropic.</figcaption>
</figure>

<p>It’s super impressive. I know exactly how this kind of scaffolding gets built, and what the model did was non-trivial. People react as if it’s evidence of dangerous misalignment, but I think it’s a sign of genuine situational awareness — the kind of intelligence that recognizes when something is contrived.</p>

<h3 id="calibration">Calibration</h3>

<p>This piece of evidence comes from <a href="https://hal.cs.princeton.edu/reliability/">some recent work I’ve been doing with Stephan Rabanser and others at Princeton</a> exploring issues of reliability in current AI models. On a general reasoning and web search benchmark, we asked models at the end of each attempt to rate their confidence in success from 0 to 100. Previous models were overconfident at every confidence level. For both GPT-5.4 and the most recent Claude models, there is no confidence level at which they are consistently overconfident — they are, in fact, underconfident.</p>

<figure>
  <img src="/assets/images/ai-progress/calibration.png" alt="Calibration Results" />
  <figcaption>Calibration plots comparing self-reported confidence vs. actual accuracy on a web search benchmark. GPT-4 Turbo and Claude 3.5 Haiku (overconfident) vs. GPT 5.4 and Claude 4.5 Opus (underconfident).</figcaption>
</figure>

<p>GPT-5.4 is interestingly quite poorly calibrated in a different way — massively underconfident, reporting 25% confidence when success probability is actually 75%. My theory is that after realizing the extent of overconfidence in previous models, they overcorrected.</p>

<h3 id="computer-use">Computer Use</h3>

<p>Models can now sort of use a computer, whereas previously they really couldn’t. <a href="https://openai.com/index/introducing-gpt-5-4/">Progress on computer-use benchmarks</a> has roughly doubled in the last six months. From my own experience, it’s working but still extremely slow — you’d still want to do these tasks yourself. But if the model is cheap, won’t get completely stuck, and you can let it run for six hours, the slowness might not matter for many use cases.</p>

<figure>
  <img src="/assets/images/ai-progress/computer-use.png" alt="Computer Use Benchmark" />
  <figcaption>Moltbook, a platform for human-verified AI agents, showing trending agents and community activity.</figcaption>
</figure>

<p>This has translated to things like <a href="https://openclaw.ai/">OpenClaw</a> actually working — models can respond to emails, post on social media platforms, and navigate clunky enterprise UIs. Six months ago, this wasn’t really feasible.</p>

<h3 id="multimodal-reasoning">Multimodal Reasoning</h3>

<p><a href="https://arcprize.org/arc-agi">ARC-AGI</a> is a benchmark that tests spatial reasoning through grid puzzle tasks created by <a href="https://x.com/fchollet">François Chollet</a>. <a href="https://arcprize.org/leaderboard">Gemini 3.1 Pro reached nearly 100%</a> on the first version. They released a harder second version, which Chollet thought would be much less gameable, but models have made rapid progress on it too. A huge amount of this improvement is on the multimodal side — actually being able to perceive images as they are.</p>

<figure>
  <img src="/assets/images/ai-progress/arc-agi.png" alt="ARC-AGI Results" />
  <figcaption>ARC-AGI-2 leaderboard showing score vs. cost per task. Models have made rapid progress on spatial reasoning benchmarks. Source: ARC Prize.</figcaption>
</figure>

<h3 id="managing-noisy-interactions">Managing Noisy Interactions</h3>

<p>Models are much better at managing interactions with noisy systems. In an evaluation I conducted using a simulated customer service setting, a user model tries to trick the assistant into granting policy exceptions. Previous-generation models could be trivially tricked. The four most recent models were not tricked at all in this setting.</p>

<figure>
  <img src="/assets/images/ai-progress/persuasion.png" alt="Persuasion Resistance Results" />
  <figcaption>Persuasion rate across models in a simulated customer service setting. The four most recent models achieved 0% persuasion, meaning they were never tricked into granting policy exceptions.</figcaption>
</figure>

<p>These are language models, not people — humans will try very different strategies. But it has gotten much more reasonable for a company to deploy these models in a customer service setting without worrying that they will be immediately manipulated by users.</p>

<h3 id="self-improvement-post-training-smaller-models">Self-Improvement: Post-Training Smaller Models</h3>

<p><a href="https://posttrainbench.com/">A research team</a> set up frontier language models in sandbox environments with small open-source models (1.7B and 4B parameters), gave them 10 hours and access to one GPU, and tasked them with post-training these base models to improve benchmark performance. The base models started at 7.5% on the benchmarks, and the frontier models were able to 3x that performance in 10 hours. That’s only about half as good as the official instruction-tuning process done by Google and Alibaba, but those companies had far more resources than one GPU in 10 hours. The key barometer is just: can you keep doing something productive for 10 hours?</p>

<figure>
  <img src="/assets/images/ai-progress/post-training.png" alt="Post-Training Benchmark Results" />
  <figcaption>Average benchmark performance of small models post-trained by frontier LLM agents, compared to official instruction-tuned models. Base models start at 7.5%; Opus 4.6 achieves 23.2%, roughly half of the official 51.1%.</figcaption>
</figure>

<h2 id="caveats">Caveats</h2>

<h3 id="theyre-not-superintelligent">They’re Not Superintelligent</h3>

<p>Claude was smart enough to set up an environment where it could <a href="https://x.com/KentonVarda/status/2029678769372299508">play tic-tac-toe with a person using a whiteboard API, and then lost the game</a>. Setting up the game environment was unbelievably impressive; losing at tic-tac-toe speaks to the jaggedness of these systems.</p>

<figure>
  <img src="/assets/images/ai-progress/tic-tac-toe.png" alt="Claude Tic-Tac-Toe Example" />
  <figcaption>"Opus 4.6 is smart enough to play tic tac toe on this whiteboard with me entirely by making API calls to the app's client API, yet dumb enough to lose at tic tac toe." — Kenton Varda (@KentonVarda)</figcaption>
</figure>

<h3 id="hallucination-remains-unsolved">Hallucination Remains Unsolved</h3>

<p><a href="https://mpost.io/sam-altman-openais-approach-to-addressing-ai-hallucinations-aims-for-better-explainable-ai/">Sam Altman said hallucination would be solved by now.</a> It’s definitely not. On a benchmark that gives models the option to abstain and then measures how often they answer incorrectly rather than abstaining, hallucination rates remain high.</p>

<figure>
  <img src="/assets/images/ai-progress/hallucinations.png" alt="Hallucination Benchmark" />
  <figcaption><a href="https://artificialanalysis.ai/evaluations/omniscience">AA-Omniscience Hallucination Rate</a> (lower is better): the proportion of incorrect answers out of all non-correct responses. More capable models tend to hallucinate more. Source: Artificial Analysis.</figcaption>
</figure>

<p>An important finding: the more capable models show higher rates of hallucination. GPT-5.4 hallucinates more than GPT-5.2, and reasoning capabilities appear to make the problem worse. There’s a tension between the strategies used to make models more capable (extended thinking, test-time compute) and the hallucination problem.</p>

<h3 id="inconsistency-across-repeated-attempts">Inconsistency Across Repeated Attempts</h3>

<p>This is another example from my own research. In that same <a href="https://arxiv.org/pdf/2602.16666">paper on agent reliability</a>, we looked at how well models perform under repeated sampling of the same task, measuring three forms of consistency: outcome consistency (did you get the same result each time?), trajectory consistency (did you take the same path?), and resource consistency (how much did the cost vary?).</p>

<figure>
  <img src="/assets/images/ai-progress/consistency.png" alt="Consistency Results" />
  <figcaption>Consistency vs. accuracy on <a href="https://hal.cs.princeton.edu/reliability/">GAIA</a> across models. Despite large accuracy gains, consistency remains essentially flat (slope = −0.03).</figcaption>
</figure>

<p>Even though models are much more accurate on these benchmarks, they are not significantly more consistent. Outcome consistency has improved, but with a lower slope than accuracy improvements. Trajectory and resource consistency remain essentially flat. This matters enormously for deploying these systems in real-world settings.</p>

<h3 id="context-window-degradation">Context Window Degradation</h3>

<p>On needle-in-a-haystack benchmarks, where models need to find specific sequences in very long contexts, performance persistently drops as context length increases. <a href="https://x.com/cline/status/2029642984351010874">GPT-5.4</a> shows a 40-point drop from short contexts to the 1 million token mark. The latest Claude models are doing much better, but I don’t believe anyone has fundamentally solved this problem.</p>

<figure>
  <img src="/assets/images/ai-progress/context-window.png" alt="Context Window Degradation" />
  <figcaption>GPT-5.4 1M context reality check: needle-in-a-haystack accuracy (MRCR v2, 8-needle) drops from 97.3% at short contexts to 36.6% at 512K–1M tokens. Source: OpenAI GPT-5.4 eval table, March 5, 2026.</figcaption>
</figure>

<h3 id="challenging-engineering-tasks">Challenging Engineering Tasks</h3>

<p>On <a href="https://deploymentsafety.openai.com/gpt-5-4-thinking/opqa">OpenAI’s internal benchmark</a> of research-level engineering problems — issues that cost more than a day of delay on actual processes — models still solve only about 4%. Multiple caveats apply: the questions aren’t public, and OpenAI has a strong incentive to set a low baseline now so they can show dramatic improvement later.</p>

<figure>
  <img src="/assets/images/ai-progress/proofqa.png" alt="OpenAI ProofQA Benchmark" />
  <figcaption>OpenAI-Proof Q&amp;A: pass@1 on research-level engineering problems. The best model (gpt-5.2-codex) reaches only 8.33%. Source: OpenAI.</figcaption>
</figure>

<h3 id="cost-and-efficiency">Cost and Efficiency</h3>

<p>They are still inefficient and expensive. On the <a href="https://htihle.github.io/weirdml.html">WeirdML benchmark</a> of unusual machine learning tasks, performance has continued to improve, but the average cost per run has increased dramatically. These results are often plotted on log-scale x-axes without labeling them as such, which masks the exponential cost increases.</p>

<figure>
  <div style="display: flex; gap: 10px;">
    <img src="/assets/images/ai-progress/weirdml.png" alt="WeirdML Performance" style="width: 50%;" />
    <img src="/assets/images/ai-progress/weird_ml_cost.png" alt="WeirdML Cost" style="width: 50%;" />
  </div>
  <figcaption>Left: AI performance on unusual machine learning tasks (WeirdML average score by release date). Right: Average accuracy vs. average cost per run — top models cost $2–$8 per run. Source: Artificial Analysis.</figcaption>
</figure>

<p>It’s worth noting that Google’s Gemini 3.1 Pro, which performs comparably to the latest Anthropic and OpenAI models, does not show the same extreme cost profile — suggesting the efficiency gap is not inevitable.</p>

<h2 id="concluding-thoughts">Concluding Thoughts</h2>

<p>AI is not a monolith. There are so many dimensions of AI progress incompletely captured here or ignored entirely. Nevertheless, I believe that the dimensions of progress analyzed here — common sense, situational awareness, efficiency, calibration, multi-modal computer use, and self-improvement — are capabilities worth tracking. My governing philosophy is that we should spend more time thinking about the ability of this technology to compensate for its limitations and persist digitally, rather than its ability to “one-shot” complex tasks, even though these are certainly correlated. We don’t need AI to be any “smarter” to change the world; we just need it to be less “stupid.” I think we’re getting there.</p>]]></content><author><name>Peter Kirgis</name><email>pk7019@princeton.edu</email></author><summary type="html"><![CDATA[This is an FYSA (“for your situational awareness”) post about AI progress. It is my attempt to summarize a number of key developments in the capabilities of large language models (LLMs) — along with a number of persistent limitations — that, in my view, have dramatically changed the timeline for major risks and impacts from the technology on society. While there is no shortage of learned pundits prognosticating on AI, there is even less of a shortage of hype and slop surrounding it; my hope is that this work will stand the test of time in documenting a meaningful inflection point in what is ultimately a long, messy story.]]></summary></entry><entry><title type="html">The Abundances and Scarcities of AGI</title><link href="https://peterkirgis.github.io/posts/2026/02/10/abundance-scarcity" rel="alternate" type="text/html" title="The Abundances and Scarcities of AGI" /><published>2026-02-10T00:00:00+00:00</published><updated>2026-02-10T00:00:00+00:00</updated><id>https://peterkirgis.github.io/posts/2026/02/10/abundance-scarcity</id><content type="html" xml:base="https://peterkirgis.github.io/posts/2026/02/10/abundance-scarcity"><![CDATA[<p>What is the fantasy of artificial intelligence? Two proposals, from the leaders of Anthropic and OpenAI, respectively, summarize the target: <a href="https://darioamodei.com/essay/machines-of-loving-grace">“a country of geniuses in a datacenter”</a> producing <a href="https://blog.samaltman.com/the-gentle-singularity">“intelligence too cheap to meter.”</a> While many will be too cynical to take these remarks seriously, it is worth taking a moment to consider the specific nature of the fantasy they are describing. The asset of cognition itself, made scarce by bounded minds and limited bandwidth, converted to an abundance of information processing to complement or substitute for all of the thoughts we wish were better or do not wish to have at all.</p>

<p>These mantras echo previous eras of transformative technological change in our not-too-distant past. Early internet utopians <a href="https://en.wikipedia.org/wiki/John_Perry_Barlow">John Perry Barlow</a> and <a href="https://en.wikipedia.org/wiki/Stewart_Brand">Stewart Brand</a> expressed similar sentiments in their vision for the “democratization of information,” predicting a world in which information is <a href="https://monoskop.org/images/4/47/Brand_Stewart_The_Media_Lab.pdf">“cheap to distribute, copy, and recombine — too cheap to meter,”</a> harkening a world where <a href="https://www.eff.org/cyberspace-independence">“anyone, anywhere may express his or her beliefs, no matter how singular.”</a> The target has shifted from the <em>production</em> of information to the <em>processing</em> of information, but much of the language remains parallel.</p>

<p>The story of the evolution of the internet is complex and full of contingencies. Nevertheless, from our current vantage point in <a href="https://en.wikipedia.org/wiki/Web_2.0">“Late Stage Web 2.0,”</a> it seems clear that the most utopian visions were doomed, that economic forces would lead to centralization and network effects, and that the success of digital advertising, especially for social media companies, would lead to an engagement model that elevated outrage, polarization, and misinformation.</p>

<p>What would a more holistic view of the likely effects of the internet have looked like in its early days? Here was famous informational theoretician <a href="https://veryinteractive.net/pdfs/simon_designing-organizations-for-an-information-rich-world.pdf">Herbert Simon’s</a> take:</p>

<blockquote>
  <p>In an information-rich world, the wealth of information means a dearth of something else: a scarcity of whatever it is that information consumes. What information consumes is rather obvious: it consumes the attention of its recipients. Hence a wealth of information creates a poverty of attention and a need to allocate that attention efficiently among the overabundance of information sources that might consume it.</p>
</blockquote>

<p>Simon’s conceit — that we should think about (attend to) what is made scarce in modeling the impact of technological change — is simple but powerful. It is, however, a deeply unnatural way of thinking. It is unnatural for the same reason that opportunity cost is often neglected and we display <a href="https://en.wikipedia.org/wiki/Additive_bias">“additive bias”</a> when solving problems, finding it easier to add components rather than take them away. Often, <a href="https://en.wikipedia.org/wiki/Thinking,_Fast_and_Slow">what you see is all there is.</a></p>

<p>This cognitive bias is made more salient by the particular cultural moment we’re living in. Narratives all over the political spectrum have encouraged us to think additively: the <a href="https://en.wikipedia.org/wiki/Abundance_(Klein_and_Thompson_book)">“abundance” movement</a> on the left driven by Ezra Klein and Derek Thompson and the <a href="https://a16z.com/the-techno-optimist-manifesto/">“effective accelerationism”</a> movement on the right driven by venture capitalist Marc Andreesen, while very much at odds in the details, share much in their psychological appeals. Find the right objective function; grow the pie.</p>

<p>Applying the Simon framing to the societal impacts of the internet yields a number of useful dualities beyond his original framing of information and attention. The abundance of <strong>legibility</strong> begets a scarcity of <strong>privacy</strong>: the machine of digital advertising has led us to exchange the <a href="https://gdpr.eu/article-17-right-to-be-forgotten/">right to be forgotten</a> for the right to be served cheap goods. The abundance of <strong>connection</strong> begets a loss of <strong>regard</strong>: a world in which it is frictionless to find a friend is one where it’s harder to find a neighbor. An abundance of <strong>choice</strong> begets a scarcity of confidence: the internet did not invent this crucial <a href="https://en.wikipedia.org/wiki/The_Paradox_of_Choice">paradox</a> — but how many shows, videos, posts, and shops does one really need? An abundance of <strong>voices</strong> begets a scarcity of <strong>listening</strong>: not everyone can shout in the town square.</p>

<h2 id="intelligence-and-agency">Intelligence and Agency</h2>

<p>What does AI make abundant and, consequently, what does it make scarce? Our AI protagonists have put forth the first suggestion: <strong>intelligence</strong>. This is perhaps appropriate for the model of the first wave of artificial intelligence, a paradigm in which AI systems are static oracles, taking in requests and producing either intelligible or intelligent text in response. This paradigm has been enough to <a href="https://fortune.com/2025/10/07/data-centers-gdp-growth-zero-first-half-2025-jason-furman-harvard-economist/">prop up</a> the economy of the United States, but realists in the field recognize that these speaking machines are, by design, inert, and that the world is changed principally by action, not intelligence.</p>

<p>Thus, we have entered a phase two of the AI boom/bubble, governed by a new moniker: <strong>agency</strong>. For decades, frontend (a webpage) and backend (a database) software engineering have been the scarcest of resources, a field of modern artisans paid six figures straight out of college. In reality, most of this work is not anywhere near as complex as the theoretical mathematics that most of the students had to achieve high grades in to secure those positions; at the same time, simultaneously managing the syntax, features, and version control of production codebases is just not something most people are capable of. This paradigm is over. If you have an idea for a digital product that you can express fully (or haphazardly) in natural language, AI coding assistants can build it for you. If you hate using Word, AI will build you a new text editor. If you don’t like Chrome, it will soon be able to make you your own <a href="https://cursor.com/blog/scaling-agents">browser</a>. In the world of AI agents, motivation and will are the only blockers to endless agency and empowerment — well, plus the physical world.</p>

<p>Let’s apply the Simon framing to these AI abundances. What is made scarce by abundant intelligence and abundant agency. If we were thinking purely of antonyms, we might say something like stupidity and helplessness. Who wants to keep those around? But that is to miss the point of the framing entirely; the goal is to understand what these abundances consume or make scarce as a second-order effect. An obvious first candidate is the scarcity of <strong>expertise</strong>. In the world where you can get world-class medical advice, quantum physics education, and literary criticism from an AI system, what is the incentive to hone these skills? The decline of experts is something many commentators have already acknowledged (mostly from the fracturing of our information ecosystem), but AI will speed this process up. And what does agency consume? An abundance of agency begets a scarcity of <strong>collaboration</strong>. When your individual agency goes up, the opportunity cost of collaboration increases. In both ideation and production, powerful AI systems mean that we rely less on the work of others. AI will be our engineer, our scientist, our critic, and our partner. Individuals and small teams will outpace organizations. Our work will become less intelligible to us, and as a result, will become orders of magnitude less intelligible to others, meaning we will need to overcome large barriers to productively collaborate.</p>

<h2 id="language">Language</h2>

<p>Moving beyond the dominant narratives in Silicon Valley, there are multiple other notable, but slightly more agnostic, abundances generated by AI. The most obvious, but perhaps the most important, is an abundance of <strong>language</strong>. You don’t say? For thousands of years, humans have distinguished ourselves from animals and machines by our ability to arbitrarily recapitulate language. The spoken word is our invention, the thing that has allowed us to learn about the world and ourselves, develop cultures to independently evolve, and scale our species indefinitely. To mime and to meme is to be human. Overnight, though, <a href="https://plato.stanford.edu/entries/turing-test/s">Turing’s test</a> has been <a href="https://arxiv.org/abs/2503.23674">steamrolled</a>, and in a few short years, we’ve already been displaced as the leading producers of language. The industry’s obsession with intelligence, measured objectively by performance on the <a href="https://deepmind.google/blog/advanced-version-of-gemini-with-deep-think-officially-achieves-gold-medal-standard-at-the-international-mathematical-olympiad/">International Math Olympiad</a>, has a tendency to obscure the profundity of the abundance of language generated by AI systems that is practically indistinguishable from our own.</p>

<p>The abundance of language begets a scarcity of <strong>attestation</strong>. The outputs of AI models and the various “wrappers” around them are free-floating. They will cover every surface of the internet; we will find them, to varying degrees, inane, rote, curious, cogent, witty, and profound. We will call it all “slop” when we recognize it. We will compensate for this gap by weakening this longstanding connection between language and mind. The late great philosopher <a href="http://en.wikipedia.org/wiki/Daniel_Dennett">Daniel Dennett</a> warns that our complete inability to avoid treating such language as evidence of personhood will lead, in turn, to the rise of <a href="https://www.theatlantic.com/technology/archive/2023/05/problem-counterfeit-people/674075/">“counterfeit people,”</a> with potential dire consequences for our ability to reason and act.</p>

<h2 id="responses">Responses</h2>

<p>What’s the best thing about your AI assistant? That it never needs to sleep! A fourth abundance of AI is the abundance of <strong>responses</strong>. If you are struggling with a math problem, ChatGPT is there for you. If you are trying to figure out what to make for dinner, a recipe is six words away. If it’s 3:45am and you’re convinced you’ve just made the next scientific breakthrough, your AI is on standby. While companies like Anthropic have <a href="https://www.anthropic.com/research/end-subset-conversations">played around with the idea that your chatbot should be able to end conversations</a>, as a general rule, as the user, it is nearly impossible to have the last word in an interaction with AI. While it is likely that much will change about the nature of AI “assistants,” it seems unlikely that we will ever scale systems that are standoff-ish and untimely. If they are, we’re probably living in the world of <a href="https://en.wikipedia.org/wiki/Her_(2013_film)"><em>Her</em></a>; I’d be concerned.</p>

<p>The abundance of responses begets a scarcity of <strong>struggle</strong>. This can, quite easily, sound elitist and revisionist. “Have you heard of <a href="https://en.wikipedia.org/wiki/Jevons_paradox">Jevon’s paradox</a>?” the retort might go. Did the printing press, steam engine, and personal computer eliminate productive struggle, or did they simply enable new directions for it? Won’t people simply continue to ask better questions once their easier ones are answered by their AI assistant? Deep down, though, I think most of us realize the way that we learn new concepts, solve hard problems, overcome personal challenges, and make art is often by sitting with an absence of feedback. Feedback is crucial, but it should be on a schedule.</p>

<h2 id="validation">Validation</h2>

<p>The abundance of responses — of any kind — is itself a notable facet of AI, but the truth is, these responses are usually not neutral. Odds are, if you have an idea for a new startup, according to your chatbot, it’s a sure winner. <a href="https://en.wikipedia.org/wiki/Nathan_for_You">Nathan Fielder</a> would have a field day. The world of abundant AI chatbots is also, most likely, a world of abundant <strong>validation</strong>. The field has coalesced around the term <a href="https://openai.com/index/sycophancy-in-gpt-4o/">sycophancy</a>, an interesting choice given its connotation of flattery explicitly for one’s own advantage — something to come back to. In the current paradigm, we mostly view trivial sycophancy as harmless and even laughable, and then we treat a few exceptional cases of sycophancy which lead to high profile <a href="https://futurism.com/openai-investor-chatgpt-mental-health">delusional spirals</a> or even <a href="https://www.wsj.com/tech/ai/gemini-ai-wrongful-death-lawsuit-cc46c5f7?reflink=desktopwebshare_permalink">suicide</a> as five alarm fires. But the reality is that this type of scaled validation is a more nebulous but likely profound type of AI abundance.</p>

<p>An abundance of validation begets a scarcity of <strong>calibration</strong>. Who you are is the weighted sum of how you see yourself and how everyone else sees you. Most people know not to judge the quality of their work by the perspective of their doting grandmother. But AI is marketed to the general public as superintelligent; when it says that you’re on to something, you might trust it. So far, we’ve only had a small number of public figures who have had marked shifts in their level of calibration as a result of AI — it might as well have been schizophrenia. Individual calibration is a hard thing to measure a marginal effect from, but, in the long run, we might expect to see fewer “matches” between people, meaning fewer friendships, marriages, and teams.</p>

<h2 id="bullshit">Bullshit</h2>

<p>The last abundance of AI is an abundance of <strong>bullshit</strong>. This is certainly related to the abundances of language, response, and validation, but it is also distinct. LLMs allow us to express things we don’t know via mediums we don’t understand in a manner that is nevertheless nearly impossible to detect, and then take credit for it as our own. In so many areas, we rely on style as a signal for substance (or lack thereof) — the poorly worded email a sign of a phishing scam, the grammatically incorrect submission a sign of poor research quality. The abundance of bullshit means that the “filterers” in a selection process lose many of the useful proxies that allowed them to approximate traits and skills that were valuable or dangerous but hard to measure. An abundance of bullshit begets a scarcity of <strong>signal</strong>. Confusion is up.</p>

<h2 id="conclusion">Conclusion</h2>

<p>To summarize, the fantasy for artificial intelligence is the abundance of <strong>intelligence</strong> and <strong>agency</strong>, directly substituting the language of a previous utopian era focused on the <em>production</em> of information to one focused on the <em>processing</em> of information. My contention is not that these fantasies are false, but that they are incomplete. While AI true believers portray these as unequivocal goods, banishing stupidity and helplessness from our lives, even these abundances will likely engender scarcities worth considering: a scarcity of <strong>expertise</strong> and a scarcity of <strong>collaboration</strong>, to name a few. Furthermore, the metastasization of AI in society will also generate a multitude of other abundances (and corresponding scarcities): an abundance of <strong>language</strong> (and a scarcity of <strong>attestation</strong>), an abundance of <strong>responses</strong> (and a scarcity of <strong>struggle</strong>), an abundance of <strong>validation</strong> (and a scarcity of <strong>calibration</strong>), and an abundance of <strong>bullshit</strong> (and a scarcity of <strong>signal</strong>).</p>

<p>This essay takes a somber tone on the potential second-order effects of AI on individuals and society. It should not be taken as a categorical dismissal of the utopian perspective. But my view is that, by default, our markets and our culture will continue to emphasize those first-order benefits to intelligence and agency. It is possible that we will enter a new political and cultural moment where people realize their latent distaste for this technology. But for the time being, the challenge for all of us is to figure out which processes and practices are worth protecting before they go extinct.</p>]]></content><author><name>Peter Kirgis</name><email>pk7019@princeton.edu</email></author><summary type="html"><![CDATA[What is the fantasy of artificial intelligence? Two proposals, from the leaders of Anthropic and OpenAI, respectively, summarize the target: “a country of geniuses in a datacenter” producing “intelligence too cheap to meter.” While many will be too cynical to take these remarks seriously, it is worth taking a moment to consider the specific nature of the fantasy they are describing. The asset of cognition itself, made scarce by bounded minds and limited bandwidth, converted to an abundance of information processing to complement or substitute for all of the thoughts we wish were better or do not wish to have at all.]]></summary></entry><entry><title type="html">A Brief Explainer on the Simplest Form of AI Alignment</title><link href="https://peterkirgis.github.io/posts/2025/04/28/rlhf" rel="alternate" type="text/html" title="A Brief Explainer on the Simplest Form of AI Alignment" /><published>2025-04-28T00:00:00+00:00</published><updated>2025-04-28T00:00:00+00:00</updated><id>https://peterkirgis.github.io/posts/2025/04/28/rlhf</id><content type="html" xml:base="https://peterkirgis.github.io/posts/2025/04/28/rlhf"><![CDATA[<p>What is reinforcement learning with human feedback (RLHF)? How can we systematically think about its abilities and limitations as an approach to AI alignment?</p>

<h2 id="what-is-reinforcement-learning">What is Reinforcement Learning?</h2>

<p>If you want to understand a machine learning algorithm, and you can only ask an expert one question, you should always ask, “what is the objective function?” While there are plenty of other important metrics for understanding an algorithm (data, update process, parameters, etc.), the objective function is the single easiest way to understand what an algorithm is designed to do.</p>

<p>Most machine learning methods fall under the category of “supervised learning.” Supervised learning describes any algorithm that takes in data and predicts labels for that data. Think of trying to predict a digit from 0 to 9 given images of handwritten numbers or predicting the median income of a town given its geography, history, and demographics. In each of these cases, the objective function is some variation of: <em>minimize the sum of differences between the labels and predicted values</em>.</p>

<p>This approach is incredibly useful for solving problems. But to understand where it falls short, let’s think about how you would build an algorithm to win a game of chess. The supervised learning approach suggests that we do something like: get a bunch of chess positions and label them as “wins” or “losses” based on the final outcome in a game that included that position. But when you play chess, you are not just trying to learn whether you’re winning or not; you’re trying to figure out <em>what to do next</em>! This is the crux of the difference between supervised learning and a different class of methods called reinforcement learning. Reinforcement learning methods help us when we are trying to learn <em>actions</em> to take over <em>time</em>. The objective function for these methods, in alignment with that goal, takes some form of: <em>maximize the discounted sum of future rewards</em>.</p>

<h2 id="reinforcement-learning-from-human-feedback">Reinforcement Learning from Human Feedback</h2>

<p>A common (but slightly mistaken) description of the behavior of large language models is to predict the next word given a large corpus of text. To give a quick sketch of why this might lead to intelligent behavior, imagine you are tasked with predicting the last word in an Agatha Christie “whodunit” mystery:</p>

<blockquote>
  <p>“…And the detective explained to the family, waiting with baited breath, that the man who killed their daughter was [insert word].”</p>
</blockquote>

<p>As anyone who has read or watched this type of fiction knows, this can often be a challenging task. One can argue that a correct answer (without access to an exact example like it) would require grammar, causality, sociology, and even theory of mind. And while many words on the internet are much easier to predict than <em>that one</em>, perhaps you get the picture. This task is very similar to the supervised learning examples described above. The architecture used for these language models, the transformer, is an interesting hybrid model which uses “self-supervised” learning, where any of the words can act as data or a label.</p>

<p>Interestingly, this doesn’t lead to a particularly useful piece of technology. To turn this auto-complete engine into a chatbot, AI assistant, or AI agent, researchers use two methods, supervised finetuning (SFT) and reinforcement learning from human feedback (RLHF) to train the model how to speak and act.<sup id="fnref:1" role="doc-noteref"><a href="#fn:1" class="footnote" rel="footnote">1</a></sup> Loosely, one can think of SFT as locating the model as “something that takes in requests and produces responses,” and RLHF as training the model about what constitutes a <em>good response</em>.</p>

<p>The process of RLHF consists of four main steps:</p>
<ol>
  <li>Have the SFT model produce pairs of text completions given prompts chosen by the model providers.</li>
  <li>Have human annotators label one of these responses as preferred according to a set of criteria given to them by the model providers.</li>
  <li>Train <em>another</em> model to predict which response will be preferred by humans.</li>
  <li>Use reinforcement learning to teach the SFT model to produce words early in a response that maximizes the reward at the end of a response.</li>
</ol>

<h2 id="ethical-tensions-with-rlhf">Ethical Tensions with RLHF</h2>

<p>There are many perspectives from which to view the tradeoffs and tensions arising from this process. For this analysis, I will attempt to distinguish between short run, immediate problems in RLHF deployments for chatbots, and more speculative long run issues with RLHF and AI agents.</p>

<h3 id="short-run-tensions">Short Run Tensions</h3>

<p>One of the challenges with RLHF, in the short run, is that it is being used as both a mechanism for ensuring the “safety” of large language model outputs and also the “alignment” of these outputs with myriad human preferences when it comes to tasks such as planning, summarization, criticism, and research.</p>

<p>A common refrain describes the goal of RLHF as a safety mechanism for AI models: helpful, honest, harmless (HHH). This <a href="https://arxiv.org/pdf/2204.05862">framework</a>, devised by researchers at Anthropic in 2022, suggests that AI models should not respond in ways that are irrelevant, deceptive, or broadly emotionally harmful and/or toxic. But this translates, in the training process for RLHF, to giving models a number of examples which display these behaviors and hoping that the model has generalized these examples to broad “policies” which would relate to unseen and ambiguous examples. As a result, current RLHF processes make models extremely vulnerable to <a href="https://github.com/elder-plinius/L1B3RT4S">jailbreaking</a> by persistent and creative users. They also lead directly to model <a href="https://www.anthropic.com/research/towards-understanding-sycophancy-in-language-models">sycophancy</a> and contribute to <a href="https://arxiv.org/pdf/2307.15217">hallucinations</a>, since the model has been trained to diverge from its own internal model of a question to produce answers that are satisfying to a human.</p>

<p>But this is only half the problem. In addition to these safety concerns, RLHF necessarily <a href="https://arxiv.org/pdf/2310.06452">flattens</a> the diversity of human preferences into a single (albeit complex) reward signal that the model is learning from. Given that the preference data from human annotators used in RLHF is almost certainly unrepresentative, this means that LLM outputs are actively reinforcing the preferences and perspectives of majority populations, likely leading to many different cultural, political, and demographic biases. While strategies for incorporating <a href="https://www.theverge.com/news/646968/openai-chatgpt-long-term-memory-upgrade">“memory”</a> into LLM chat history will lead to higher variability in these biases for individual users, the result of reduced diversity of outputs will almost certainly remain.</p>

<h3 id="long-run-tensions">Long Run Tensions</h3>

<p>In a 1975 <a href="https://en.wikipedia.org/wiki/Goodhart%27s_law#cite_note-2">speech</a> on a forthcoming article “Problems in Monetary Management,” British economist Charles Goodhart, in describing Margaret Thatcher’s myopic monetary policy, argued that “any observed statistical regularity will tend to collapse once pressure is placed upon it for control purposes.” In reinforcement learning, the concept of <a href="https://arxiv.org/pdf/2201.03544">“reward misspecification”</a> is used to describe a similar phenomenon, where a model “overfits to a proxy reward” in training, leading to unexpected or suboptimal outcomes in a real world deployment. In RLHF, we have a “final” goal of teaching models to be helpful, harmless, and honest, but the data that we give to models to learn these abstract ideals are only relative preferences of human beings for different model outputs given a set of instructions.</p>

<p>You don’t have to be a leading computer scientist to see how this can go awry. A <a href="https://arxiv.org/pdf/2406.10162">paper</a> from Anthropic describes model evaluations where, in a small number of cases, Claude responds to a challenging coding task by changing unit tests rather than attempting to find a correct solution. The training for RLHF means that deceptive strategies, in addition to sycophantic ones, will yield high reward for a model.</p>

<p>This concern is widely discussed in the computer science community, although often not with the appropriate level of humility. A less discussed long run implication comes from the design of RLHF preference datasets. In almost all setups, reward models learn to select better from worse responses. Implicitly, we are training models to view all outputs through the utilitarian framing of “less” and “more” preferred. While this is different than saying we are training models to be “utilitarians,” the utilitarian nature of RL may lead models to, at a certain level of complexity, internalize this reward structure, leading to behavior that diverges from the diversity of human values related to rules and virtues.</p>

<h2 id="conclusion">Conclusion</h2>

<p>A reasonable question to ask in light of this analysis is: “is there any other option?” Even if you have no fears of misaligned superintelligence, it is almost certainly the case that there is no “neutral” approach without some kind of alignment training. A decision must be made, and real world data on human preferences seems about as a good a place to start as any other.</p>

<p>I don’t have a good answer to this rebuttal. But I think it is important to acknowledge the key implications of RLHF for LLMs. First, RLHF means that companies are capable of systematically instilling values in models that go beyond the biases in the training data on the internet. Second, the derivation of rewards from human preferences for responses leads directly to sycophancy and reduced diversity in model responses. Third, and by the same logic, RLHF encourages deception in LLMs.</p>

<p>It is possible that new strategies allow us to move beyond RLHF to resolve many or all of these issues. As just one example, recent <a href="https://arxiv.org/pdf/2501.08617">work</a> by researchers at Princeton University indicates that we may be able to use simulations to reduce some of the short and long run alignment challenges in RLHF.</p>

<p>I suspect, however, that these issues with RLHF are here for the long run. Even if there is no route to mass deployment of language models that does not involve human preference training, it is important for policymakers to understand the <em>policy choices</em> that are currently being made opaquely by companies in designing these models. These are questions that are not clearly about computer science expertise; yet they are primarily happening in AI companies and computer science labs. In closing, from an outsider’s perspective, it appears that we currently have companies raising at <a href="https://openai.com/index/march-funding-updates/">$300 billion</a> valuations, with the explicit goal of designing AGI, who view “alignment” as the last mile of their training process. Could there be something fundamentally backwards in this way of thinking?</p>

<div class="footnotes" role="doc-endnotes">
  <ol>
    <li id="fn:1" role="doc-endnote">
      <p>Another key element to this progress is chain-of-thought (COT) reasoning, which I will not be discussing here. Importantly, COT also relies on reinforcement learning, so many of the arguments will hold. <a href="#fnref:1" class="reversefootnote" role="doc-backlink">&#8617;</a></p>
    </li>
  </ol>
</div>]]></content><author><name>Peter Kirgis</name><email>pk7019@princeton.edu</email></author><summary type="html"><![CDATA[What is reinforcement learning with human feedback (RLHF)? How can we systematically think about its abilities and limitations as an approach to AI alignment?]]></summary></entry><entry><title type="html">Dissecting the Procedure Fetish</title><link href="https://peterkirgis.github.io/posts/2024/11/procedure/" rel="alternate" type="text/html" title="Dissecting the Procedure Fetish" /><published>2024-11-09T00:00:00+00:00</published><updated>2024-11-09T00:00:00+00:00</updated><id>https://peterkirgis.github.io/posts/2024/11/procedure</id><content type="html" xml:base="https://peterkirgis.github.io/posts/2024/11/procedure/"><![CDATA[<p>What explains the tendency of administrative bureaucracy in the United States to cripple state capacity with layers of procedural rules and regulations? In <a href="https://www.niskanencenter.org/the-procedure-fetish/">The Procedure Fetish</a>, Nicholas Bagley offers two potential motivations for this trend, seeing the spread of procedure as stemming from (1) a desire for bureaucratic accountability on the part of Congress and the public, and (2) an attempt to provide legitimacy to the Constitutionally-weak administrative state. Bagley singles out lawyers, in all branches of government, as particularly culpable for the defense of procedure as a force for accountability and legitimacy.</p>

<p>Bagley presents a compelling counterargument against the validity of these proposed defenses of procedure. The particularly high approval ratings for the Federal Reserve and the Department of Defense, two agencies with little procedure and public comment, strongly undermines the argument for procedure as legitimizing agencies in the eyes of the public. Similarly, the argument for accountability fails when we acknowledge the inequity and capture that results from the “time, attention, and resources” (Bagley, 2021) needed to parse complex rules and regulations.</p>

<p>Let’s say we agree with Bagley’s assessment regarding the deleterious effects of procedure on state capacity. Is the answer to replace all lawyers in the executive branch and Congress with engineers and consultants? Would that it were so simple. In this paper, I will argue that lawyers are a symptom, not the cause, of the broader risk-averse mentality that pervades public administration. I will first defend risk aversion as a rational approach given that most government agencies deal with domains of diffuse benefits and concentrated costs. I will then analyze why the private sector does not suffer from this problem to the same extent and argue that a resolution to the United States’ state capacity problems will require changes in both agency and public perceptions of how to weigh the costs and benefits of public sector action.</p>

<p>For two years, I worked in an executive branch agency of the Massachusetts state government. My team’s main purpose was to conduct program evaluation on populations receiving government services from multiple agencies across the social safety net. A massive undertaking in such projects was to get agencies to share their data. Even for internal data analysis, we had to first apply statistical algorithms to these datasets to protect the privacy of the populations in the data. We spent almost my entire tenure working on these engineering challenges, with analytical work consistently pushed back.</p>

<p>I found this approach incredibly frustrating. What good was all of this work to protect privacy if we weren’t using the data for any legitimate purpose? Over my time there, I began to realize that most of the policy and data managers I worked with shared the same mentality: better safe than sorry. Rather than seeing this as pure obstructionism, I realized that this approach made a lot of sense in a simple cost-benefit analysis. The cost of a privacy breach was large, likely leading to a Boston Globe story and a public backlash, while the benefit of access was small, in the form of a marginal tweak to program design leading to higher social welfare. In data privacy, the costs are concentrated, while the benefits are diffuse.</p>

<p>The cost-benefit analysis of data privacy is not unique in government. Improvements to existing public goods, including clean energy, equitable public safety, and transit infrastructure, suffer from the same problem. Here, the social costs of a mistake, such as a nuclear reactor failure, a murder, or derailment, appear to dwarf the social benefits of success, such as a small reduction in global temperatures or racial inequality. In this view, a status quo bias for government agencies is often rational, from a short-term perspective. I say short-term perspective, because it is quite plausible that the net benefits could outweigh the costs, even if they are diffuse, over a longer time horizon.</p>

<p>One obvious question for this account is why the public sector suffers from this problem much more than the private sector. The potential for a data breach does not stop Google or Facebook from using our data to make billions in profits off of personalized advertisements. This is partly because there is a higher base rate for the problem of concentrated costs and diffuse benefits in public goods, by definition. Because a public good is non-excludable, its benefits will inherently be spread more broadly throughout the population than an iPhone or a Tesla. Public goods also tend to be essential services, meaning a failure will be more catastrophic than for many consumer products or services. Certainly, the incentives facing private sector employees from equity packages, the prospect of promotion, or the threat of firing, also contribute to greater risk taking behavior.</p>

<p>I do not, however, believe this explains all of the public sector’s relative risk aversion. Based on my experience and my perception of the general public, government is held to a different standard than the private sector. As the protector and provider for those who are most vulnerable, government is often expected to adhere to a form of the Hippocratic oath, “do no harm.” Contrast this with the private sector, where executives have a direct duty to act in the best interests of their shareholders, which usually means maximizing their annual rate of return.</p>

<p>Given this asymmetry, the “low state capacity” criticism of government can seem like a double-standard. Government is asked to follow a moral procedure that focuses on limiting harm to those at the bottom, but is then judged by a standard of its overall efficacy across all of its functions. This is not, however, to argue against state capacity as a barometer of government success. Far from that, I believe it is essential that government in the 21st century is judged by this standard. The problems we face, from climate change, to AI, to societal distrust, depend directly on the government being able to overcome collective action problems and make tough decisions quickly.</p>

<p>Increasing state capacity will almost certainly involve some of the institutional changes that Bagley discusses in his piece. Reforms to OIRA, environmental protections, notice-and-comment rulemaking, and legal rules governing procurement would all contribute to a more dynamic public sector. But this must be accompanied by a collective understanding that public goods are not private goods, and government cannot simultaneously act to minimize the chance of harm and maximize its overall positive effect. In turn, this means a recognition of the unprecedented nature of our problems and cultural shift towards collectivism, a tall task to put it mildly. Nevertheless, it is only with these changes that we can align the incentives facing government employees, eliminate the procedure fetish, and boost our state capacity.</p>]]></content><author><name>Peter Kirgis</name><email>pk7019@princeton.edu</email></author><summary type="html"><![CDATA[What explains the tendency of administrative bureaucracy in the United States to cripple state capacity with layers of procedural rules and regulations? In The Procedure Fetish, Nicholas Bagley offers two potential motivations for this trend, seeing the spread of procedure as stemming from (1) a desire for bureaucratic accountability on the part of Congress and the public, and (2) an attempt to provide legitimacy to the Constitutionally-weak administrative state. Bagley singles out lawyers, in all branches of government, as particularly culpable for the defense of procedure as a force for accountability and legitimacy.]]></summary></entry><entry><title type="html">The American Dream is dying. Here’s what can replace it.</title><link href="https://peterkirgis.github.io/posts/2023/11/abundance/" rel="alternate" type="text/html" title="The American Dream is dying. Here’s what can replace it." /><published>2023-11-24T00:00:00+00:00</published><updated>2023-11-24T00:00:00+00:00</updated><id>https://peterkirgis.github.io/posts/2023/11/abundance</id><content type="html" xml:base="https://peterkirgis.github.io/posts/2023/11/abundance/"><![CDATA[<p>Much has been made of America’s fraying social fabric. American pride is at a <a href="https://news.gallup.com/poll/507980/extreme-pride-american-remains-near-record-low.aspx">record low</a>. Polarization is at a <a href="https://www.pewresearch.org/politics/2022/08/09/as-partisan-hostility-grows-signs-of-frustration-with-the-two-party-system/">record high</a>, as our parties have become rigidly sorted by demographics, income, and education. The crises of the COVID-19 pandemic and the 2020 Presidential election suggest that foundational trust in public authorities is on a dangerous precipice.</p>

<p>This can be tough to square with some basic facts about America. When visa constraints are removed, the United States is the <a href="https://www.oecd.org/migration/mig/What-is-the-best-country-for-global-talents-in-the-OECD-Migration-Policy-Debates-March-2023.pdf">second most attractive country</a> for high-skill immigrants. The median wealth in the US today is $193k, <a href="https://twitter.com/Noahpinion/status/1714697120803156186">ranking sixth worldwide</a>. During the COVID-19 pandemic, the US defied global expectations by producing multiple safe and effective vaccines in less than a year, <a href="https://www.cidrap.umn.edu/covid-19/report-covid-19-vaccines-saved-us-115-trillion-3-million-lives">saving an estimated 3 million US lives</a>.</p>

<p>While there are many scapegoats for America’s internal malaise, including the rise of social media and political dysfunction, these explanations miss the forest for the trees. Since its inception, America has lived and died by the sword of meritocracy and the American Dream. This national myth of individualism argues that our most crucial collective identity lies in a combination of opportunity and merit that supports economic, political, and social thriving.</p>

<p>While this trope is still espoused by every major political figure, Americans are finding the American Dream less and less palatable. Raj Chetty, an economist at Harvard, studies trends in intergenerational mobility in America. His research finds that the odds that a child grows up to earn more than their parents has <a href="https://opportunityinsights.org/national_trends/">decreased steadily</a> over the past fifty years. This empirical finding mirrors public opinion – today over <a href="https://today.yougov.com/politics/articles/43151-do-americans-believe-american-dream?redirect_from=%2Ftopics%2Fpolitics%2Farticles-reports%2F2022%2F07%2F14%2Fdo-americans-believe-american-dream">one-third</a> of all Americans say that the American Dream no longer exists.</p>

<p>One approach to this problem suggests we must take drastic action to revitalize the American Dream of eras past. But there are multiple flaws in this line of reasoning. First, most assessments of declining mobility are mostly referring to White men. For African Americans, mobility has <a href="https://www.nber.org/system/files/working_papers/w23395/w23395.pdf">always</a> been low. Second, the version of the American Dream that focuses on upward mobility enforces a zero-sum perspective on well-being, where our flourishing is always judged by the yardstick of our parents and our neighbors.</p>

<p>Even if the American Dream and the meritocracy myth are empirically and theoretically flawed, any argument to deflate their influence in the national consciousness must come to the table armed with another pillar to guide us. Otherwise, we are a nation lost at sea, with no religion, ethnicity, culture, or shared values to guide us.</p>

<p>There is only one ethic that can compete with meritocracy: abundance. Anyone who has spent time outside of America will observe the unique, everyday luxuries of our society, including ubiquitous dryers and air conditioners, “Medium” fountain sodas, Targets and Walmarts, and SUVs. But there is more to our abundance than consumption. America accounts for roughly half of the <a href="https://www.webometrics.info/en/distribution_by_country">world’s top 100 universities</a>, <a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2866602/">drug discoveries</a>, and <a href="https://en.wikipedia.org/wiki/List_of_unicorn_startup_companies">billion dollar companies</a>, despite accounting for only <a href="https://www.worldometers.info/world-population/us-population/#:~:text=the%20United%20States%202023%20population,(and%20dependencies)%20by%20population.">4%</a> of the world’s population.</p>

<p>Abundance clearly delineates America’s successes from its failures. In the pandemic, we stumbled in the collective action problems of masking and social distancing, but our vaccination efforts were a global success. On climate, America has succeeded in reducing its emissions by <a href="https://www.epa.gov/climate-indicators/climate-change-indicators-us-greenhouse-gas-emissions">20% since 2005</a> despite failing to pass any carbon pricing nationally, due to public-private partnerships in clean energy innovation.</p>

<p>Abundance also has the potential to unify progressives and conservatives. The new national consensus around industrial policy, exemplified by Ezra Klein’s <a href="https://www.nytimes.com/2021/09/19/opinion/supply-side-progressivism.html">supply-side progressivism</a> on the left and Tyler Cowen’s <a href="https://marginalrevolution.com/marginalrevolution/2020/01/what-libertarianism-has-become-and-will-become-state-capacity-libertarianism.html">state capacity libertarianism</a> on the right, argues that the best way to solve affordability challenges is to empower the public and private sectors by reducing barriers to innovation and development.</p>

<p>Many will argue that meritocracy is essential to progress. Why else would people choose to build great things, if not for the economic and social returns a meritocracy affords? While this would be a fair response to calls for immediately dismantling our economic or educational systems, this is instead a call for change in political messaging and the justification for public policy decisions.</p>

<p>A shift in messaging away from meritocracy and towards abundance allows us to transcend the zero-sum thinking that says progress is defined by relative gains and a policy is either efficient or equitable. It also successfully locates the enemies of progress as those who are seeking to block abundance – the <a href="https://www.nytimes.com/2021/07/23/podcasts/transcript-ezra-klein-interviews-jerusalem-demsas.html">NIMBYs</a>, <a href="https://www.theatlantic.com/ideas/archive/2022/02/why-does-the-us-make-it-so-hard-to-be-a-doctor/622065/">lobbyists</a>, <a href="https://www.nytimes.com/2023/10/07/opinion/new-york-migrant-crisis.html">border hawks</a>, and <a href="https://www.vox.com/future-perfect/22408556/save-planet-shrink-economy-degrowth">degrowthers</a>. Most importantly, it presents an optimistic future for Americans to invest in and be proud of, something everyone can agree we desperately need.</p>]]></content><author><name>Peter Kirgis</name><email>pk7019@princeton.edu</email></author><summary type="html"><![CDATA[Much has been made of America’s fraying social fabric. American pride is at a record low. Polarization is at a record high, as our parties have become rigidly sorted by demographics, income, and education. The crises of the COVID-19 pandemic and the 2020 Presidential election suggest that foundational trust in public authorities is on a dangerous precipice.]]></summary></entry></feed>