Justified Posteriors
By: Seth Benzell and Andrey Fradkin
Language: en
Categories: Business, Society, Culture
Explorations into the economics of AI and innovation. Seth Benzell and Andrey Fradkin discuss academic papers and essays at the intersection of economics and technology. empiricrafting.substack.com
Episodes
Are We There Yet? Evaluating METR’s Eval of AI’s Ability to Complete Tasks of Different Lengths
Dec 15, 2025Seth and Andrey are back to evaluating an AI evaluation, this time discussing METR’s paper “Measuring AI Ability to Complete Long Tasks.” The paper’s central claim is that the “effective horizon” of AI agents—the length of tasks they can complete autonomously—is doubling every 7 months. Extrapolate that, and AI handles month-long projects by decade’s end.
They discuss the data and the assumptions that go into this benchmark. Seth and Andrey start by walking through the tests of task length, from simple atomic actions to the 8-hour research simulations in RE-Bench. They discuss whether the paper properly...
Duration: 01:06:09Epistemic Apocalypse and Prediction Markets (Bo Cowgill Pt. 2)
Dec 02, 2025We continue our conversation with Columbia professor Bo Cowgill. We start with a detour through Roman Jakobson’s six functions of language (plus two bonus functions Seth insists on adding: performative and incantatory). Can LLMs handle the referential? The expressive? The poetic? What about magic?
The conversation gets properly technical as we dig into Crawford-Sobel cheap talk models, the collapse of costly signaling, and whether “pay to apply” is the inevitable market response to a world where everyone can produce indistinguishable text. Bo argues we’ll see more referral hiring (your network as the last remaining credible signal...
Duration: 01:02:19Does AI Cheapen Talk? (Bo Cowgill Pt. 1)
Nov 18, 2025In this episode, we brought on our friend Bo Cowgill, to dissect his forthcoming Management Science paper, Does AI Cheapen Talk? The core question is one economists have been circling since Spence drew a line on the blackboard: What happens when a technology makes costly signals cheap? If GenAI allows anyone to produce polished pitches, résumés, and cover letters, what happens to screening, hiring, and the entire communication equilibrium?
Bo’s answer: it depends. Under some conditions, GenAI induces an epistemic apocalypse, flattening signals and confusing recruiters. In others, it reveals skill even more sharply, givi...
Duration: 00:53:26Evaluating GDPVal, OpenAI's Eval for Economic Value
Nov 04, 2025In this episode of Justified Posteriors podcast, Seth and Andrey discuss “GDPVal” a new set of AI evaluations, really a novel approach to AI evaluation, from OpenAI. The metric is debuted in a new OpenAI paper, “GDP Val: Evaluating AI Model Performance on Real-World, Economically Valuable Tasks.” We discuss this “bottom-up” approach to the possible economic impact of AI (which evaluates hundreds of specific tasks, multiplying them by estimated economic value in the economy of each), and contrast it with Daron Acemoglu’s “top-down” “Simple Macroeconomics of AI” paper (which does the same, but only for aggregate averages), as well as with measures...
Duration: 01:04:04Will Super-Intelligence's Opportunity Costs Save Human Labor?
Oct 21, 2025In this episode, Seth Benzell and Andrey Fradkin read “We Won’t Be Missed: Work and Growth in the AGI World” by Pascual Restrepo (Yale) to understand what how AGI will change work in the long run. A common metaphor for the post AGI economy is to compare AGIs to humans and men to ants. Will the AGI want to keep the humans around? Some argue that they would — there’s the possibility of useful exchange with the ants, even if they are small and weak, because an AGI will, definitionally, have opportunity costs. You might view Pascual’s paper as a f...
Duration: 00:51:39Can political science contribute to the AI discourse?
Oct 07, 2025Economists generally see AI as a production technology, or input into production. But maybe AI is actually more impactful as unlocking a new way of organizing society. Finish this story:
* The printing press unlocked the Enlightenment — along with both liberal democracy and France’s Reign of Terror
* Communism is primitive socialism plus electricity
* The radio was an essential prerequisite for fascism
* AI will unlock ????
We read “AI as Governance” by Henry Farrell in order to understand whether and how political scientists are thinking about this question.
* Concepts or other...
Duration: 00:58:19Should AI Read Without Permission?
Sep 22, 2025Many of today’s thinkers and journalists worry that AI models are eating their lunch: hoovering up these authors’ best ideas and giving them away for free or nearly free. Beyond fairness, there is a worry that these authors will stop producing valuable content if they can’t be compensated for their work. On the other hand, making lots of data freely accessible makes AI models better, potentially increasing the utility of everyone using them. Lawsuits are working their way through the courts as we speak of AI with property rights. Society needs a better of understanding the harms and be...
Duration: 00:55:06EMERGENCY POD: Is AI already causing youth unemployment?
Sep 09, 2025In our first ever EMERGENCY PODCAST, co-host Seth Benzell is summoned out of paternity leave by Andrey Fradkin to discuss the AI automation paper that’s making headlines around the world. The paper is Canaries in the Coal Mine? Six Facts about the Recent Employment Effects of Artificial Intelligence by Erik Brynjolfsson, Bharat Chandar, and Ruyu Chen. The paper is being heralded as the first evidence that AI is negatively impacting employment for young workers in certain careers. Seth and Andrey dive in, and ask — what do we believe about AI’s effect on youth employment going in, and what c...
Duration: 00:51:45AI and its labor market effects in the knowledge economy
Aug 25, 2025In this episode, we discuss a new theoretical framework for understanding how AI integrates into the economy. We read the paper Artificial Intelligence and the Knowledge Economy (Ide & Talamas, JPE), and debate whether AI will function as a worker, a manager, or an expert. Read on to learn more about the model, our thoughts, timestamp, and at the end, you can spoil yourself on Andrey and Seth’s prior beliefs and posterior conclusions — Thanks to Abdullahi Hassan for compiling these notes to make this possible.
The Ide & Talamas Model
Our discussion was based on the pape...
Duration: 00:56:30One LLM to rule them all?
Aug 12, 2025In this special episode of the Justified Posteriors Podcast, hosts Seth Benzell and Andrey Fradkin dive into the competitive dynamics of large language models (LLMs). Using Andrey’s working paper, Demand for LLMs: Descriptive Evidence on Substitution, Market Expansion, and Multihoming, they explore how quickly new models gain market share, why some cannibalize predecessors while others expand the user base, and how apps often integrate multiple models simultaneously.
Host’s note, this episode was recorded in May 2025, and things have been rapidly evolving. Look for an update sometime soon.
Transcript
Seth: Welcome to Justif...
Duration: 01:02:01What can we learn from AI exposure measures?
Jul 28, 2025In a Justified Posteriors first, hosts Seth Benzell and Andrey Fradkin sit down with economist Daniel Rock, assistant professor at Wharton and AI2050 Schmidt Science Fellow, to unpack his groundbreaking research on generative AI, productivity, exposure scores, and the future of work. Through a wide-ranging and insightful conversation, the trio examines how exposure to AI reshapes job tasks and why the difference between exposure and automation matters deeply.
Links to the referenced papers, as well as a lightly edited transcript of our conversation, with timestamps are below:
Timestamps:
[00:08] – Meet Daniel Rock[02:04] – Why AI? The...
Duration: 01:11:43A Resource Curse for AI?
Jul 14, 2025In this episode of Justified Posteriors, we tackle the provocative essay “The Intelligence Curse” by Luke Drago and Rudolf Laine. What if AI is less like a productivity booster and more like oil in a failed state? Drawing from economics, political theory, and dystopian sci-fi, we explore the analogy between AI-driven automation and the classic resource curse.
* [00:03:30] Introducing The Intelligence Curse – A speculative essay that blends LessWrong rationalism, macroeconomic theory, and political pessimism.
* [00:07:55] Running through the six economic mechanisms behind the curse, including volatility, Dutch disease, and institutional decay.
* [00:13:10] Prior #1: Will AI-enabled automation make e...
Duration: 01:09:33Robots for the retired?
Jun 30, 2025In this episode of Justified Posteriors, we examine the paper "Demographics and Automation" by economists Daron Acemoglu and Pascual Restrepo. The central hypothesis of this paper is that aging societies, facing a scarcity of middle-aged labor for physical production tasks, are more likely to invest in industrial automation.
Going in, we were split. One of us thought the idea made basic economic sense, while the other was skeptical, worrying that a vague trend of "modernity" might be the real force causing both aging populations and a rise in automation. The paper threw a mountain of data at...
Duration: 01:00:54When Humans and Machines Don't Say What They Think
Jun 16, 2025Andrey and Seth examine two papers exploring how both humans and AI systems don't always say what they think. They discuss Luca Braghieri's study on political correctness among UC San Diego students, which finds surprisingly small differences (0.1-0.2 standard deviations) between what students report privately versus publicly on hot-button issues. We then pivot to Anthropic's research showing that AI models can produce chain-of-thought reasoning that doesn't reflect their actual decision-making process. Throughout, we grapple with fundamental questions about truth, social conformity, and whether any intelligent system can fully understand or honestly represent its own thinking.
Timestamps (Transcript...
Duration: 01:09:34Scaling Laws Meet Persuasion
Jun 03, 2025In this episode, we tackle the thorny question of AI persuasion with a fresh study: "Scaling Language Model Size Yields Diminishing Returns for Single-Message Political Persuasion." The headline? Bigger AI models plateau in their persuasive power around the 70B parameter mark—think LLaMA 2 70B or Qwen-1.5 72B.
As you can imagine, this had us diving deep into what this means for AI safety concerns and the future of digital influence. Seth came in worried that super-persuasive AIs might be the top existential risk (60% confidence!), while Andrey was far more skeptical (less than 1%).
Before jumping into th...
Duration: 00:40:16Techno-prophets try macroeconomics: are they hallucinating?
May 19, 2025In this episode, we tackle a brand new paper from the folks at Epoch AI called the "GATE model" (Growth and AI Transition Endogenous model). It makes some bold claims. The headline grabber? Their default scenario projects a whopping 23% global GDP growth in 2027!
As you can imagine, that had us both (especially Andrey) practically falling out of our chairs. Before diving into GATE, Andrey shared a bit about the challenge of picking readings for his PhD course on AGI and business – a tough task when the future hasn't happened yet! Then, we broke down the GATE model it...
Duration: 01:06:23Did Meta's Algorithms Swing the 2020 Election?
May 05, 2025We hear it constantly: social media algorithms are driving polarization, feeding us echo chambers, and maybe even swinging elections. But what does the evidence actually say?
In the darkest version of this narrative, social media platform owners are shadow king-makers and puppet masters who can select the winner of close election by selectively promoting narratives. Amorally, they disregard the heightened political polarization and mental anxiety which are the consequence of their manipulations of the public psyche.
In this episode, we dive into an important study published in Science (How do social media feed algorithms affect...
Duration: 00:50:25Claude Just Refereed the Anthropic Economic Index
Apr 21, 2025In this episode of Justified Posteriors, we dive into the paper "Which Economic Tasks Are Performed with AI: Evidence from Millions of Claude Conversations." We analyze Anthropic's effort to categorize how people use their Claude AI assistant across different economic tasks and occupations, examining both the methodology and implications with a critical eye.
We came into this discussion expecting coding and writing to dominate AI usage patterns—and while the data largely confirms this, our conversation highlights several surprising insights. Why are computer and mathematical tasks so heavily overrepresented, while office administrative work lag behind? What explains th...
Duration: 01:01:56How much should we invest in AI safety?
Apr 07, 2025In this episode, we tackle one of the most pressing questions of our technological age: how much risk of human extinction should we accept in exchange for unprecedented economic growth from AI?
The podcast explores research by Stanford economist Chad Jones, who models scenarios where AI might deliver a staggering 10% annual GDP growth but carry a small probability of triggering an existential catastrophe. We dissect how our risk tolerance depends on fundamental assumptions about utility functions, time horizons, and what actually constitutes an "existential risk."
We discuss how Jones’ model presents some stark calculations: with ce...
Duration: 01:09:06Can AI make better decisions than an ER doctor?
Mar 24, 2025Dive into the intersection of economics and healthcare with our latest podcast episode. How much can AI systems enhance high-stakes medical decision-making? In this episode, we explore the implications of a research paper titled “Diagnosing Physician Error: A Machine Learning Approach to Low Value Health Care” by Sendhil Mullainathan and Ziad Obermeyer.
The paper argues that physicians often make predictable and costly errors in deciding who to test for heart attacks. The authors claim that incorporating machine learning could significantly improve the efficiency and outcome of such tests, reducing the cost per life year saved while maintaining or i...
Duration: 00:43:52If the Robots Are Coming, Why Aren't Interest Rates Higher?
Mar 10, 2025In this episode, we tackle an intriguing question inspired by a recent working paper: If artificial general intelligence (AGI) is imminent, why are real interest rates so low?
The discussion centers on the provocative paper, "Transformative AI, Existential Risk, and Real Interest Rates", authored by Trevor Chow, Basil Halperin, and Jay Zachary Maslisch. This research argues that today's historically low real interest rates signal market skepticism about the near-term arrival of transformative AI—defined here as either technology generating massive economic growth (over 30% annually) or catastrophic outcomes like total extinction.
We found ourselves initially at od...
Duration: 00:59:43High Prices, Higher Welfare? The Auto Industry as a Case Study
Feb 24, 2025Does the U.S. auto industry prioritize consumers or corporate profits? In this episode of Justified Posteriors, hosts Seth Benzell and Andrey Fradkin explore the evidence behind this question through the lens of the research paper “The Evolution of Market Power and the U.S. Automobile Industry” by Paul Grieco, Charles Murry, and Ali Yurukoglu.
Join them as they unpack trends in car prices, market concentration, and consumer surplus, critique the methodology, and consider how competition and innovation shape the auto industry. Could a different competitive structure have driven even greater innovation? Tune in to find out!
Scaling Laws in AI
Feb 10, 2025Does scaling alone hold the key to transformative AI?
In this episode of Justified Posteriors, we dive into the topic of scaling laws in artificial intelligence (AI), discussing a set of paradigmatic papers.
We discuss the idea that as more compute, data, and parameters are added to machine learning models, their performance improves predictably. Referencing several pivotal papers, including early works from OpenAI and empirical studies, they explore how scaling laws translate to model performance and potential economic value. We also debate the ultimate usefulness and limitations of scaling laws, considering whether purely increasing compute...
Duration: 00:59:06Is Social Media a Trap?
Jan 26, 2025Are we trapped by the social media we love? In this episode of the “Justified Posteriors” podcast, hosts Seth Benzell and Andrey Fradkin discuss a research paper examining the social and economic impacts of TikTok and Instagram usage among college students. The paper, authored by Leonardo Bursztyn, Benjamin Handel, Rafael Jimenez, and Christopher Roth, suggests that these platforms may create a “collective trap” where users prefer a world where no one used social media, despite the platforms' popularity. Through surveys, the researchers found that students place significant value on these platforms but also experience negative social externalities. The discussion explores...
Duration: 00:37:02Beyond Task Replacement
Jan 11, 2025In this episode, we discuss Artificial Intelligence Technologies and Aggregate Growth Prospects by Timothy Bresnahan.
* We contrast Tim Bresnahan's paper on AI's impact on economic growth, with Daron Acemoglu's task-replacement focused approach from the previous episode.
* Bresnahan argues that AI's main economic benefits will come through:
* Reorganizing organizations and tasks
* Capital deepening (improving existing machine capabilities)
* Creating new products and services rather than simply replacing human jobs
* We discuss examples from big tech companies:
* Amazon's product recommendations
* Google's search capabilities
* Voice...
Duration: 00:44:48The Simple Macroeconomics of AI
Dec 21, 2024Will AI's impact be as modest as predicted, or could it exceed expectations in reshaping economic productivity? In this episode, hosts Seth Benzell and Andrey Fradkin discuss the paper “The Simple Macroeconomics of AI” by Daron Acemoglu, an economist and an institute professor at MIT.
Additional notes from friend of the podcast Daniel Rock of Wharton, coauthor of “GPTs are GPTs: An Early Look at the Labor Market Impact Potential of Large Language Models” one of the papers cited in the show, and a main data source for Acemoglu’s paper: (1) Acemoglu does not use the paper’s ‘main’ est...
Duration: 00:53:06Situational Awareness
Dec 07, 2024How close are we to AGI, and what might its impact be on the global stage? In this episode, hosts Seth Benzell and Andrey Fradkin tackle the high-stakes world of artificial intelligence, focusing on the transformative potential of Artificial General Intelligence (AGI). The conversation is based on Leopold Aschenbrenner’s essay 'Situational Awareness', which argues that AI's development follows a predictable scaling law that allow for reliable projections about when AGI will emerge. The hosts also discuss Leopold’s thoughts on the geopolitical implications of AGI, including the influence of AI on military and social conflicts.
Sp...
Duration: 01:02:51