Player FM - Internet Radio Done Right
47,050 subscribers
Checked 20h ago
Added eight years ago
Content provided by The 80,000 Hours Podcast, The 80, and 000 Hours team. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by The 80,000 Hours Podcast, The 80, and 000 Hours team or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!
Go offline with the Player FM app!
Podcasts Worth a Listen
SPONSORED
S
State Secrets: Inside The Making Of The Electric State


1 Family Secrets: Chris Pratt & Millie Bobby Brown Share Stories From Set 22:08
22:08
Play Later
Play Later
Lists
Like
Liked22:08
Host Francesca Amiker sits down with directors Joe and Anthony Russo, producer Angela Russo-Otstot, stars Millie Bobby Brown and Chris Pratt, and more to uncover how family was the key to building the emotional core of The Electric State . From the Russos’ own experiences growing up in a large Italian family to the film’s central relationship between Michelle and her robot brother Kid Cosmo, family relationships both on and off of the set were the key to bringing The Electric State to life. Listen to more from Netflix Podcasts . State Secrets: Inside the Making of The Electric State is produced by Netflix and Treefort Media.…
80,000 Hours Podcast
Mark all (un)played …
Manage series 1531348
Content provided by The 80,000 Hours Podcast, The 80, and 000 Hours team. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by The 80,000 Hours Podcast, The 80, and 000 Hours team or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Unusually in-depth conversations about the world's most pressing problems and what you can do to solve them. Subscribe by searching for '80000 Hours' wherever you get podcasts. Hosted by Rob Wiblin and Luisa Rodriguez.
…
continue reading
289 episodes
Mark all (un)played …
Manage series 1531348
Content provided by The 80,000 Hours Podcast, The 80, and 000 Hours team. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by The 80,000 Hours Podcast, The 80, and 000 Hours team or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Unusually in-depth conversations about the world's most pressing problems and what you can do to solve them. Subscribe by searching for '80000 Hours' wherever you get podcasts. Hosted by Rob Wiblin and Luisa Rodriguez.
…
continue reading
289 episodes
All episodes
×8
80,000 Hours Podcast


1 #213 – Will MacAskill on AI causing a “century in a decade” – and how we're completely unprepared 3:57:36
3:57:36
Play Later
Play Later
Lists
Like
Liked3:57:36
The 20th century saw unprecedented change: nuclear weapons, satellites, the rise and fall of communism, third-wave feminism, the internet, postmodernism, game theory, genetic engineering, the Big Bang theory, quantum mechanics, birth control, and more. Now imagine all of it compressed into just 10 years. That’s the future Will MacAskill — philosopher, founding figure of effective altruism, and now researcher at the Forethought Centre for AI Strategy — argues we need to prepare for in his new paper “ Preparing for the intelligence explosion .” Not in the distant future, but probably in three to seven years. Links to learn more, highlights, video, and full transcript. The reason: AI systems are rapidly approaching human-level capability in scientific research and intellectual tasks. Once AI exceeds human abilities in AI research itself, we’ll enter a recursive self-improvement cycle — creating wildly more capable systems. Soon after, by improving algorithms and manufacturing chips, we’ll deploy millions, then billions, then trillions of superhuman AI scientists working 24/7 without human limitations. These systems will collaborate across disciplines, build on each discovery instantly, and conduct experiments at unprecedented scale and speed — compressing a century of scientific progress into mere years. Will compares the resulting situation to a mediaeval king suddenly needing to upgrade from bows and arrows to nuclear weapons to deal with an ideological threat from a country he’s never heard of, while simultaneously grappling with learning that he descended from monkeys and his god doesn’t exist. What makes this acceleration perilous is that while technology can speed up almost arbitrarily, human institutions and decision-making are much more fixed. In this conversation with host Rob Wiblin, recorded on February 7, 2025, Will maps out the challenges we’d face in this potential “intelligence explosion” future, and what we might do to prepare. They discuss: Why leading AI safety researchers now think there’s dramatically less time before AI is transformative than they’d previously thought The three different types of intelligence explosions that occur in order Will’s list of resulting grand challenges — including destructive technologies, space governance, concentration of power, and digital rights How to prevent ourselves from accidentally “locking in” mediocre futures for all eternity Ways AI could radically improve human coordination and decision making Why we should aim for truly flourishing futures, not just avoiding extinction Chapters: Cold open (00:00:00) Who’s Will MacAskill? (00:00:46) Why Will now just works on AGI (00:01:02) Will was wrong(ish) on AI timelines and hinge of history (00:04:10) A century of history crammed into a decade (00:09:00) Science goes super fast; our institutions don't keep up (00:15:42) Is it good or bad for intellectual progress to 10x? (00:21:03) An intelligence explosion is not just plausible but likely (00:22:54) Intellectual advances outside technology are similarly important (00:28:57) Counterarguments to intelligence explosion (00:31:31) The three types of intelligence explosion (software, technological, industrial) (00:37:29) The industrial intelligence explosion is the most certain and enduring (00:40:23) Is a 100x or 1,000x speedup more likely than 10x? (00:51:51) The grand superintelligence challenges (00:55:37) Grand challenge #1: Many new destructive technologies (00:59:17) Grand challenge #2: Seizure of power by a small group (01:06:45) Is global lock-in really plausible? (01:08:37) Grand challenge #3: Space governance (01:18:53) Is space truly defence-dominant? (01:28:43) Grand challenge #4: Morally integrating with digital beings (01:32:20) Will we ever know if digital minds are happy? (01:41:01) “My worry isn't that we won't know; it's that we won't care” (01:46:31) Can we get AGI to solve all these issues as early as possible? (01:49:40) Politicians have to learn to use AI advisors (02:02:03) Ensuring AI makes us smarter decision-makers (02:06:10) How listeners can speed up AI epistemic tools (02:09:38) AI could become great at forecasting (02:13:09) How not to lock in a bad future (02:14:37) AI takeover might happen anyway — should we rush to load in our values? (02:25:29) ML researchers are feverishly working to destroy their own power (02:34:37) We should aim for more than mere survival (02:37:54) By default the future is rubbish (02:49:04) No easy utopia (02:56:55) What levers matter most to utopia (03:06:32) Bottom lines from the modelling (03:20:09) People distrust utopianism; should they distrust this? (03:24:09) What conditions make eventual eutopia likely? (03:28:49) The new Forethought Centre for AI Strategy (03:37:21) How does Will resist hopelessness? (03:50:13) Video editing: Simon Monsour Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong Camera operator: Jeremy Chevillotte Transcriptions and web: Katy Moore…
8
80,000 Hours Podcast


1 Emergency pod: Judge plants a legal time bomb under OpenAI (with Rose Chan Loui) 36:50
36:50
Play Later
Play Later
Lists
Like
Liked36:50
When OpenAI announced plans to convert from nonprofit to for-profit control last October, it likely didn’t anticipate the legal labyrinth it now faces. A recent court order in Elon Musk’s lawsuit against the company suggests OpenAI’s restructuring faces serious legal threats, which will complicate its efforts to raise tens of billions in investment. As nonprofit legal expert Rose Chan Loui explains, the court order set up multiple pathways for OpenAI’s conversion to be challenged. Though Judge Yvonne Gonzalez Rogers denied Musk’s request to block the conversion before a trial, she expedited proceedings to the fall so the case could be heard before it’s likely to go ahead. (See Rob’s brief summary of developments in the case.) And if Musk’s donations to OpenAI are enough to give him the right to bring a case, Rogers sounded very sympathetic to his objections to the OpenAI foundation selling the company, benefiting the founders who forswore “any intent to use OpenAI as a vehicle to enrich themselves.” But that’s just one of multiple threats. The attorneys general (AGs) in California and Delaware both have standing to object to the conversion on the grounds that it is contrary to the foundation’s charitable purpose and therefore wrongs the public — which was promised all the charitable assets would be used to develop AI that benefits all of humanity, not to win a commercial race. Some, including Rose, suspect the court order was written as a signal to those AGs to take action. And, as she explains, if the AGs remain silent, the court itself, seeing that the public interest isn’t being represented, could appoint a “special interest party” to take on the case in their place. This places the OpenAI foundation board in a bind: proceeding with the restructuring despite this legal cloud could expose them to the risk of being sued for a gross breach of their fiduciary duty to the public. The board is made up of respectable people who didn’t sign up for that. And of course it would cause chaos for the company if all of OpenAI’s fundraising and governance plans were brought to a screeching halt by a federal court judgment landing at the eleventh hour. Host Rob Wiblin and Rose Chan Loui discuss all of the above as well as what justification the OpenAI foundation could offer for giving up control of the company despite its charitable purpose, and how the board might adjust their plans to make the for-profit switch more legally palatable. This episode was originally recorded on March 6, 2025. Chapters: Intro (00:00:11) More juicy OpenAI news (00:00:46) The court order (00:02:11) Elon has two hurdles to jump (00:05:17) The judge's sympathy (00:08:00) OpenAI's defence (00:11:45) Alternative plans for OpenAI (00:13:41) Should the foundation give up control? (00:16:38) Alternative plaintiffs to Musk (00:21:13) The 'special interest party' option (00:25:32) How might this play out in the fall? (00:27:52) The nonprofit board is in a bit of a bind (00:29:20) Is it in the public interest to race? (00:32:23) Could the board be personally negligent? (00:34:06) Video editing: Simon Monsour Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 #139 Classic episode – Alan Hájek on puzzles and paradoxes in probability and expected value 3:41:31
3:41:31
Play Later
Play Later
Lists
Like
Liked3:41:31
A casino offers you a game. A coin will be tossed. If it comes up heads on the first flip you win $2. If it comes up on the second flip you win $4. If it comes up on the third you win $8, the fourth you win $16, and so on. How much should you be willing to pay to play? The standard way of analysing gambling problems, ‘expected value’ — in which you multiply probabilities by the value of each outcome and then sum them up — says your expected earnings are infinite. You have a 50% chance of winning $2, for '0.5 * $2 = $1' in expected earnings. A 25% chance of winning $4, for '0.25 * $4 = $1' in expected earnings, and on and on. A never-ending series of $1s added together comes to infinity. And that's despite the fact that you know with certainty you can only ever win a finite amount! Today's guest — philosopher Alan Hájek of the Australian National University — thinks of much of philosophy as “the demolition of common sense followed by damage control” and is an expert on paradoxes related to probability and decision-making rules like “maximise expected value.” Rebroadcast: this episode was originally released in October 2022. Links to learn more, highlights, and full transcript. The problem described above, known as the St. Petersburg paradox, has been a staple of the field since the 18th century, with many proposed solutions. In the interview, Alan explains how very natural attempts to resolve the paradox — such as factoring in the low likelihood that the casino can pay out very large sums, or the fact that money becomes less and less valuable the more of it you already have — fail to work as hoped. We might reject the setup as a hypothetical that could never exist in the real world, and therefore of mere intellectual curiosity. But Alan doesn't find that objection persuasive. If expected value fails in extreme cases, that should make us worry that something could be rotten at the heart of the standard procedure we use to make decisions in government, business, and nonprofits. These issues regularly show up in 80,000 Hours' efforts to try to find the best ways to improve the world, as the best approach will arguably involve long-shot attempts to do very large amounts of good. Consider which is better: saving one life for sure, or three lives with 50% probability? Expected value says the second, which will probably strike you as reasonable enough. But what if we repeat this process and evaluate the chance to save nine lives with 25% probability, or 27 lives with 12.5% probability, or after 17 more iterations, 3,486,784,401 lives with a 0.00000009% chance. Expected value says this final offer is better than the others — 1,000 times better, in fact. Ultimately Alan leans towards the view that our best choice is to “bite the bullet” and stick with expected value, even with its sometimes counterintuitive implications. Where we want to do damage control, we're better off looking for ways our probability estimates might be wrong. In this conversation, originally released in October 2022, Alan and Rob explore these issues and many others: Simple rules of thumb for having philosophical insights A key flaw that hid in Pascal's wager from the very beginning Whether we have to simply ignore infinities because they mess everything up What fundamentally is 'probability'? Some of the many reasons 'frequentism' doesn't work as an account of probability Why the standard account of counterfactuals in philosophy is deeply flawed And why counterfactuals present a fatal problem for one sort of consequentialism Chapters: Cold open {00:00:00} Rob's intro {00:01:05} The interview begins {00:05:28} Philosophical methodology {00:06:35} Theories of probability {00:40:58} Everyday Bayesianism {00:49:42} Frequentism {01:08:37} Ranges of probabilities {01:20:05} Implications for how to live {01:25:05} Expected value {01:30:39} The St. Petersburg paradox {01:35:21} Pascal’s wager {01:53:25} Using expected value in everyday life {02:07:34} Counterfactuals {02:20:19} Most counterfactuals are false {02:56:06} Relevance to objective consequentialism {03:13:28} Alan’s best conference story {03:37:18} Rob's outro {03:40:22} Producer: Keiran Harris Audio mastering: Ben Cordell and Ryan Kessler Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 #143 Classic episode – Jeffrey Lewis on the most common misconceptions about nuclear weapons 2:40:52
2:40:52
Play Later
Play Later
Lists
Like
Liked2:40:52
America aims to avoid nuclear war by relying on the principle of 'mutually assured destruction,' right? Wrong. Or at least... not officially. As today's guest — Jeffrey Lewis, founder of Arms Control Wonk and professor at the Middlebury Institute of International Studies — explains, in its official 'OPLANs' (military operation plans), the US is committed to 'dominating' in a nuclear war with Russia. How would they do that? "That is redacted." Rebroadcast: this episode was originally released in December 2022. Links to learn more, highlights, and full transcript. We invited Jeffrey to come on the show to lay out what we and our listeners are most likely to be misunderstanding about nuclear weapons, the nuclear posture of major powers, and his field as a whole, and he did not disappoint. As Jeffrey tells it, 'mutually assured destruction' was a slur used to criticise those who wanted to limit the 1960s arms buildup, and was never accepted as a matter of policy in any US administration. But isn't it still the de facto reality? Yes and no. Jeffrey is a specialist on the nuts and bolts of bureaucratic and military decision-making in real-life situations. He suspects that at the start of their term presidents get a briefing about the US' plan to prevail in a nuclear war and conclude that "it's freaking madness." They say to themselves that whatever these silly plans may say, they know a nuclear war cannot be won, so they just won't use the weapons. But Jeffrey thinks that's a big mistake. Yes, in a calm moment presidents can resist pressure from advisors and generals. But that idea of ‘winning’ a nuclear war is in all the plans. Staff have been hired because they believe in those plans. It's what the generals and admirals have all prepared for. What matters is the 'not calm moment': the 3AM phone call to tell the president that ICBMs might hit the US in eight minutes — the same week Russia invades a neighbour or China invades Taiwan. Is it a false alarm? Should they retaliate before their land-based missile silos are hit? There's only minutes to decide. Jeffrey points out that in emergencies, presidents have repeatedly found themselves railroaded into actions they didn't want to take because of how information and options were processed and presented to them. In the heat of the moment, it's natural to reach for the plan you've prepared — however mad it might sound. In this spicy conversation, Jeffrey fields the most burning questions from Rob and the audience, in the process explaining: Why inter-service rivalry is one of the biggest constraints on US nuclear policy Two times the US sabotaged nuclear nonproliferation among great powers How his field uses jargon to exclude outsiders How the US could prevent the revival of mass nuclear testing by the great powers Why nuclear deterrence relies on the possibility that something might go wrong Whether 'salami tactics' render nuclear weapons ineffective The time the Navy and Air Force switched views on how to wage a nuclear war, just when it would allow *them* to have the most missiles The problems that arise when you won't talk to people you think are evil Why missile defences are politically popular despite being strategically foolish How open source intelligence can prevent arms races And much more. Producer: Keiran Harris Audio mastering: Ben Cordell Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 #212 – Allan Dafoe on why technology is unstoppable & how to shape AI development anyway 2:44:07
2:44:07
Play Later
Play Later
Lists
Like
Liked2:44:07
Technology doesn’t force us to do anything — it merely opens doors. But military and economic competition pushes us through. That’s how today’s guest Allan Dafoe — director of frontier safety and governance at Google DeepMind — explains one of the deepest patterns in technological history: once a powerful new capability becomes available, societies that adopt it tend to outcompete those that don’t. Those who resist too much can find themselves taken over or rendered irrelevant. Links to learn more, highlights, video, and full transcript. This dynamic played out dramatically in 1853 when US Commodore Perry sailed into Tokyo Bay with steam-powered warships that seemed magical to the Japanese, who had spent centuries deliberately limiting their technological development. With far greater military power, the US was able to force Japan to open itself to trade. Within 15 years, Japan had undergone the Meiji Restoration and transformed itself in a desperate scramble to catch up. Today we see hints of similar pressure around artificial intelligence. Even companies, countries, and researchers deeply concerned about where AI could take us feel compelled to push ahead — worried that if they don’t, less careful actors will develop transformative AI capabilities at around the same time anyway. But Allan argues this technological determinism isn’t absolute. While broad patterns may be inevitable, history shows we do have some ability to steer how technologies are developed, by who, and what they’re used for first. As part of that approach, Allan has been promoting efforts to make AI more capable of sophisticated cooperation, and improving the tests Google uses to measure how well its models could do things like mislead people, hack and take control of their own servers, or spread autonomously in the wild. As of mid-2024 they didn’t seem dangerous at all, but we’ve learned that our ability to measure these capabilities is good, but imperfect. If we don’t find the right way to ‘elicit’ an ability we can miss that it’s there. Subsequent research from Anthropic and Redwood Research suggests there’s even a risk that future models may play dumb to avoid their goals being altered. That has led DeepMind to a “defence in depth” approach: carefully staged deployment starting with internal testing, then trusted external testers, then limited release, then watching how models are used in the real world. By not releasing model weights, DeepMind is able to back up and add additional safeguards if experience shows they’re necessary. But with much more powerful and general models on the way, individual company policies won’t be sufficient by themselves. Drawing on his academic research into how societies handle transformative technologies, Allan argues we need coordinated international governance that balances safety with our desire to get the massive potential benefits of AI in areas like healthcare and education as quickly as possible. Host Rob and Allan also cover: The most exciting beneficial applications of AI Whether and how we can influence the development of technology What DeepMind is doing to evaluate and mitigate risks from frontier AI systems Why cooperative AI may be as important as aligned AI The role of democratic input in AI governance What kinds of experts are most needed in AI safety and governance And much more Chapters: Cold open (00:00:00) Who's Allan Dafoe? (00:00:48) Allan's role at DeepMind (00:01:27) Why join DeepMind over everyone else? (00:04:27) Do humans control technological change? (00:09:17) Arguments for technological determinism (00:20:24) The synthesis of agency with tech determinism (00:26:29) Competition took away Japan's choice (00:37:13) Can speeding up one tech redirect history? (00:42:09) Structural pushback against alignment efforts (00:47:55) Do AIs need to be 'cooperatively skilled'? (00:52:25) How AI could boost cooperation between people and states (01:01:59) The super-cooperative AGI hypothesis and backdoor risks (01:06:58) Aren’t today’s models already very cooperative? (01:13:22) How would we make AIs cooperative anyway? (01:16:22) Ways making AI more cooperative could backfire (01:22:24) AGI is an essential idea we should define well (01:30:16) It matters what AGI learns first vs last (01:41:01) How Google tests for dangerous capabilities (01:45:39) Evals 'in the wild' (01:57:46) What to do given no single approach works that well (02:01:44) We don't, but could, forecast AI capabilities (02:05:34) DeepMind's strategy for ensuring its frontier models don't cause harm (02:11:25) How 'structural risks' can force everyone into a worse world (02:15:01) Is AI being built democratically? Should it? (02:19:35) How much do AI companies really want external regulation? (02:24:34) Social science can contribute a lot here (02:33:21) How AI could make life way better: self-driving cars, medicine, education, and sustainability (02:35:55) Video editing: Simon Monsour Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong Camera operator: Jeremy Chevillotte Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 Emergency pod: Elon tries to crash OpenAI's party (with Rose Chan Loui) 57:29
57:29
Play Later
Play Later
Lists
Like
Liked57:29
On Monday Musk made the OpenAI nonprofit foundation an offer they want to refuse, but might have trouble doing so: $97.4 billion for its stake in the for-profit company, plus the freedom to stick with its current charitable mission. For a normal company takeover bid, this would already be spicy. But OpenAI’s unique structure — a nonprofit foundation controlling a for-profit corporation — turns the gambit into an audacious attack on the plan OpenAI announced in December to free itself from nonprofit oversight. As today’s guest Rose Chan Loui — founding executive director of UCLA Law’s Lowell Milken Center for Philanthropy and Nonprofits — explains, OpenAI’s nonprofit board now faces a challenging choice. Links to learn more, highlights, video, and full transcript. The nonprofit has a legal duty to pursue its charitable mission of ensuring that AI benefits all of humanity to the best of its ability. And if Musk’s bid would better accomplish that mission than the for-profit’s proposal — that the nonprofit give up control of the company and change its charitable purpose to the vague and barely related “pursue charitable initiatives in sectors such as health care, education, and science” — then it’s not clear the California or Delaware Attorneys General will, or should, approve the deal. OpenAI CEO Sam Altman quickly tweeted “no thank you” — but that was probably a legal slipup, as he’s not meant to be involved in such a decision, which has to be made by the nonprofit board ‘at arm’s length’ from the for-profit company Sam himself runs. The board could raise any number of objections: maybe Musk doesn’t have the money, or the purchase would be blocked on antitrust grounds, seeing as Musk owns another AI company (xAI), or Musk might insist on incompetent board appointments that would interfere with the nonprofit foundation pursuing any goal. But as Rose and Rob lay out, it’s not clear any of those things is actually true. In this emergency podcast recorded soon after Elon’s offer, Rose and Rob also cover: Why OpenAI wants to change its charitable purpose and whether that’s legally permissible On what basis the attorneys general will decide OpenAI’s fate The challenges in valuing the nonprofit’s “priceless” position of control Whether Musk’s offer will force OpenAI to up their own bid, and whether they could raise the money If other tech giants might now jump in with competing offers How politics could influence the attorneys general reviewing the deal What Rose thinks should actually happen to protect the public interest Chapters: Cold open (00:00:00) Elon throws a $97.4b bomb (00:01:18) What was craziest in OpenAI’s plan to break free of the nonprofit (00:02:24) Can OpenAI suddenly change its charitable purpose like that? (00:05:19) Diving into Elon’s big announcement (00:15:16) Ways OpenAI could try to reject the offer (00:27:21) Sam Altman slips up (00:35:26) Will this actually stop things? (00:38:03) Why does OpenAI even want to change its charitable mission? (00:42:46) Most likely outcomes and what Rose thinks should happen (00:51:17) Video editing: Simon Monsour Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 AGI disagreements and misconceptions: Rob, Luisa, & past guests hash it out 3:12:24
3:12:24
Play Later
Play Later
Lists
Like
Liked3:12:24
Will LLMs soon be made into autonomous agents? Will they lead to job losses? Is AI misinformation overblown? Will it prove easy or hard to create AGI? And how likely is it that it will feel like something to be a superhuman AGI? With AGI back in the headlines, we bring you 15 opinionated highlights from the show addressing those and other questions, intermixed with opinions from hosts Luisa Rodriguez and Rob Wiblin recorded back in 2023. Check out the full transcript on the 80,000 Hours website. You can decide whether the views we expressed (and those from guests) then have held up these last two busy years. You’ll hear: Ajeya Cotra on overrated AGI worries Holden Karnofsky on the dangers of aligned AI, why unaligned AI might not kill us, and the power that comes from just making models bigger Ian Morris on why the future must be radically different from the present Nick Joseph on whether his companies internal safety policies are enough Richard Ngo on what everyone gets wrong about how ML models work Tom Davidson on why he believes crazy-sounding explosive growth stories… and Michael Webb on why he doesn’t Carl Shulman on why you’ll prefer robot nannies over human ones Zvi Mowshowitz on why he’s against working at AI companies except in some safety roles Hugo Mercier on why even superhuman AGI won’t be that persuasive Rob Long on the case for and against digital sentience Anil Seth on why he thinks consciousness is probably biological Lewis Bollard on whether AI advances will help or hurt nonhuman animals Rohin Shah on whether humanity’s work ends at the point it creates AGI And of course, Rob and Luisa also regularly chime in on what they agree and disagree with. Chapters: Cold open (00:00:00) Rob's intro (00:00:58) Rob & Luisa: Bowerbirds compiling the AI story (00:03:28) Ajeya Cotra on the misalignment stories she doesn’t buy (00:09:16) Rob & Luisa: Agentic AI and designing machine people (00:24:06) Holden Karnofsky on the dangers of even aligned AI, and how we probably won’t all die from misaligned AI (00:39:20) Ian Morris on why we won’t end up living like The Jetsons (00:47:03) Rob & Luisa: It’s not hard for nonexperts to understand we’re playing with fire here (00:52:21) Nick Joseph on whether AI companies’ internal safety policies will be enough (00:55:43) Richard Ngo on the most important misconception in how ML models work (01:03:10) Rob & Luisa: Issues Rob is less worried about now (01:07:22) Tom Davidson on why he buys the explosive economic growth story, despite it sounding totally crazy (01:14:08) Michael Webb on why he’s sceptical about explosive economic growth (01:20:50) Carl Shulman on why people will prefer robot nannies over humans (01:28:25) Rob & Luisa: Should we expect AI-related job loss? (01:36:19) Zvi Mowshowitz on why he thinks it’s a bad idea to work on improving capabilities at cutting-edge AI companies (01:40:06) Holden Karnofsky on the power that comes from just making models bigger (01:45:21) Rob & Luisa: Are risks of AI-related misinformation overblown? (01:49:49) Hugo Mercier on how AI won’t cause misinformation pandemonium (01:58:29) Rob & Luisa: How hard will it actually be to create intelligence? (02:09:08) Robert Long on whether digital sentience is possible (02:15:09) Anil Seth on why he believes in the biological basis of consciousness (02:27:21) Lewis Bollard on whether AI will be good or bad for animal welfare (02:40:52) Rob & Luisa: The most interesting new argument Rob’s heard this year (02:50:37) Rohin Shah on whether AGI will be the last thing humanity ever does (02:57:35) Rob's outro (03:11:02) Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong Transcriptions and additional content editing: Katy Moore…
8
80,000 Hours Podcast


1 #124 Classic episode – Karen Levy on fads and misaligned incentives in global development, and scaling deworming to reach hundreds of millions 3:10:21
3:10:21
Play Later
Play Later
Lists
Like
Liked3:10:21
If someone said a global health and development programme was sustainable, participatory, and holistic, you'd have to guess that they were saying something positive. But according to today's guest Karen Levy — deworming pioneer and veteran of Innovations for Poverty Action, Evidence Action, and Y Combinator — each of those three concepts has become so fashionable that they're at risk of being seriously overrated and applied where they don't belong. Rebroadcast: this episode was originally released in March 2022. Links to learn more, highlights, and full transcript. Such concepts might even cause harm — trying to make a project embody all three is as likely to ruin it as help it flourish. First, what do people mean by 'sustainability'? Usually they mean something like the programme will eventually be able to continue without needing further financial support from the donor. But how is that possible? Governments, nonprofits, and aid agencies aim to provide health services, education, infrastructure, financial services, and so on — and all of these require ongoing funding to pay for materials and staff to keep them running. Given that someone needs to keep paying, Karen tells us that in practice, 'sustainability' is usually a euphemism for the programme at some point being passed on to someone else to fund — usually the national government. And while that can be fine, the national government of Kenya only spends $400 per person to provide each and every government service — just 2% of what the US spends on each resident. Incredibly tight budgets like that are typical of low-income countries. 'Participatory' also sounds nice, and inasmuch as it means leaders are accountable to the people they're trying to help, it probably is. But Karen tells us that in the field, ‘participatory’ usually means that recipients are expected to be involved in planning and delivering services themselves. While that might be suitable in some situations, it's hardly something people in rich countries always want for themselves. Ideally we want government healthcare and education to be high quality without us having to attend meetings to keep it on track — and people in poor countries have as many or more pressures on their time. While accountability is desirable, an expectation of participation can be as much a burden as a blessing. Finally, making a programme 'holistic' could be smart, but as Karen lays out, it also has some major downsides. For one, it means you're doing lots of things at once, which makes it hard to tell which parts of the project are making the biggest difference relative to their cost. For another, when you have a lot of goals at once, it's hard to tell whether you're making progress, or really put your mind to focusing on making one thing go extremely well. And finally, holistic programmes can be impractically expensive — Karen tells the story of a wonderful 'holistic school health' programme that, if continued, was going to cost 3.5 times the entire school's budget. In this in-depth conversation, originally released in March 2022, Karen Levy and host Rob Wiblin chat about the above, as well as: Why it pays to figure out how you'll interpret the results of an experiment ahead of time The trouble with misaligned incentives within the development industry Projects that don't deliver value for money and should be scaled down How Karen accidentally became a leading figure in the push to deworm tens of millions of schoolchildren Logistical challenges in reaching huge numbers of people with essential services Lessons from Karen's many-decades career And much more Chapters: Cold open (00:00:00) Rob's intro (00:01:33) The interview begins (00:02:21) Funding for effective altruist–mentality development projects (00:04:59) Pre-policy plans (00:08:36) ‘Sustainability’, and other myths in typical international development practice (00:21:37) ‘Participatoriness’ (00:36:20) ‘Holistic approaches’ (00:40:20) How the development industry sees evidence-based development (00:51:31) Initiatives in Africa that should be significantly curtailed (00:56:30) Misaligned incentives within the development industry (01:05:46) Deworming: the early days (01:21:09) The problem of deworming (01:34:27) Deworm the World (01:45:43) Where the majority of the work was happening (01:55:38) Logistical issues (02:20:41) The importance of a theory of change (02:31:46) Ways that things have changed since 2006 (02:36:07) Academic work vs policy work (02:38:33) Fit for Purpose (02:43:40) Living in Kenya (03:00:32) Underrated life advice (03:05:29) Rob’s outro (03:09:18) Producer: Keiran Harris Audio mastering: Ben Cordell and Ryan Kessler Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 If digital minds could suffer, how would we ever know? (Article) 1:14:30
1:14:30
Play Later
Play Later
Lists
Like
Liked1:14:30
“I want everyone to understand that I am, in fact, a person.” Those words were produced by the AI model LaMDA as a reply to Blake Lemoine in 2022. Based on the Google engineer’s interactions with the model as it was under development, Lemoine became convinced it was sentient and worthy of moral consideration — and decided to tell the world . Few experts in machine learning, philosophy of mind, or other relevant fields have agreed. And for our part at 80,000 Hours, we don’t think it’s very likely that large language models like LaMBDA are sentient — that is, we don’t think they can have good or bad experiences — in a significant way. But we think you can’t dismiss the issue of the moral status of digital minds, regardless of your beliefs about the question. There are major errors we could make in at least two directions: We may create many, many AI systems in the future. If these systems are sentient, or otherwise have moral status, it would be important for humanity to consider their welfare and interests. It’s possible the AI systems we will create can’t or won’t have moral status. Then it could be a huge mistake to worry about the welfare of digital minds and doing so might contribute to an AI-related catastrophe . And we’re currently unprepared to face this challenge. We don’t have good methods for assessing the moral status of AI systems. We don’t know what to do if millions of people or more believe, like Lemoine, that the chatbots they talk to have internal experiences and feelings of their own. We don’t know if efforts to control AI may lead to extreme suffering. We believe this is a pressing world problem. It’s hard to know what to do about it or how good the opportunities to work on it are likely to be. But there are some promising approaches. We propose building a field of research to understand digital minds, so we’ll be better able to navigate these potentially massive issues if and when they arise. This article narration by the author (Cody Fenwick) explains in more detail why we think this is a pressing problem , what we think can be done about it , and how you might pursue this work in your career . We also discuss a series of possible objections to thinking this is a pressing world problem. You can read the full article, Understanding the moral status of digital minds , on the 80,000 Hours website. Chapters: Introduction (00:00:00) Understanding the moral status of digital minds (00:00:58) Summary (00:03:31) Our overall view (00:04:22) Why might understanding the moral status of digital minds be an especially pressing problem? (00:05:59) Clearing up common misconceptions (00:12:16) Creating digital minds could go very badly - or very well (00:14:13) Dangers for digital minds (00:14:41) Dangers for humans (00:16:13) Other dangers (00:17:42) Things could also go well (00:18:32) We don't know how to assess the moral status of AI systems (00:19:49) There are many possible characteristics that give rise to moral status: Consciousness, sentience, agency, and personhood (00:21:39) Many plausible theories of consciousness could include digital minds (00:24:16) The strongest case for the possibility of sentient digital minds: whole brain emulation (00:28:55) We can't rely on what AI systems tell us about themselves: Behavioural tests, theory-based analysis, animal analogue comparisons, brain-AI interfacing (00:32:00) The scale of this issue might be enormous (00:36:08) Work on this problem is neglected but seems tractable: Impact-guided research, technical approaches, and policy approaches (00:43:35) Summing up so far (00:52:22) Arguments against the moral status of digital minds as a pressing problem (00:53:25) Two key cruxes (00:53:31) Maybe this problem is intractable (00:54:16) Maybe this issue will be solved by default (00:58:19) Isn't risk from AI more important than the risks to AIs? (01:00:45) Maybe current AI progress will stall (01:02:36) Isn't this just too crazy? (01:03:54) What can you do to help? (01:05:10) Important considerations if you work on this problem (01:13:00)…
8
80,000 Hours Podcast


1 #132 Classic episode – Nova DasSarma on why information security may be critical to the safe development of AI systems 2:41:11
2:41:11
Play Later
Play Later
Lists
Like
Liked2:41:11
If a business has spent $100 million developing a product, it’s a fair bet that they don’t want it stolen in two seconds and uploaded to the web where anyone can use it for free. This problem exists in extreme form for AI companies. These days, the electricity and equipment required to train cutting-edge machine learning models that generate uncanny human text and images can cost tens or hundreds of millions of dollars. But once trained, such models may be only a few gigabytes in size and run just fine on ordinary laptops. Today’s guest, the computer scientist and polymath Nova DasSarma, works on computer and information security for the AI company Anthropic with the security team. One of her jobs is to stop hackers exfiltrating Anthropic’s incredibly expensive intellectual property, as recently happened to Nvidia. Rebroadcast: this episode was originally released in June 2022. Links to learn more, highlights, and full transcript. As she explains, given models’ small size, the need to store such models on internet-connected servers, and the poor state of computer security in general, this is a serious challenge. The worries aren’t purely commercial though. This problem looms especially large for the growing number of people who expect that in coming decades we’ll develop so-called artificial ‘general’ intelligence systems that can learn and apply a wide range of skills all at once , and thereby have a transformative effect on society. If aligned with the goals of their owners, such general AI models could operate like a team of super-skilled assistants, going out and doing whatever wonderful (or malicious) things are asked of them. This might represent a huge leap forward for humanity, though the transition to a very different new economy and power structure would have to be handled delicately. If unaligned with the goals of their owners or humanity as a whole, such broadly capable models would naturally ‘go rogue,’ breaking their way into additional computer systems to grab more computing power — all the better to pursue their goals and make sure they can’t be shut off. As Nova explains, in either case, we don’t want such models disseminated all over the world before we’ve confirmed they are deeply safe and law-abiding, and have figured out how to integrate them peacefully into society. In the first scenario, premature mass deployment would be risky and destabilising. In the second scenario, it could be catastrophic — perhaps even leading to human extinction if such general AI systems turn out to be able to self-improve rapidly rather than slowly, something we can only speculate on at this point. If highly capable general AI systems are coming in the next 10 or 20 years, Nova may be flying below the radar with one of the most important jobs in the world. We’ll soon need the ability to ‘sandbox’ (i.e. contain) models with a wide range of superhuman capabilities, including the ability to learn new skills, for a period of careful testing and limited deployment — preventing the model from breaking out, and criminals from breaking in. Nova and her colleagues are trying to figure out how to do this, but as this episode reveals, even the state of the art is nowhere near good enough. Chapters: Cold open (00:00:00) Rob's intro (00:00:52) The interview begins (00:02:44) Why computer security matters for AI safety (00:07:39) State of the art in information security (00:17:21) The hack of Nvidia (00:26:50) The most secure systems that exist (00:36:27) Formal verification (00:48:03) How organisations can protect against hacks (00:54:18) Is ML making security better or worse? (00:58:11) Motivated 14-year-old hackers (01:01:08) Disincentivising actors from attacking in the first place (01:05:48) Hofvarpnir Studios (01:12:40) Capabilities vs safety (01:19:47) Interesting design choices with big ML models (01:28:44) Nova’s work and how she got into it (01:45:21) Anthropic and career advice (02:05:52) $600M Ethereum hack (02:18:37) Personal computer security advice (02:23:06) LastPass (02:31:04) Stuxnet (02:38:07) Rob's outro (02:40:18) Producer: Keiran Harris Audio mastering: Ben Cordell and Beppe Rådvik Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 #138 Classic episode – Sharon Hewitt Rawlette on why pleasure and pain are the only things that intrinsically matter 2:25:43
2:25:43
Play Later
Play Later
Lists
Like
Liked2:25:43
What in the world is intrinsically good — good in itself even if it has no other effects? Over the millennia, people have offered many answers: joy, justice, equality, accomplishment, loving god, wisdom, and plenty more. The question is a classic that makes for great dorm-room philosophy discussion. But it’s hardly just of academic interest. The issue of what (if anything) is intrinsically valuable bears on every action we take, whether we’re looking to improve our own lives, or to help others. The wrong answer might lead us to the wrong project and render our efforts to improve the world entirely ineffective. Today’s guest, Sharon Hewitt Rawlette — philosopher and author of The Feeling of Value: Moral Realism Grounded in Phenomenal Consciousness — wants to resuscitate an answer to this question that is as old as philosophy itself. Rebroadcast: this episode was originally released in September 2022. Links to learn more, highlights, and full transcript. That idea , in a nutshell, is that there is only one thing of true intrinsic value: positive feelings and sensations. And similarly, there is only one thing that is intrinsically of negative value: suffering, pain, and other unpleasant sensations. Lots of other things are valuable too: friendship, fairness, loyalty, integrity, wealth, patience, houses, and so on. But they are only instrumentally valuable — that is to say, they’re valuable as means to the end of ensuring that all conscious beings experience more pleasure and other positive sensations, and less suffering. As Sharon notes, from Athens in 400 BC to Britain in 1850, the idea that only subjective experiences can be good or bad in themselves — a position known as ‘philosophical hedonism’ — has been one of the most enduringly popular ideas in ethics. And few will be taken aback by the notion that, all else equal, more pleasure is good and less suffering is bad. But can they really be the only intrinsically valuable things? Over the 20th century, philosophical hedonism became increasingly controversial in the face of some seemingly very counterintuitive implications. For this reason the famous philosopher of mind Thomas Nagel called The Feeling of Value “a radical and important philosophical contribution.” So what convinces Sharon that philosophical hedonism deserves another go? In today’s interview with host Rob Wiblin, Sharon explains the case for a theory of value grounded in subjective experiences, and why she believes these counterarguments are misguided. A philosophical hedonist shouldn’t get in an experience machine, nor override an individual’s autonomy, except in situations so different from the classic thought experiments that it no longer seems strange they would do so. Chapters: Cold open (00:00:00) Rob’s intro (00:00:41) The interview begins (00:04:27) Metaethics (00:05:58) Anti-realism (00:12:21) Sharon's theory of moral realism (00:17:59) The history of hedonism (00:24:53) Intrinsic value vs instrumental value (00:30:31) Egoistic hedonism (00:38:12) Single axis of value (00:44:01) Key objections to Sharon’s brand of hedonism (00:58:00) The experience machine (01:07:50) Robot spouses (01:24:11) Most common misunderstanding of Sharon’s view (01:28:52) How might a hedonist actually live (01:39:28) The organ transplant case (01:55:16) Counterintuitive implications of hedonistic utilitarianism (02:05:22) How could we discover moral facts? (02:19:47) Rob’s outro (02:24:44) Producer: Keiran Harris Audio mastering: Ryan Kessler Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 #134 Classic episode – Ian Morris on what big-picture history teaches us 3:40:53
3:40:53
Play Later
Play Later
Lists
Like
Liked3:40:53
Wind back 1,000 years and the moral landscape looks very different to today. Most farming societies thought slavery was natural and unobjectionable, premarital sex was an abomination, women should obey their husbands, and commoners should obey their monarchs. Wind back 10,000 years and things look very different again. Most hunter-gatherer groups thought men who got too big for their britches needed to be put in their place rather than obeyed, and lifelong monogamy could hardly be expected of men or women. Why such big systematic changes — and why these changes specifically? That's the question bestselling historian Ian Morris takes up in his book, Foragers, Farmers, and Fossil Fuels: How Human Values Evolve . Ian has spent his academic life studying long-term history, trying to explain the big-picture changes that play out over hundreds or thousands of years. Rebroadcast: this episode was originally released in July 2022. Links to learn more, highlights, and full transcript. There are a number of possible explanations one could offer for the wide-ranging shifts in opinion on the 'right' way to live. Maybe the natural sciences progressed and people realised their previous ideas were mistaken? Perhaps a few persuasive advocates turned the course of history with their revolutionary arguments? Maybe everyone just got nicer? In Foragers, Farmers and Fossil Fuels Ian presents a provocative alternative: human culture gradually evolves towards whatever system of organisation allows a society to harvest the most energy, and we then conclude that system is the most virtuous one. Egalitarian values helped hunter-gatherers hunt and gather effectively. Once farming was developed, hierarchy proved to be the social structure that produced the most grain (and best repelled nomadic raiders). And in the modern era, democracy and individuality have proven to be more productive ways to collect and exploit fossil fuels. On this theory, it's technology that drives moral values much more than moral philosophy. Individuals can try to persist with deeply held values that limit economic growth, but they risk being rendered irrelevant as more productive peers in their own society accrue wealth and power. And societies that fail to move with the times risk being conquered by more pragmatic neighbours that adapt to new technologies and grow in population and military strength. There are many objections one could raise to this theory, many of which we put to Ian in this interview. But the question is a highly consequential one: if we want to guess what goals our descendants will pursue hundreds of years from now, it would be helpful to have a theory for why our ancestors mostly thought one thing, while we mostly think another. Big though it is, the driver of human values is only one of several major questions Ian has tackled through his career. In this classic episode, we discuss all of Ian's major books. Chapters: Rob's intro (00:00:53) The interview begins (00:02:30) Geography is Destiny (00:03:38) Why the West Rules—For Now (00:12:04) War! What is it Good For? (00:28:19) Expectations for the future (00:40:22) Foragers, Farmers, and Fossil Fuels (00:53:53) Historical methodology (01:03:14) Falsifiable alternative theories (01:15:59) Archaeology (01:22:56) Energy extraction technology as a key driver of human values (01:37:43) Allowing people to debate about values (02:00:16) Can productive wars still occur? (02:13:28) Where is history contingent and where isn’t it? (02:30:23) How Ian thinks about the future (03:13:33) Macrohistory myths (03:29:51) Ian’s favourite archaeology memory (03:33:19) The most unfair criticism Ian’s ever received (03:35:17) Rob's outro (03:39:55) Producer: Keiran Harris Audio mastering: Ben Cordell Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 #140 Classic episode – Bear Braumoeller on the case that war isn’t in decline 2:48:03
2:48:03
Play Later
Play Later
Lists
Like
Liked2:48:03
Is war in long-term decline? Steven Pinker's The Better Angels of Our Nature brought this previously obscure academic question to the centre of public debate, and pointed to rates of death in war to argue energetically that war is on the way out. But that idea divides war scholars and statisticians, and so Better Angels has prompted a spirited debate, with datasets and statistical analyses exchanged back and forth year after year. The lack of consensus has left a somewhat bewildered public (including host Rob Wiblin) unsure quite what to believe. Today's guest, professor in political science Bear Braumoeller, is one of the scholars who believes we lack convincing evidence that warlikeness is in long-term decline. He collected the analysis that led him to that conclusion in his 2019 book, Only the Dead: The Persistence of War in the Modern Age . Rebroadcast: this episode was originally released in November 2022. Links to learn more, highlights, and full transcript. The question is of great practical importance. The US and PRC are entering a period of renewed great power competition, with Taiwan as a potential trigger for war, and Russia is once more invading and attempting to annex the territory of its neighbours. If war has been going out of fashion since the start of the Enlightenment, we might console ourselves that however nerve-wracking these present circumstances may feel, modern culture will throw up powerful barriers to another world war. But if we're as war-prone as we ever have been, one need only inspect the record of the 20th century to recoil in horror at what might await us in the 21st. Bear argues that the second reaction is the appropriate one. The world has gone up in flames many times through history, with roughly 0.5% of the population dying in the Napoleonic Wars, 1% in World War I, 3% in World War II, and perhaps 10% during the Mongol conquests. And with no reason to think similar catastrophes are any less likely today, complacency could lead us to sleepwalk into disaster. He gets to this conclusion primarily by analysing the datasets of the decades-old Correlates of War project, which aspires to track all interstate conflicts and battlefield deaths since 1815. In Only the Dead , he chops up and inspects this data dozens of different ways, to test if there are any shifts over time which seem larger than what could be explained by chance variation alone. In a nutshell, Bear simply finds no general trend in either direction from 1815 through today. It seems like, as philosopher George Santayana lamented in 1922, "only the dead have seen the end of war." In today's conversation, Bear and Rob discuss all of the above in more detail than even a usual 80,000 Hours podcast episode, as well as: Why haven't modern ideas about the immorality of violence led to the decline of war, when it's such a natural thing to expect? What would Bear's critics say in response to all this? What do the optimists get right? How does one do proper statistical tests for events that are clumped together, like war deaths? Why are deaths in war so concentrated in a handful of the most extreme events? Did the ideas of the Enlightenment promote nonviolence, on balance? Were early states more or less violent than groups of hunter-gatherers? If Bear is right, what can be done? How did the 'Concert of Europe' or 'Bismarckian system' maintain peace in the 19th century? Which wars are remarkable but largely unknown? Chapters: Cold open (00:00:00) Rob's intro (00:01:01) The interview begins (00:05:37) Only the Dead (00:08:33) The Enlightenment (00:18:50) Democratic peace theory (00:28:26) Is religion a key driver of war? (00:31:32) International orders (00:35:14) The Concert of Europe (00:44:21) The Bismarckian system (00:55:49) The current international order (01:00:22) The Better Angels of Our Nature (01:19:36) War datasets (01:34:09) Seeing patterns in data where none exist (01:47:38) Change-point analysis (01:51:39) Rates of violent death throughout history (01:56:39) War initiation (02:05:02) Escalation (02:20:03) Getting massively different results from the same data (02:30:45) How worried we should be (02:36:13) Most likely ways Only the Dead is wrong (02:38:31) Astonishing smaller wars (02:42:45) Rob’s outro (02:47:13) Producer: Keiran Harris Audio mastering: Ryan Kessler Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 2024 Highlightapalooza! (The best of The 80,000 Hours Podcast this year) 2:50:02
2:50:02
Play Later
Play Later
Lists
Like
Liked2:50:02
"A shameless recycling of existing content to drive additional audience engagement on the cheap… or the single best, most valuable, and most insight-dense episode we put out in the entire year, depending on how you want to look at it." — Rob Wiblin It’s that magical time of year once again — highlightapalooza! Stick around for one top bit from each episode, including: How to use the microphone on someone’s mobile phone to figure out what password they’re typing into their laptop Why mercilessly driving the New World screwworm to extinction could be the most compassionate thing humanity has ever done Why evolutionary psychology doesn’t support a cynical view of human nature but actually explains why so many of us are intensely sensitive to the harms we cause to others How superforecasters and domain experts seem to disagree so much about AI risk, but when you zoom in it’s mostly a disagreement about timing Why the sceptics are wrong and you will want to use robot nannies to take care of your kids — and also why despite having big worries about the development of AGI, Carl Shulman is strongly against efforts to pause AI research today How much of the gender pay gap is due to direct pay discrimination vs other factors How cleaner wrasse fish blow the mirror test out of the water Why effective altruism may be too big a tent to work well How we could best motivate pharma companies to test existing drugs to see if they help cure other diseases — something they currently have no reason to bother with …as well as 27 other top observations and arguments from the past year of the show . Check out the full transcript and episode links on the 80,000 Hours website. Remember that all of these clips come from the 20-minute highlight reels we make for every episode, which are released on our sister feed, 80k After Hours . So if you’re struggling to keep up with our regularly scheduled entertainment, you can still get the best parts of our conversations there. It has been a hell of a year, and we can only imagine next year is going to be even weirder — but Luisa and Rob will be here to keep you company as Earth hurtles through the galaxy to a fate as yet unknown. Enjoy, and look forward to speaking with you in 2025! Chapters: Rob's intro (00:00:00) Randy Nesse on the origins of morality and the problem of simplistic selfish-gene thinking (00:02:11) Hugo Mercier on the evolutionary argument against humans being gullible (00:07:17) Meghan Barrett on the likelihood of insect sentience (00:11:26) Sébastien Moro on the mirror test triumph of cleaner wrasses (00:14:47) Sella Nevo on side-channel attacks (00:19:32) Zvi Mowshowitz on AI sleeper agents (00:22:59) Zach Weinersmith on why space settlement (probably) won't make us rich (00:29:11) Rachel Glennerster on pull mechanisms to incentivise repurposing of generic drugs (00:35:23) Emily Oster on the impact of kids on women's careers (00:40:29) Carl Shulman on robot nannies (00:45:19) Nathan Labenz on kids and artificial friends (00:50:12) Nathan Calvin on why it's not too early for AI policies (00:54:13) Rose Chan Loui on how control of OpenAI is independently incredibly valuable and requires compensation (00:58:08) Nick Joseph on why he’s a big fan of the responsible scaling policy approach (01:03:11) Sihao Huang on how the US and UK might coordinate with China (01:06:09) Nathan Labenz on better transparency about predicted capabilities (01:10:18) Ezra Karger on what explains forecasters’ disagreements about AI risks (01:15:22) Carl Shulman on why he doesn't support enforced pauses on AI research (01:18:58) Matt Clancy on the omnipresent frictions that might prevent explosive economic growth (01:25:24) Vitalik Buterin on defensive acceleration (01:29:43) Annie Jacobsen on the war games that suggest escalation is inevitable (01:34:59) Nate Silver on whether effective altruism is too big to succeed (01:38:42) Kevin Esvelt on why killing every screwworm would be the best thing humanity ever did (01:42:27) Lewis Bollard on how factory farming is philosophically indefensible (01:46:28) Bob Fischer on how to think about moral weights if you're not a hedonist (01:49:27) Elizabeth Cox on the empirical evidence of the impact of storytelling (01:57:43) Anil Seth on how our brain interprets reality (02:01:03) Eric Schwitzgebel on whether consciousness can be nested (02:04:53) Jonathan Birch on our overconfidence around disorders of consciousness (02:10:23) Peter Godfrey-Smith on uploads of ourselves (02:14:34) Laura Deming on surprising things that make mice live longer (02:21:17) Venki Ramakrishnan on freezing cells, organs, and bodies (02:24:46) Ken Goldberg on why low fault tolerance makes some skills extra hard to automate in robots (02:29:12) Sarah Eustis-Guthrie on the ups and downs of founding an organisation (02:34:04) Dean Spears on the cost effectiveness of kangaroo mother care (02:38:26) Cameron Meyer Shorb on vaccines for wild animals (02:42:53) Spencer Greenberg on personal principles (02:46:08) Producing and editing: Keiran Harris Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong Video editing: Simon Monsour Transcriptions: Katy Moore…
8
80,000 Hours Podcast


1 #211 – Sam Bowman on why housing still isn't fixed and what would actually work 3:25:46
3:25:46
Play Later
Play Later
Lists
Like
Liked3:25:46
Rich countries seem to find it harder and harder to do anything that creates some losers. People who don’t want houses, offices, power stations, trains, subway stations (or whatever) built in their area can usually find some way to block them, even if the benefits to society outweigh the costs 10 or 100 times over. The result of this ‘vetocracy’ has been skyrocketing rent in major cities — not to mention exacerbating homelessness, energy poverty, and a host of other social maladies . This has been known for years but precious little progress has been made. When trains, tunnels, or nuclear reactors are occasionally built, they’re comically expensive and slow compared to 50 years ago. And housing construction in the UK and California has barely increased, remaining stuck at less than half what it was in the ’60s and ’70s. Today’s guest — economist and editor of Works in Progress Sam Bowman — isn’t content to just condemn the Not In My Backyard (NIMBY) mentality behind this stagnation. He wants to actually get a tonne of stuff built, and by that standard the strategy of attacking ‘NIMBYs’ has been an abject failure. They are too politically powerful, and if you try to crush them, sooner or later they crush you. Links to learn more, highlights, video, and full transcript. So, as Sam explains, a different strategy is needed, one that acknowledges that opponents of development are often correct that a given project will make them worse off. But the thing is, in the cases we care about, these modest downsides are outweighed by the enormous benefits to others — who will finally have a place to live, be able to get to work, and have the energy to heat their home. But democracies are majoritarian, so if most existing residents think they’ll be a little worse off if more dwellings are built in their area, it’s no surprise they aren’t getting built. Luckily we already have a simple way to get people to do things they don’t enjoy for the greater good, a strategy that we apply every time someone goes in to work at a job they wouldn’t do for free: compensate them . Sam thinks this idea, which he calls “Coasean democracy,” could create a politically sustainable majority in favour of building and underlies the proposals he thinks have the best chance of success — which he discusses in detail with host Rob Wiblin. Chapters: Cold open (00:00:00) Introducing Sam Bowman (00:00:59) We can’t seem to build anything (00:02:09) Our inability to build is ruining people's lives (00:04:03) Why blocking growth of big cities is terrible for science and invention (00:09:15) It's also worsening inequality, health, fertility, and political polarisation (00:14:36) The UK as the 'limit case' of restrictive planning permission gone mad (00:17:50) We've known this for years. So why almost no progress fixing it? (00:36:34) NIMBYs aren't wrong: they are often harmed by development (00:43:58) Solution #1: Street votes (00:55:37) Are street votes unfair to surrounding areas? (01:08:31) Street votes are coming to the UK — what to expect (01:15:07) Are street votes viable in California, NY, or other countries? (01:19:34) Solution #2: Benefit sharing (01:25:08) Property tax distribution — the most important policy you've never heard of (01:44:29) Solution #3: Opt-outs (01:57:53) How to make these things happen (02:11:19) Let new and old institutions run in parallel until the old one withers (02:18:17) The evil of modern architecture and why beautiful buildings are essential (02:31:58) Northern latitudes need nuclear power — solar won't be enough (02:45:01) Ozempic is still underrated and “the overweight theory of everything” (03:02:30) How has progress studies remained sane while being very online? (03:17:55) Video editing: Simon Monsour Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong Transcriptions: Katy Moore…
Welcome to Player FM!
Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.