URP public
[search 0]
Download the App!
show episodes
 
AXRP (pronounced axe-urp) is the AI X-risk Research Podcast where I, Daniel Filan, have conversations with researchers about their papers. We discuss the paper, and hopefully get a sense of why it's been written and how it might reduce the risk of AI causing an existential catastrophe: that is, permanently and drastically curtailing humanity's future potential. You can visit the website and read transcripts at axrp.net.
  continue reading
 
Every human being deserves to live their life to the fullest. Yet, every single day, drug addiction robs people and families of that opportunity. That problem is the driving force of this podcast. We provide a helping hand for people going through recovery; and a credible source for the truth about beating addiction. Every episode is a judgment-free space for healing, hope, and empowerment.
  continue reading
 
Your new favorite sports show. Tune in with Danache and special guests as she breaks down NFL hot topics, dissecting everything NFL on and off the field, game analysis, fantasy tips, game picks and more.
  continue reading
 
Loading …
show series
 
Reinforcement Learning from Human Feedback, or RLHF, is one of the main ways that makers of large language models make them 'aligned'. But people have long noted that there are difficulties with this approach when the models are smarter than the humans providing feedback. In this episode, I talk with Scott Emmons about his work categorizing the pro…
  continue reading
 
What's the difference between a large language model and the human brain? And what's wrong with our theories of agency? In this episode, I chat about these questions with Jan Kulveit, who leads the Alignment of Complex Systems research group. Patreon: patreon.com/axrpodcast Ko-fi: ko-fi.com/axrpodcast The transcript: axrp.net/episode/2024/05/30/epi…
  continue reading
 
What's going on with deep learning? What sorts of models get learned, and what are the learning dynamics? Singular learning theory is a theory of Bayesian statistics broad enough in scope to encompass deep neural networks that may help answer these questions. In this episode, I speak with Daniel Murfet about this research program and what it tells …
  continue reading
 
Top labs use various forms of "safety training" on models before their release to make sure they don't do nasty stuff - but how robust is that? How can we ensure that the weights of powerful AIs don't get leaked or stolen? And what can AI even do these days? In this episode, I speak with Jeffrey Ladish about security and AI. Patreon: patreon.com/ax…
  continue reading
 
In 2022, it was announced that a fairly simple method can be used to extract the true beliefs of a language model on any given topic, without having to actually understand the topic at hand. Earlier, in 2021, it was announced that neural networks sometimes 'grok': that is, when training them on certain tasks, they initially memorize their training …
  continue reading
 
How should the law govern AI? Those concerned about existential risks often push either for bans or for regulations meant to ensure that AI is developed safely - but another approach is possible. In this episode, Gabriel Weil talks about his proposal to modify tort law to enable people to sue AI companies for disasters that are "nearly catastrophic…
  continue reading
 
A lot of work to prevent AI existential risk takes the form of ensuring that AIs don't want to cause harm or take over the world---or in other words, ensuring that they're aligned. In this episode, I talk with Buck Shlegeris and Ryan Greenblatt about a different approach, called "AI control": ensuring that AI systems couldn't take over the world, e…
  continue reading
 
The events of this year have highlighted important questions about the governance of artificial intelligence. For instance, what does it mean to democratize AI? And how should we balance benefits and dangers of open-sourcing powerful AI systems such as large language models? In this episode, I speak with Elizabeth Seger about her research on these …
  continue reading
 
Imagine a world where there are many powerful AI systems, working at cross purposes. You could suppose that different governments use AIs to manage their militaries, or simply that many powerful AIs have their own wills. At any rate, it seems valuable for them to be able to cooperatively work together and minimize pointless conflict. How do we ensu…
  continue reading
 
Recently, OpenAI made a splash by announcing a new "Superalignment" team. Lead by Jan Leike and Ilya Sutskever, the team would consist of top researchers, attempting to solve alignment for superintelligent AIs in four years by figuring out how to build a trustworthy human-level AI alignment researcher, and then using it to solve the rest of the pro…
  continue reading
 
Is there some way we can detect bad behaviour in our AI system without having to know exactly what it looks like? In this episode, I speak with Mark Xu about mechanistic anomaly detection: a research direction based on the idea of detecting strange things happening in neural networks, in the hope that that will alert us of potential treacherous tur…
  continue reading
 
What can we learn about advanced deep learning systems by understanding how humans learn and form values over their lifetimes? Will superhuman AI look like ruthless coherent utility optimization, or more like a mishmash of contextually activated desires? This episode's guest, Quintin Pope, has been thinking about these questions as a leading resear…
  continue reading
 
Lots of people in the field of machine learning study 'interpretability', developing tools that they say give us useful information about neural networks. But how do we know if meaningful progress is actually being made? What should we want out of these tools? In this episode, I speak to Stephen Casper about these questions, as well as about a benc…
  continue reading
 
How should we scientifically think about the impact of AI on human civilization, and whether or not it will doom us all? In this episode, I speak with Scott Aaronson about his views on how to make progress in AI alignment, as well as his work on watermarking the output of language models, and how he moved from a background in quantum complexity the…
  continue reading
 
Danache and Jerry dive into Super Bowl LVII, the history and importance of the black quarterback, what a win or loss could mean for either Patrick Mahomes and Jalen Hurts. Are these two on the path to becoming one of the greatest to ever play the game? Andy Reid's legacy as a coach. Updates on some coaching moves around the league, LamarJackson and…
  continue reading
 
How good are we at understanding the internal computation of advanced machine learning models, and do we have a hope at getting better? In this episode, Neel Nanda talks about the sub-field of mechanistic interpretability research, as well as papers he's contributed to that explore the basics of transformer circuits, induction heads, and grokking. …
  continue reading
 
Danache and Jerry are joined by What Just Happened Sports founder Melissa Anthony. Tune in as they discuss, her journey, the importance of equipping and empowering women with sports knowledge, preview of this weekends conference championship games, and answering the question of what’s wrong with the Buffalo Bills?…
  continue reading
 
Only thing better than 2 is 3. Jon, Jerry, & Ed talk Deion Sanders and the weight of the move. Are the Cowboys an OBJ away from best team in the league? The Bucs are all but guaranteed to make the playoffs but what's the point? The Bengals could be the most slept on team in the league and the 49ers might just be alright with Mr Irrelevant…
  continue reading
 
The NFL season is starting to turn up and the next few weeks can change everything. The AFC and NFC East have turned to probably the best divisons in football. Kirk Cousins isn't looking like a joke anymore (well 1 pm kirk that is). Tune is as we cover whats been going on around the league.By Danache & Jeremy
  continue reading
 
I have a new podcast, where I interview whoever I want about whatever I want. It's called "The Filan Cabinet", and you can find it wherever you listen to podcasts. The first three episodes are about pandemic preparedness, God, and cryptocurrency. For more details, check out the podcast website (thefilancabinet.com), or search "The Filan Cabinet" in…
  continue reading
 
Jeremy is back!!!! These two jump right in and discuss all things week 4 and preview what’s to come for week 5. #packers #nfc east #ny giants #green bay packers #all the smoke #joy Taylor #taylor rooks #skip bayless #london #new England patriots #bill belichick #josh McDaniels #miami #miami Dolphins #tua #teddy bridgewater #detriot #detriot lions #…
  continue reading
 
Special Guest @UGOBOY88 joins the. show and we talk about it all. Is there a new king in the AFC South? The AFC West clearly isn't living up to expectations (yet) and are the Cowboys back? This and sooooo much more #dallas cowboys #dak Prescott # Tom Brady #the Pivot #under Review #under review podcast #undisputed #first Take #jaguars #miami #dolph…
  continue reading
 
Concept extrapolation is the idea of taking concepts an AI has about the world - say, "mass" or "does this picture contain a hot dog" - and extending them sensibly to situations where things are different - like learning that the world works via special relativity, or seeing a picture of a novel sausage-bread combination. For a while, Stuart Armstr…
  continue reading
 
Sometimes, people talk about making AI systems safe by taking examples where they fail and training them to do well on those. But how can we actually do this well, especially when we can't use a computer program to say what a 'failure' is? In this episode, I speak with Daniel Ziegler about his research group's efforts to try doing this with present…
  continue reading
 
Many people in the AI alignment space have heard of AI safety via debate - check out AXRP episode 6 (axrp.net/episode/2021/04/08/episode-6-debate-beth-barnes.html) if you need a primer. But how do we get language models to the stage where they can usefully implement debate? In this episode, I talk to Geoffrey Irving about the role of language model…
  continue reading
 
Why does anybody care about natural abstractions? Do they somehow relate to math, or value learning? How do E. coli bacteria find sources of sugar? All these questions and more will be answered in this interview with John Wentworth, where we talk about his research plan of understanding agency via natural abstractions. Topics we discuss, and timest…
  continue reading
 
Danache is joined by Jonathan and they dive into the NBA playoffs. Grizzlies/Warriors a new found rivalry?!?! Best team in the AFC West, best team in the AFC North, what players are we looking forward to seeing this upcoming season, and what is up with these notable free agents? Julio Jones, Odell Beckham, Jarvis Landry, and more.…
  continue reading
 
Why no one wants Baker..right now anyway. Cardinals pick up Kyler Murray’s 5th year option. AFC is on fire and the NFC looks like prime real estate, who is ready to take it? Brian Flores lawsuit update. Deshaun Watson, the Browns, and what needs to happen for them to move on. And of course the NFL draft. Twitter: @U_goboy @mzdanache IG: @ugoboy88 @…
  continue reading
 
Did you miss us?! We’re back for Season 3 and we’re kicking this season off with our guy Jon hosting Eddy Mr ESPN, where we chop it up and touch on all things sports from the new Rooney rule, what the MLS has to do to be successful, black ownership in sports, the state of the NBA playoffs, and much more. Don’t wanna miss it!! Ed/eddyjoseph_pes eddy…
  continue reading
 
Late last year, Vanessa Kosoy and Alexander Appel published some research under the heading of "Infra-Bayesian physicalism". But wait - what was infra-Bayesianism again? Why should we care? And what does any of this have to do with physicalism? In this episode, I talk with Vanessa Kosoy about these questions, and get a technical overview of how inf…
  continue reading
 
How should we think about artificial general intelligence (AGI), and the risks it might pose? What constraints exist on technical solutions to the problem of aligning superhuman AI systems with human intentions? In this episode, I talk to Richard Ngo about his report analyzing AGI safety from first principles, and recent conversations he had with E…
  continue reading
 
Danache is joined by special guests, Miami Herald Writer Fabian Lyon and football coach Jerry Joseph. Listen in as these three talk about the Rooney Rule, minority and black coaches, lack of minority and black GM’s as well as who they think is winning wild card weekend. Guest: @jaystepperforeva, @game_changers_underscore_Fabian_lyon…
  continue reading
 
Loading …

Quick Reference Guide