Go offline with the Player FM app!
Lawfare Daily: Kevin Frazier on Prioritizing AI Research
Manage episode 437840983 series 56794
Associate Professor at the University of Minnesota Law School and Lawfare Senior Editor Alan Rozenshtein sits down with Kevin Frazier, Assistant Professor of Law at St. Thomas University College of Law, Co-Director of the Center for Law and AI Risk, and a Tarbell Fellow at Lawfare. They discuss a new paper that Kevin has published as part of Lawfare’s ongoing Digital Social Contract paper series titled “Prioritizing International AI Research, Not Regulations.”
Frazier sheds light on the current state of AI regulation, noting that it's still in its early stages and is often under-theorized and under-enforced. He underscores the need for more targeted research to better understand the specific risks associated with AI models. Drawing parallels to risk research in the automobile industry, Frazier also explores the potential role of international institutions in consolidating expertise and establishing legitimacy in AI risk research and regulation.
To receive ad-free podcasts, become a Lawfare Material Supporter at www.patreon.com/lawfare. You can also support Lawfare by making a one-time donation at https://givebutter.com/c/trumptrials.
Support this show http://supporter.acast.com/lawfare.
Hosted on Acast. See acast.com/privacy for more information.
2316 episodes
Manage episode 437840983 series 56794
Associate Professor at the University of Minnesota Law School and Lawfare Senior Editor Alan Rozenshtein sits down with Kevin Frazier, Assistant Professor of Law at St. Thomas University College of Law, Co-Director of the Center for Law and AI Risk, and a Tarbell Fellow at Lawfare. They discuss a new paper that Kevin has published as part of Lawfare’s ongoing Digital Social Contract paper series titled “Prioritizing International AI Research, Not Regulations.”
Frazier sheds light on the current state of AI regulation, noting that it's still in its early stages and is often under-theorized and under-enforced. He underscores the need for more targeted research to better understand the specific risks associated with AI models. Drawing parallels to risk research in the automobile industry, Frazier also explores the potential role of international institutions in consolidating expertise and establishing legitimacy in AI risk research and regulation.
To receive ad-free podcasts, become a Lawfare Material Supporter at www.patreon.com/lawfare. You can also support Lawfare by making a one-time donation at https://givebutter.com/c/trumptrials.
Support this show http://supporter.acast.com/lawfare.
Hosted on Acast. See acast.com/privacy for more information.
2316 episodes
All episodes
×Welcome to Player FM!
Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.