Artwork

Content provided by Brian Carter. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Brian Carter or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!

Will o1 Ever Escape ChatGPT's Old Training?

7:41
 
Share
 

Manage episode 444202216 series 3605861
Content provided by Brian Carter. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Brian Carter or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.

This study investigates whether the reasoning abilities of large language models (LLMs) are still influenced by their origins in next-word prediction. The authors examine the performance of a new LLM from OpenAI called o1, which is specifically optimized for reasoning, on tasks that highlight the limitations of LLMs based on their autoregressive nature. While o1 shows significant improvements compared to previous LLMs, it still displays a sensitivity to the probability of both the task and the output, suggesting that reasoning optimization may not fully overcome the probabilistic biases ingrained during training. The study provides evidence for the "teleological perspective," which argues that understanding AI systems requires considering the pressures and optimizations that have shaped them.

Read more: https://arxiv.org/abs/2410.01792

  continue reading

71 episodes

Artwork
iconShare
 
Manage episode 444202216 series 3605861
Content provided by Brian Carter. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Brian Carter or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.

This study investigates whether the reasoning abilities of large language models (LLMs) are still influenced by their origins in next-word prediction. The authors examine the performance of a new LLM from OpenAI called o1, which is specifically optimized for reasoning, on tasks that highlight the limitations of LLMs based on their autoregressive nature. While o1 shows significant improvements compared to previous LLMs, it still displays a sensitivity to the probability of both the task and the output, suggesting that reasoning optimization may not fully overcome the probabilistic biases ingrained during training. The study provides evidence for the "teleological perspective," which argues that understanding AI systems requires considering the pressures and optimizations that have shaped them.

Read more: https://arxiv.org/abs/2410.01792

  continue reading

71 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Quick Reference Guide