Hosted by award-winning author Robert Rodriguez, Something About The Beatles is an intelligent but entertaining examination of The Beatles' music and career. Smart, funny and surprising - just like the Fab Four.
…
continue reading
The C-Squared Podcast is an in depth weekly discussion about the chess world with your hosts, Grandmasters Fabiano Caruana and Cristian Chirila. Bringing the biggest news and chess stars to your doorstep!
…
continue reading
A forum for discussing matters on student growth for the Oak Grove Elementary School community (Fayette County Public School, GA).
…
continue reading
Geek Mythology Podcast brings you all of the latest news in TV, Movies and Gaming! Join hosts Toby, Tara, Lel and Andrew as they jump into the world of geek and discuss all of the latest ongoing topics that surround the industry that’s kept us on the couch for years. Themed episodes every week, a new geeky discussion topic each episode and the 2014 Geek Mythology Olympiad are just some of the things you can expect in the Geek Mythology Podcast. What are you waiting for? Jump in.
…
continue reading
Running out of time to catch up with new arXiv papers? We take the most impactful papers and present them as convenient podcasts. If you're a visual learner, we offer these papers in an engaging video format. Our service fills the gap between overly brief paper summaries and time-consuming full paper reads. You gain academic insights in a time-efficient, digestible format. Code behind this work: https://github.com/imelnyk/ArxivPapers Support this podcast: https://podcasters.spotify.com/pod/s ...
…
continue reading
Wordquest Devon is a project of Aune Head Arts, Cyprus Well, Devon Libraries (Devon County Council), and the University of Exeter. It began life as a response to the 'questing' themes of the Cultural Olympiad, and still retains many of these playful and game-playing aspects. The project's live presence began in June 2011 and continues until September 2012. Around the County (and perhaps beyond) you can pick up copies of our 'Literary Map for Devon' - but you can also explore the online map h ...
…
continue reading
Counsellor Charlie brings you the latest from Camp Eden Woods in Harcourt, Ontario. Listen in to find out about camp news, funny stories and hear from the campers themselves!
…
continue reading
A SOLO school going TEENAGER takes on the challenge of delivering the GREATEST podcast to mankind. To see how he conquers the podcast world tune in to see a new episode EVERY WEEK on EVERY PLATFORM. Witness greatness!!!!
…
continue reading
My Name is Anjali Mathur, I am currently working as an Education Counselor at SSSi Online Tutoring Services. SSSI Tutoring Services is a prominent online coaching platform that provides Best online Tuition for class 1 to 12 in all the subjects
…
continue reading
#1 Cambridge University BNOC (Big Name on Campus) - and ‘the best interviewer [The Body Coach Joe Wicks] has ever met’ - I am striving to give guests legacy-worthy interviews that listeners can enjoy while cooking, commuting, relaxing, or walking their pets. 10000+ Downloads in the first year. Please nominate a guest now: https://linktr.ee/DavidQuan
…
continue reading
India’s leading publisher of educational books. They publish help books for all the leading boards like CBSE, ISC, ICSE, Karnataka. Apart from this, Oswaal Books also publish books for some of the most important national-level competitive exams like JEE – Mains & Advanced, NEET, RRB-NTPC, CAT, CLAT, etc. Official Website: - https://www.amazon.in/stores/page/C25CEB9D-D428-4AB0-865B-6CB330B0C7F5
…
continue reading
Three of our 5th graders, Abby, Elly, and Jerrick, recently competed at the Science Olympiad. They sit down to talk about their experiences in this hands-on Science fun!By Kenneth Blum
…
continue reading
1
[QA] Can Language Models Solve Olympiad Programming?
11:43
11:43
Play later
Play later
Lists
Like
Liked
11:43
This paper introduces the USACO benchmark for evaluating language models on computing olympiad problems, highlighting challenges and proposing novel inference methods. https://arxiv.org/abs//2404.10952 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arx…
…
continue reading
1
[QA] OpenELM: An Efficient Language Model Family with Open-source Training and Inference Framework
8:13
8:13
Play later
Play later
Lists
Like
Liked
8:13
OpenELM, a state-of-the-art open language model, enhances accuracy using layer-wise scaling. Released with complete training framework, it empowers open research community. Available on GitHub and HuggingFace. https://arxiv.org/abs//2404.14619 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts:…
…
continue reading
1
OpenELM: An Efficient Language Model Family with Open-source Training and Inference Framework
8:58
8:58
Play later
Play later
Lists
Like
Liked
8:58
OpenELM, a state-of-the-art open language model, enhances accuracy using layer-wise scaling. Released with complete training framework, it empowers open research community. Available on GitHub and HuggingFace. https://arxiv.org/abs//2404.14619 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts:…
…
continue reading
1
[QA] Achieving >97% on GSM8K: Deeply Understanding the Problems Makes LLMs Perfect Reasoners
7:30
7:30
Play later
Play later
Lists
Like
Liked
7:30
The paper introduces DUP prompting strategy to improve Large Language Models' performance on complex reasoning tasks, outperforming Zero-Shot CoT on diverse datasets, achieving state-of-the-art results. https://arxiv.org/abs//2404.14963 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https:…
…
continue reading
1
Achieving >97% on GSM8K: Deeply Understanding the Problems Makes LLMs Perfect Reasoners
10:56
10:56
Play later
Play later
Lists
Like
Liked
10:56
The paper introduces DUP prompting strategy to improve Large Language Models' performance on complex reasoning tasks, outperforming Zero-Shot CoT on diverse datasets, achieving state-of-the-art results. https://arxiv.org/abs//2404.14963 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https:…
…
continue reading
1
[QA] SnapKV: LLM Knows What You are Looking for Before Generation
10:00
10:00
Play later
Play later
Lists
Like
Liked
10:00
SnapKV is a fine-tuning-free method that efficiently reduces Key-Value cache size in Large Language Models, maintaining performance while enhancing memory and time efficiency for long input sequences. https://arxiv.org/abs//2404.14469 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
…
continue reading
1
SnapKV: LLM Knows What You are Looking for Before Generation
17:09
17:09
Play later
Play later
Lists
Like
Liked
17:09
SnapKV is a fine-tuning-free method that efficiently reduces Key-Value cache size in Large Language Models, maintaining performance while enhancing memory and time efficiency for long input sequences. https://arxiv.org/abs//2404.14469 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
…
continue reading
MH-MoE addresses low expert activation and lack of fine-grained analysis in SMoE by using a multi-head mechanism to enhance context understanding and expert activation. https://arxiv.org/abs//2404.15045 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/ar…
…
continue reading
MH-MoE addresses low expert activation and lack of fine-grained analysis in SMoE by using a multi-head mechanism to enhance context understanding and expert activation. https://arxiv.org/abs//2404.15045 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/ar…
…
continue reading
1
[QA] The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions
9:21
9:21
Play later
Play later
Lists
Like
Liked
9:21
LLMs are vulnerable to attacks due to equal priority given to all prompts. Proposed instruction hierarchy teaches models to ignore lower-priority instructions, enhancing robustness with minimal impact on capabilities. https://arxiv.org/abs//2404.13208 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple P…
…
continue reading
1
The Instruction Hierarchy: Training LLMs to Prioritize Privileged Instructions
13:06
13:06
Play later
Play later
Lists
Like
Liked
13:06
LLMs are vulnerable to attacks due to equal priority given to all prompts. Proposed instruction hierarchy teaches models to ignore lower-priority instructions, enhancing robustness with minimal impact on capabilities. https://arxiv.org/abs//2404.13208 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple P…
…
continue reading
1
[QA] Preference Fine-Tuning of LLMs Should Leverage Suboptimal, On-Policy Data
9:37
9:37
Play later
Play later
Lists
Like
Liked
9:37
https://arxiv.org/abs//2404.14367 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
…
continue reading
1
Preference Fine-Tuning of LLMs Should Leverage Suboptimal, On-Policy Data
18:47
18:47
Play later
Play later
Lists
Like
Liked
18:47
https://arxiv.org/abs//2404.14367 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
…
continue reading
1
[QA] Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
9:02
9:02
Play later
Play later
Lists
Like
Liked
9:02
Introducing phi-3-mini, a high-performing language model trained on a large dataset, with smaller versions phi-3-small and phi-3-medium showing even better performance. https://arxiv.org/abs//2404.14219 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/ar…
…
continue reading
1
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
5:58
5:58
Play later
Play later
Lists
Like
Liked
5:58
Introducing phi-3-mini, a high-performing language model trained on a large dataset, with smaller versions phi-3-small and phi-3-medium showing even better performance. https://arxiv.org/abs//2404.14219 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/ar…
…
continue reading
1
[QA] Towards Reliable Latent Knowledge Estimation in LLMs: In-Context Learning vs. Prompting Based Factual Knowledge Extraction
8:22
8:22
Play later
Play later
Lists
Like
Liked
8:22
Approach estimates latent knowledge in large language models using in-context learning, showing differences in factual knowledge across models and sizes. https://arxiv.org/abs//2404.12957 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id16…
…
continue reading
1
Towards Reliable Latent Knowledge Estimation in LLMs: In-Context Learning vs. Prompting Based Factual Knowledge Extraction
23:51
23:51
Play later
Play later
Lists
Like
Liked
23:51
Approach estimates latent knowledge in large language models using in-context learning, showing differences in factual knowledge across models and sizes. https://arxiv.org/abs//2404.12957 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id16…
…
continue reading
1
[QA] HalluciBot: Is There No Such Thing as a Bad Question?
11:45
11:45
Play later
Play later
Lists
Like
Liked
11:45
HalluciBot predicts hallucination probability before generation in Large Language Models, aiding in query quality assessment and user accountability, potentially reducing computational waste. https://arxiv.org/abs//2404.12535 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.…
…
continue reading
1
HalluciBot: Is There No Such Thing as a Bad Question?
14:33
14:33
Play later
Play later
Lists
Like
Liked
14:33
HalluciBot predicts hallucination probability before generation in Large Language Models, aiding in query quality assessment and user accountability, potentially reducing computational waste. https://arxiv.org/abs//2404.12535 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.…
…
continue reading
1
[QA] Stronger Random Baselines for In-Context Learning
9:48
9:48
Play later
Play later
Lists
Like
Liked
9:48
Evaluating language models' in-context learning performance faces challenges. A stronger random baseline is proposed, improving evaluation accuracy and predicting held-out performance effectively. https://arxiv.org/abs//2404.13020 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podc…
…
continue reading
1
Stronger Random Baselines for In-Context Learning
14:58
14:58
Play later
Play later
Lists
Like
Liked
14:58
Evaluating language models' in-context learning performance faces challenges. A stronger random baseline is proposed, improving evaluation accuracy and predicting held-out performance effectively. https://arxiv.org/abs//2404.13020 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podc…
…
continue reading
1
[QA] Is DPO Superior to PPO for LLM Alignment? A Comprehensive Study
6:51
6:51
Play later
Play later
Lists
Like
Liked
6:51
The paper compares Direct Preference Optimization (DPO) and Proximal Policy Optimization (PPO) in aligning large language models with human feedback, showing PPO outperforms DPO in various RLHF testbeds. https://arxiv.org/abs//2404.10719 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https…
…
continue reading
1
Is DPO Superior to PPO for LLM Alignment? A Comprehensive Study
12:38
12:38
Play later
Play later
Lists
Like
Liked
12:38
The paper compares Direct Preference Optimization (DPO) and Proximal Policy Optimization (PPO) in aligning large language models with human feedback, showing PPO outperforms DPO in various RLHF testbeds. https://arxiv.org/abs//2404.10719 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https…
…
continue reading
1
[QA] The Illusion of State in State-Space Models
7:50
7:50
Play later
Play later
Lists
Like
Liked
7:50
State-space models (SSMs) are not more expressive than transformers for state tracking due to limitations in computational complexity, as shown through analysis and experiments. https://arxiv.org/abs//2404.08819 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/p…
…
continue reading
1
The Illusion of State in State-Space Models
19:21
19:21
Play later
Play later
Lists
Like
Liked
19:21
State-space models (SSMs) are not more expressive than transformers for state tracking due to limitations in computational complexity, as shown through analysis and experiments. https://arxiv.org/abs//2404.08819 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/p…
…
continue reading
1
[QA] Chinchilla Scaling: A replication attempt
7:55
7:55
Play later
Play later
Lists
Like
Liked
7:55
Hoffmann et al. (2022) propose three methods for estimating a compute-optimal scaling law. Replication of their third method reveals inconsistencies and implausibly narrow confidence intervals. https://arxiv.org/abs//2404.10102 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcast…
…
continue reading
Hoffmann et al. (2022) propose three methods for estimating a compute-optimal scaling law. Replication of their third method reveals inconsistencies and implausibly narrow confidence intervals. https://arxiv.org/abs//2404.10102 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcast…
…
continue reading
1
[QA] From R to Q: Your Language Model is Secretly a Q-Function
8:06
8:06
Play later
Play later
Lists
Like
Liked
8:06
The paper addresses the mismatch between Direct Preference Optimization (DPO) and standard Reinforcement Learning From Human Feedback (RLHF) setups, proposing a token-level approach for improved performance. https://arxiv.org/abs//2404.12358 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: h…
…
continue reading
1
From R to Q: Your Language Model is Secretly a Q-Function
15:17
15:17
Play later
Play later
Lists
Like
Liked
15:17
The paper addresses the mismatch between Direct Preference Optimization (DPO) and standard Reinforcement Learning From Human Feedback (RLHF) setups, proposing a token-level approach for improved performance. https://arxiv.org/abs//2404.12358 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: h…
…
continue reading
1
[QA] Toward Self-Improvement of LLMs via Imagination, Searching, and Criticizing
9:38
9:38
Play later
Play later
Lists
Like
Liked
9:38
ALPHALLM integrates Monte Carlo Tree Search with Large Language Models for self-improvement, enhancing reasoning abilities without additional annotations, addressing challenges in complex tasks. https://arxiv.org/abs//2404.12253 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
…
continue reading
1
Toward Self-Improvement of LLMs via Imagination, Searching, and Criticizing
22:01
22:01
Play later
Play later
Lists
Like
Liked
22:01
ALPHALLM integrates Monte Carlo Tree Search with Large Language Models for self-improvement, enhancing reasoning abilities without additional annotations, addressing challenges in complex tasks. https://arxiv.org/abs//2404.12253 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
…
continue reading
1
[QA] Dynamic Typography: Bringing Text to Life via Video Diffusion Prior
11:18
11:18
Play later
Play later
Lists
Like
Liked
11:18
Automated Dynamic Typography scheme deforms letters to convey meaning and adds vibrant movements based on user prompts, maintaining legibility and coherence in text animations. https://arxiv.org/abs//2404.11614 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/po…
…
continue reading
1
Dynamic Typography: Bringing Text to Life via Video Diffusion Prior
16:35
16:35
Play later
Play later
Lists
Like
Liked
16:35
Automated Dynamic Typography scheme deforms letters to convey meaning and adds vibrant movements based on user prompts, maintaining legibility and coherence in text animations. https://arxiv.org/abs//2404.11614 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/po…
…
continue reading
1
[QA] TRIFORCE: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding
8:12
8:12
Play later
Play later
Lists
Like
Liked
8:12
TRIFORCE introduces a hierarchical speculative decoding system to improve efficiency in long-sequence generation with large language models, achieving impressive speedups and scalability while maintaining generation quality. https://arxiv.org/abs//2404.11912 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers …
…
continue reading
1
TRIFORCE: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding
15:07
15:07
Play later
Play later
Lists
Like
Liked
15:07
TRIFORCE introduces a hierarchical speculative decoding system to improve efficiency in long-sequence generation with large language models, achieving impressive speedups and scalability while maintaining generation quality. https://arxiv.org/abs//2404.11912 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers …
…
continue reading
1
[QA] Reka Core, Flash, and Edge: A Series of Powerful Multimodal Language Models
8:55
8:55
Play later
Play later
Lists
Like
Liked
8:55
Reka introduces powerful multimodal language models - Core, Flash, and Edge - outperforming larger models in various tasks, approaching state-of-the-art performance. https://arxiv.org/abs//2404.12387 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv…
…
continue reading
1
Reka Core, Flash, and Edge: A Series of Powerful Multimodal Language Models
14:17
14:17
Play later
Play later
Lists
Like
Liked
14:17
Reka introduces powerful multimodal language models - Core, Flash, and Edge - outperforming larger models in various tasks, approaching state-of-the-art performance. https://arxiv.org/abs//2404.12387 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv…
…
continue reading
1
[QA] BLINK: Multimodal Large Language Models Can See but Not Perceive
10:54
10:54
Play later
Play later
Lists
Like
Liked
10:54
BLINK introduces a benchmark for multimodal language models focusing on visual perception tasks challenging for current models, with human accuracy significantly outperforming existing LLMs. https://arxiv.org/abs//2404.12390 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.a…
…
continue reading
1
BLINK: Multimodal Large Language Models Can See but Not Perceive
13:10
13:10
Play later
Play later
Lists
Like
Liked
13:10
BLINK introduces a benchmark for multimodal language models focusing on visual perception tasks challenging for current models, with human accuracy significantly outperforming existing LLMs. https://arxiv.org/abs//2404.12390 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.a…
…
continue reading
1
[QA] Fewer Truncations Improve Language Modeling
7:40
7:40
Play later
Play later
Lists
Like
Liked
7:40
Best-fit Packing method optimizes large language model training by packing documents into training sequences without unnecessary truncations, improving model coherence and performance significantly. https://arxiv.org/abs//2404.10830 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://po…
…
continue reading
1
Fewer Truncations Improve Language Modeling
12:14
12:14
Play later
Play later
Lists
Like
Liked
12:14
Best-fit Packing method optimizes large language model training by packing documents into training sequences without unnecessary truncations, improving model coherence and performance significantly. https://arxiv.org/abs//2404.10830 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://po…
…
continue reading
https://arxiv.org/abs//2404.11018 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
…
continue reading
https://arxiv.org/abs//2404.11018 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
…
continue reading
1
Can Language Models Solve Olympiad Programming?
14:18
14:18
Play later
Play later
Lists
Like
Liked
14:18
This paper introduces the USACO benchmark for evaluating language models on computing olympiad problems, highlighting challenges and proposing novel inference methods. https://arxiv.org/abs//2404.10952 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arx…
…
continue reading
1
[QA] Social Choice for Al Alignment: Dealing with Diverse Human Feedback
8:29
8:29
Play later
Play later
Lists
Like
Liked
8:29
The paper explores fine-tuning foundation models like GPT-4 to avoid problematic behavior, focusing on aggregating human input for collective preferences using social choice theory. https://arxiv.org/abs//2404.10271 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
…
continue reading
1
[short] Social Choice for Al Alignment: Dealing with Diverse Human Feedback
2:21
2:21
Play later
Play later
Lists
Like
Liked
2:21
The paper explores fine-tuning foundation models like GPT-4 to avoid problematic behavior, focusing on aggregating human input for collective preferences using social choice theory. https://arxiv.org/abs//2404.10271 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
…
continue reading
1
Social Choice for Al Alignment: Dealing with Diverse Human Feedback
23:07
23:07
Play later
Play later
Lists
Like
Liked
23:07
The paper explores fine-tuning foundation models like GPT-4 to avoid problematic behavior, focusing on aggregating human input for collective preferences using social choice theory. https://arxiv.org/abs//2404.10271 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
…
continue reading
1
[QA] Self-playing Adversarial Language Game Enhances LLM Reasoning
9:41
9:41
Play later
Play later
Lists
Like
Liked
9:41
The paper explores self-play training of large language models in an adversarial language game to enhance reasoning ability, showing performance improvement on reasoning benchmarks. https://arxiv.org/abs//2404.10642 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
…
continue reading
1
[short] Self-playing Adversarial Language Game Enhances LLM Reasoning
2:33
2:33
Play later
Play later
Lists
Like
Liked
2:33
The paper explores self-play training of large language models in an adversarial language game to enhance reasoning ability, showing performance improvement on reasoning benchmarks. https://arxiv.org/abs//2404.10642 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
…
continue reading