show episodes
 
Artwork

1
C-Squared Podcast

Fabiano Caruana & Cristian Chirila

Unsubscribe
Unsubscribe
Monthly+
 
The C-Squared Podcast is an in depth weekly discussion about the chess world with your hosts, Grandmasters Fabiano Caruana and Cristian Chirila. Bringing the biggest news and chess stars to your doorstep!
  continue reading
 
Geek Mythology Podcast brings you all of the latest news in TV, Movies and Gaming! Join hosts Toby, Tara, Lel and Andrew as they jump into the world of geek and discuss all of the latest ongoing topics that surround the industry that’s kept us on the couch for years. Themed episodes every week, a new geeky discussion topic each episode and the 2014 Geek Mythology Olympiad are just some of the things you can expect in the Geek Mythology Podcast. What are you waiting for? Jump in.
  continue reading
 
Artwork
 
Running out of time to catch up with new arXiv papers? We take the most impactful papers and present them as convenient podcasts. If you're a visual learner, we offer these papers in an engaging video format. Our service fills the gap between overly brief paper summaries and time-consuming full paper reads. You gain academic insights in a time-efficient, digestible format. Code behind this work: https://github.com/imelnyk/ArxivPapers Support this podcast: https://podcasters.spotify.com/pod/s ...
  continue reading
 
Artwork

1
WordQuest Devon

WordQuest Devon

Unsubscribe
Unsubscribe
Monthly
 
Wordquest Devon is a project of Aune Head Arts, Cyprus Well, Devon Libraries (Devon County Council), and the University of Exeter. It began life as a response to the 'questing' themes of the Cultural Olympiad, and still retains many of these playful and game-playing aspects. The project's live presence began in June 2011 and continues until September 2012. Around the County (and perhaps beyond) you can pick up copies of our 'Literary Map for Devon' - but you can also explore the online map h ...
  continue reading
 
Artwork

1
CampCast

Charles Adolphy

Unsubscribe
Unsubscribe
Monthly
 
Counsellor Charlie brings you the latest from Camp Eden Woods in Harcourt, Ontario. Listen in to find out about camp news, funny stories and hear from the campers themselves!
  continue reading
 
A SOLO school going TEENAGER takes on the challenge of delivering the GREATEST podcast to mankind. To see how he conquers the podcast world tune in to see a new episode EVERY WEEK on EVERY PLATFORM. Witness greatness!!!!
  continue reading
 
My Name is Anjali Mathur, I am currently working as an Education Counselor at SSSi Online Tutoring Services. SSSI Tutoring Services is a prominent online coaching platform that provides Best online Tuition for class 1 to 12 in all the subjects
  continue reading
 
#1 Cambridge University BNOC (Big Name on Campus) - and ‘the best interviewer [The Body Coach Joe Wicks] has ever met’ - I am striving to give guests legacy-worthy interviews that listeners can enjoy while cooking, commuting, relaxing, or walking their pets. 10000+ Downloads in the first year. Please nominate a guest now: https://linktr.ee/DavidQuan
  continue reading
 
Artwork
 
India’s leading publisher of educational books. They publish help books for all the leading boards like CBSE, ISC, ICSE, Karnataka. Apart from this, Oswaal Books also publish books for some of the most important national-level competitive exams like JEE – Mains & Advanced, NEET, RRB-NTPC, CAT, CLAT, etc. Official Website: - https://www.amazon.in/stores/page/C25CEB9D-D428-4AB0-865B-6CB330B0C7F5
  continue reading
 
Loading …
show series
 
This paper introduces the USACO benchmark for evaluating language models on computing olympiad problems, highlighting challenges and proposing novel inference methods. https://arxiv.org/abs//2404.10952 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arx…
  continue reading
 
OpenELM, a state-of-the-art open language model, enhances accuracy using layer-wise scaling. Released with complete training framework, it empowers open research community. Available on GitHub and HuggingFace. https://arxiv.org/abs//2404.14619 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts:…
  continue reading
 
OpenELM, a state-of-the-art open language model, enhances accuracy using layer-wise scaling. Released with complete training framework, it empowers open research community. Available on GitHub and HuggingFace. https://arxiv.org/abs//2404.14619 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts:…
  continue reading
 
The paper introduces DUP prompting strategy to improve Large Language Models' performance on complex reasoning tasks, outperforming Zero-Shot CoT on diverse datasets, achieving state-of-the-art results. https://arxiv.org/abs//2404.14963 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https:…
  continue reading
 
The paper introduces DUP prompting strategy to improve Large Language Models' performance on complex reasoning tasks, outperforming Zero-Shot CoT on diverse datasets, achieving state-of-the-art results. https://arxiv.org/abs//2404.14963 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https:…
  continue reading
 
SnapKV is a fine-tuning-free method that efficiently reduces Key-Value cache size in Large Language Models, maintaining performance while enhancing memory and time efficiency for long input sequences. https://arxiv.org/abs//2404.14469 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
  continue reading
 
SnapKV is a fine-tuning-free method that efficiently reduces Key-Value cache size in Large Language Models, maintaining performance while enhancing memory and time efficiency for long input sequences. https://arxiv.org/abs//2404.14469 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
  continue reading
 
MH-MoE addresses low expert activation and lack of fine-grained analysis in SMoE by using a multi-head mechanism to enhance context understanding and expert activation. https://arxiv.org/abs//2404.15045 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/ar…
  continue reading
 
MH-MoE addresses low expert activation and lack of fine-grained analysis in SMoE by using a multi-head mechanism to enhance context understanding and expert activation. https://arxiv.org/abs//2404.15045 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/ar…
  continue reading
 
LLMs are vulnerable to attacks due to equal priority given to all prompts. Proposed instruction hierarchy teaches models to ignore lower-priority instructions, enhancing robustness with minimal impact on capabilities. https://arxiv.org/abs//2404.13208 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple P…
  continue reading
 
LLMs are vulnerable to attacks due to equal priority given to all prompts. Proposed instruction hierarchy teaches models to ignore lower-priority instructions, enhancing robustness with minimal impact on capabilities. https://arxiv.org/abs//2404.13208 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple P…
  continue reading
 
https://arxiv.org/abs//2404.14367 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
https://arxiv.org/abs//2404.14367 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
Introducing phi-3-mini, a high-performing language model trained on a large dataset, with smaller versions phi-3-small and phi-3-medium showing even better performance. https://arxiv.org/abs//2404.14219 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/ar…
  continue reading
 
Introducing phi-3-mini, a high-performing language model trained on a large dataset, with smaller versions phi-3-small and phi-3-medium showing even better performance. https://arxiv.org/abs//2404.14219 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/ar…
  continue reading
 
Approach estimates latent knowledge in large language models using in-context learning, showing differences in factual knowledge across models and sizes. https://arxiv.org/abs//2404.12957 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id16…
  continue reading
 
Approach estimates latent knowledge in large language models using in-context learning, showing differences in factual knowledge across models and sizes. https://arxiv.org/abs//2404.12957 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id16…
  continue reading
 
HalluciBot predicts hallucination probability before generation in Large Language Models, aiding in query quality assessment and user accountability, potentially reducing computational waste. https://arxiv.org/abs//2404.12535 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.…
  continue reading
 
HalluciBot predicts hallucination probability before generation in Large Language Models, aiding in query quality assessment and user accountability, potentially reducing computational waste. https://arxiv.org/abs//2404.12535 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.…
  continue reading
 
Evaluating language models' in-context learning performance faces challenges. A stronger random baseline is proposed, improving evaluation accuracy and predicting held-out performance effectively. https://arxiv.org/abs//2404.13020 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podc…
  continue reading
 
Evaluating language models' in-context learning performance faces challenges. A stronger random baseline is proposed, improving evaluation accuracy and predicting held-out performance effectively. https://arxiv.org/abs//2404.13020 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podc…
  continue reading
 
The paper compares Direct Preference Optimization (DPO) and Proximal Policy Optimization (PPO) in aligning large language models with human feedback, showing PPO outperforms DPO in various RLHF testbeds. https://arxiv.org/abs//2404.10719 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https…
  continue reading
 
The paper compares Direct Preference Optimization (DPO) and Proximal Policy Optimization (PPO) in aligning large language models with human feedback, showing PPO outperforms DPO in various RLHF testbeds. https://arxiv.org/abs//2404.10719 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https…
  continue reading
 
State-space models (SSMs) are not more expressive than transformers for state tracking due to limitations in computational complexity, as shown through analysis and experiments. https://arxiv.org/abs//2404.08819 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/p…
  continue reading
 
State-space models (SSMs) are not more expressive than transformers for state tracking due to limitations in computational complexity, as shown through analysis and experiments. https://arxiv.org/abs//2404.08819 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/p…
  continue reading
 
Hoffmann et al. (2022) propose three methods for estimating a compute-optimal scaling law. Replication of their third method reveals inconsistencies and implausibly narrow confidence intervals. https://arxiv.org/abs//2404.10102 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcast…
  continue reading
 
Hoffmann et al. (2022) propose three methods for estimating a compute-optimal scaling law. Replication of their third method reveals inconsistencies and implausibly narrow confidence intervals. https://arxiv.org/abs//2404.10102 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcast…
  continue reading
 
The paper addresses the mismatch between Direct Preference Optimization (DPO) and standard Reinforcement Learning From Human Feedback (RLHF) setups, proposing a token-level approach for improved performance. https://arxiv.org/abs//2404.12358 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: h…
  continue reading
 
The paper addresses the mismatch between Direct Preference Optimization (DPO) and standard Reinforcement Learning From Human Feedback (RLHF) setups, proposing a token-level approach for improved performance. https://arxiv.org/abs//2404.12358 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: h…
  continue reading
 
ALPHALLM integrates Monte Carlo Tree Search with Large Language Models for self-improvement, enhancing reasoning abilities without additional annotations, addressing challenges in complex tasks. https://arxiv.org/abs//2404.12253 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
  continue reading
 
ALPHALLM integrates Monte Carlo Tree Search with Large Language Models for self-improvement, enhancing reasoning abilities without additional annotations, addressing challenges in complex tasks. https://arxiv.org/abs//2404.12253 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcas…
  continue reading
 
Automated Dynamic Typography scheme deforms letters to convey meaning and adds vibrant movements based on user prompts, maintaining legibility and coherence in text animations. https://arxiv.org/abs//2404.11614 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/po…
  continue reading
 
Automated Dynamic Typography scheme deforms letters to convey meaning and adds vibrant movements based on user prompts, maintaining legibility and coherence in text animations. https://arxiv.org/abs//2404.11614 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/po…
  continue reading
 
TRIFORCE introduces a hierarchical speculative decoding system to improve efficiency in long-sequence generation with large language models, achieving impressive speedups and scalability while maintaining generation quality. https://arxiv.org/abs//2404.11912 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers …
  continue reading
 
TRIFORCE introduces a hierarchical speculative decoding system to improve efficiency in long-sequence generation with large language models, achieving impressive speedups and scalability while maintaining generation quality. https://arxiv.org/abs//2404.11912 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers …
  continue reading
 
Reka introduces powerful multimodal language models - Core, Flash, and Edge - outperforming larger models in various tasks, approaching state-of-the-art performance. https://arxiv.org/abs//2404.12387 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv…
  continue reading
 
Reka introduces powerful multimodal language models - Core, Flash, and Edge - outperforming larger models in various tasks, approaching state-of-the-art performance. https://arxiv.org/abs//2404.12387 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv…
  continue reading
 
BLINK introduces a benchmark for multimodal language models focusing on visual perception tasks challenging for current models, with human accuracy significantly outperforming existing LLMs. https://arxiv.org/abs//2404.12390 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.a…
  continue reading
 
BLINK introduces a benchmark for multimodal language models focusing on visual perception tasks challenging for current models, with human accuracy significantly outperforming existing LLMs. https://arxiv.org/abs//2404.12390 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.a…
  continue reading
 
Best-fit Packing method optimizes large language model training by packing documents into training sequences without unnecessary truncations, improving model coherence and performance significantly. https://arxiv.org/abs//2404.10830 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://po…
  continue reading
 
Best-fit Packing method optimizes large language model training by packing documents into training sequences without unnecessary truncations, improving model coherence and performance significantly. https://arxiv.org/abs//2404.10830 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://po…
  continue reading
 
https://arxiv.org/abs//2404.11018 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
https://arxiv.org/abs//2404.11018 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
This paper introduces the USACO benchmark for evaluating language models on computing olympiad problems, highlighting challenges and proposing novel inference methods. https://arxiv.org/abs//2404.10952 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arx…
  continue reading
 
The paper explores fine-tuning foundation models like GPT-4 to avoid problematic behavior, focusing on aggregating human input for collective preferences using social choice theory. https://arxiv.org/abs//2404.10271 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
The paper explores fine-tuning foundation models like GPT-4 to avoid problematic behavior, focusing on aggregating human input for collective preferences using social choice theory. https://arxiv.org/abs//2404.10271 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
The paper explores fine-tuning foundation models like GPT-4 to avoid problematic behavior, focusing on aggregating human input for collective preferences using social choice theory. https://arxiv.org/abs//2404.10271 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
The paper explores self-play training of large language models in an adversarial language game to enhance reasoning ability, showing performance improvement on reasoning benchmarks. https://arxiv.org/abs//2404.10642 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
The paper explores self-play training of large language models in an adversarial language game to enhance reasoning ability, showing performance improvement on reasoning benchmarks. https://arxiv.org/abs//2404.10642 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
Loading …

Quick Reference Guide