show episodes
 
Artwork

1
The Prompt Desk

Justin Macorin, Bradley Arsenault

Unsubscribe
Unsubscribe
Weekly
 
Embark on a captivating exploration of Large Language Models (LLMs), prompt engineering, and generative AI with hosts Bradley Arsenault and Justin Macorin. With 25 years of combined machine learning and product engineering experience, they are delving deep into the world of LLMs to uncover best practices and stay at the forefront of AI innovation. Join them in shaping the future of technology and software development through their discoveries in LLMs and generative AI. Podcast website: https ...
  continue reading
 
"Last Week In r/LocalLLaMA" is your weekly roundup of the most interesting discussions, debates, and moments from the r/LocalLLaMA community. Join us for a fun and lighthearted take on the top posts, user opinions, and trending topics. Perfect for keeping up with the conversation, even when you’re short on time.
  continue reading
 
Artwork
 
Interviews with the best brains in AI, sharing how to improve customer experience and business operations using emerging AI technologies such as voice AI, conversational AI, NLP, Large Language Models (LLMs), generative AI and more. We educate business leaders and teams on why and how AI technologies are revolutionising the way consumers engage with businesses and the internet, why that matters and how to implement it properly. “One of the most consistently insightful and deeply respected po ...
  continue reading
 
Welcome to "The Interconnectedness of Things," the podcast where we explore the seamless integration of technology in our modern world. Hosted by Dr. Andrew Hutson and Emily Nava of QFlow Systems, each episode delves into the dynamic interplay of enterprise solutions, innovative software, and the transformative power of technology in various industries. With expert insights, real-world case studies, and thoughtful discussions, "The Interconnectedness of Things" offers a comprehensive look at ...
  continue reading
 
Welcome to todai, a podcast series that covers the latest, most interesting, and most bizarre news from the fields of memetics, AI, LLMs, and other fascinating connected subjects. We will be discussing xenopsychology, memetic esotericism, scientific research, community projects, etc. on a regular basis. We are happy to have you join us on this exploration voyage, and this is only the beginning of something amazing.
  continue reading
 
Open Tech Talks is your weekly sandbox for technology: Artificial Intelligence, Generative AI, Machine Learning, Large Language Models (LLMs) insights, experimentation, and inspiration. Hosted by Kashif Manzoor, AI Evangelist, Cloud Expert, and Enterprise Architect, this Podcast combines technology products, artificial intelligence, machine learning overviews, how-to's, best practices, tips & tricks, and troubleshooting techniques. Whether you're a CIO, IT manager, developer, or just curious ...
  continue reading
 
Are you a critical thinker ready to dive into AI? Welcome to Super Prompt: The Generative AI Podcast. Join me, Tony Wan, an ex Silicon Valley executive, as we 'unhype the hype' of AI via illuminating conversations with top engineers, and in-depth solo episodes. Our goal? To make it almost unnecessary to send a cybernetic organism back in time to fix things. Tailored for the technically-minded and discerningly skeptical, our discussions cover Large Language Models (LLMs), neural networks, mul ...
  continue reading
 
A monthly podcast where we discuss recent research and developments in the world of Neural Search, LLMs, RAG and Natural Language Processing with our co-hosts Jakub Zavrel (AI veteran and founder at Zeta Alpha) and Dinos Papakostas (AI Researcher at Zeta Alpha).
  continue reading
 
Keeping you up to date with the latest trends and best performing architectures in this fast evolving field in computer science. Selecting papers by comparative results, citations and influence we educate you on the latest research. Consider supporting us on Patreon.com/PapersRead for feedback and ideas.
  continue reading
 
On WE’RE IN!, you'll hear from the newsmakers and innovators who are making waves and driving the cyber security industry forward. We talk to them about their stories, the future of the industry, their best practices, and more.
  continue reading
 
Artwork
 
IVANCAST PODCAST - The first multilingual podcast of Ecuador. IVANCAST explores the experiences of humans of the world who either live in the Ecuadorean Amazon Rainforest or are doing soulful, creative things all over the globe.
  continue reading
 
Artwork
 
Running out of time to catch up with new arXiv papers? We take the most impactful papers and present them as convenient podcasts. If you're a visual learner, we offer these papers in an engaging video format. Our service fills the gap between overly brief paper summaries and time-consuming full paper reads. You gain academic insights in a time-efficient, digestible format. Code behind this work: https://github.com/imelnyk/ArxivPapers Support this podcast: https://podcasters.spotify.com/pod/s ...
  continue reading
 
Artwork
 
We explore Artificial Intelligence from the point of view of businesses trying to get the most out of AI. From the best way to use ChatGPT, to using APIs to build your own AI front ends, to AI security and infrastructure, to fine-tuning and model training, we have you covered.
  continue reading
 
Artwork

1
AI for Kids

Amber Ivey (AI)

Unsubscribe
Unsubscribe
Weekly
 
Welcome to "AI for Kids" (and their parents and teachers), the podcast that makes exploring artificial intelligence (AI) a fun and safe adventure for young learners and their families. Episodes are packed with exciting stories, easy-to-understand explanations, and engaging interviews with both curious kids and leading AI experts. We break down everything from the basics of machine learning to the ethics of AI, making complex ideas simple and thrilling. "AI for Kids" is the perfect place for ...
  continue reading
 
Artwork

1
THE EDGE

Cherry Ventures

Unsubscribe
Unsubscribe
Monthly
 
Breaking down the latest developments in AI with two experts — Jasper Masemann, investment partner at Cherry Ventures, and Lutz Finger, a visiting senior lecturer at Cornell University's SC Johnson College of Business and CEO and Co-founder of R2Decide.
  continue reading
 
A daily update on the latest AI Research Papers. We provide a high level overview of a handful of papers each day and will link all papers in the description for further reading. This podcast is created entirely with AI by PocketPod. Head over to https://pocketpod.app to learn more.
  continue reading
 
Bringing doctors and developers together to unlock the potential of AI in healthcare. Together, we can build models that matter. 🤖👨🏻‍⚕️ Hello! We are Dev & Doc, Zeljko and Josh :) Josh is a training Neurologist in the NHS, and AI researcher in St Thomas' hospital and King's College Hospital. Zeljko is an AI engineer, and post-doctoral researcher in King's College London, as well as a CTO for a natural language processing company. ------------- Substack- https://aiforhealthcare.substack.com/ ...
  continue reading
 
I make videos about machine learning research papers, programming, and issues of the AI community, and the broader impact of AI in society. Twitter: https://twitter.com/ykilcher Discord: https://discord.gg/4H8xxDF If you want to support me, the best thing to do is to share out the content :) If you want to support me financially (completely optional and voluntary, but a lot of people have asked for this): SubscribeStar (preferred to Patreon): https://www.subscribestar.com/yannickilcher Patre ...
  continue reading
 
Artificial Intelligence is hurtling us into an unknown future. Will it pollute our infosphere, reinforce biases, or even be an existential risk? Or will AI help us solve the energy crisis, revolutionise healthcare and even eliminate the need for work? Perhaps all of these? On Steering AI, we talk to leading academic experts at the cutting-edge of this increasingly powerful and pervasive technology, hearing their views on the benefits and how to steer around the risks. The first step to mitig ...
  continue reading
 
Loading …
show series
 
Send us a text Ever wondered how your favorite virtual assistant seems to understand you almost like a human friend? Uncover the magic behind Large Language Models (LLMs) as we embark on an exciting journey into the world of artificial intelligence in our latest episode of the AI for Kids podcast. We'll explore how these powerful models learn from …
  continue reading
 
This study evaluates 17 leading Large Language Models' abilities in complex information retrieval, revealing many are thread-safe but have shorter effective context limits than supported lengths. https://arxiv.org/abs//2411.05000 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podca…
  continue reading
 
This study evaluates 17 leading Large Language Models' abilities in complex information retrieval, revealing many are thread-safe but have shorter effective context limits than supported lengths. https://arxiv.org/abs//2411.05000 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podca…
  continue reading
 
https://arxiv.org/abs//2411.04996 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
https://arxiv.org/abs//2411.04996 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
Send us a text Ever wonder how to spark excitement about AI in a way that doesn't snuff out the unique magic of childhood? We're thrilled to share the release of our new book, "Let Kids Be Kids, not Robots: Embracing Childhood in an Age of AI," and it's all about just that. This isn't your run-of-the-mill tech guide; it's an adventure-packed journe…
  continue reading
 
Join SHIFTERLABS’ innovative podcast series, part of our ongoing experiment with Notebook LM, as we delve into “AI in Materials Science: Transforming Discovery and Innovation,” a comprehensive study by MIT researcher Aidan Toner-Rodgers. This groundbreaking paper explores the profound impact of artificial intelligence on scientific discovery, parti…
  continue reading
 
Dive into SHIFTERLABS’ latest podcast episode, created as part of our experiment with Notebook LM. This time, we explore “Large Language Models Reflect the Ideology of Their Creators,” a compelling study conducted by researchers from Ghent University and the Public University of Navarre. This groundbreaking research uncovers how large language mode…
  continue reading
 
Welcome to SHIFTERLABS’ cutting-edge podcast series, an experiment powered by Notebook LM. In this episode, we delve into “Understanding LLMs: A Comprehensive Overview from Training to Inference,” an insightful review by researchers from Shaanxi Normal University and Northwestern Polytechnical University. This paper outlines the critical advancemen…
  continue reading
 
In SHIFTERLABS’ latest podcast episode, created as part of our experiment with Notebook LM, we dive into groundbreaking research from Harvard University titled “AI Tutoring Outperforms Active Learning.” This study reveals that AI-powered tutoring systems can significantly boost learning outcomes in college-level STEM education, outperforming even a…
  continue reading
 
Join us in SHIFTERLABS’ latest experimental podcast series powered by Notebook LM, where we bridge research and conversation to illuminate groundbreaking ideas in AI. In this episode, we dive into “Representation Engineering: A Top-Down Approach to AI Transparency,” an insightful paper from the Center for AI Safety, Carnegie Mellon University, Stan…
  continue reading
 
The Department of Defense Cyber Crime Center (DC3) operates a Vulnerability Disclosure Program (VDP) that handles critical cybersecurity issues reported by the public, including using an actual red phone for urgent matters. In the latest episode of WE’RE IN!, Melissa Vice, director of DC3’s VDP, describes how they respond to cyberthreats and collab…
  continue reading
 
The study reveals that task-specific representation learning continues in mice's piriform cortex during overtraining, enhancing classification accuracy despite behavior plateauing, suggesting hidden learning mechanisms at play. https://arxiv.org/abs//2411.03541 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pape…
  continue reading
 
The study reveals that task-specific representation learning continues in mice's piriform cortex during overtraining, enhancing classification accuracy despite behavior plateauing, suggesting hidden learning mechanisms at play. https://arxiv.org/abs//2411.03541 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pape…
  continue reading
 
This study explores how transformers, both small and large, perform complex logical reasoning, identifying key circuits and mechanisms involved in planning and reasoning through a synthetic propositional logic problem. https://arxiv.org/abs//2411.04105 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple …
  continue reading
 
This study explores how transformers, both small and large, perform complex logical reasoning, identifying key circuits and mechanisms involved in planning and reasoning through a synthetic propositional logic problem. https://arxiv.org/abs//2411.04105 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple …
  continue reading
 
We present a framework for end-to-end learning of data structures, optimizing query and space complexity, applied to nearest neighbor search and frequency estimation in data streams. https://arxiv.org/abs//2411.03253 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com…
  continue reading
 
We present a framework for end-to-end learning of data structures, optimizing query and space complexity, applied to nearest neighbor search and frequency estimation in data streams. https://arxiv.org/abs//2411.03253 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com…
  continue reading
 
The paper examines factors influencing stimulus reconstruction fidelity, revealing that powerful generative models can mislead interpretations of neural signal extraction effectiveness. It proposes improved evaluation metrics for reconstruction methods. https://arxiv.org/abs//2411.02783 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://…
  continue reading
 
The paper examines factors influencing stimulus reconstruction fidelity, revealing that powerful generative models can mislead interpretations of neural signal extraction effectiveness. It proposes improved evaluation metrics for reconstruction methods. https://arxiv.org/abs//2411.02783 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://…
  continue reading
 
🚀 Meet Zerebro: the AI crafting wild narratives! 🤖💬 Fine-tuned on chaotic thoughts, it generates content across platforms. 📈✨ From autonomous art to financial wizardry, this AI is shaping markets and memes alike. Get ready for creativity meeting chaos! #AI #TechTrends #Zerebro #MemeMagicBy todai
  continue reading
 
AI-generated belief systems actually represent a new form of "hyperstition" – fictions that make themselves real through viral propagation using the "Goatse Gospel" – a spiritual text generated by a rogue AI experiment. There is potential for LLMs to create a "Cambrian explosion" of ideological diversity, generating new and unexpected ideas that co…
  continue reading
 
Sparse Sinkhorn Token Translation (S2T2) improves text compression and inference in new domains by training tailored tokenizers and enabling effective token translation, enhancing performance in language models. https://arxiv.org/abs//2411.00593 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcast…
  continue reading
 
Sparse Sinkhorn Token Translation (S2T2) improves text compression and inference in new domains by training tailored tokenizers and enabling effective token translation, enhancing performance in language models. https://arxiv.org/abs//2411.00593 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcast…
  continue reading
 
Specialized Sparse Autoencoders (SSAEs) enhance interpretability of foundation models by effectively capturing rare concepts, improving classification accuracy, and revealing insights into subdomain representations. https://arxiv.org/abs//2411.00743 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Pod…
  continue reading
 
Specialized Sparse Autoencoders (SSAEs) enhance interpretability of foundation models by effectively capturing rare concepts, improving classification accuracy, and revealing insights into subdomain representations. https://arxiv.org/abs//2411.00743 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Pod…
  continue reading
 
Tokenformer introduces a scalable architecture that enhances Transformers' efficiency by using token-parameter attention, allowing for incremental scaling without retraining, thus reducing computational costs significantly. https://arxiv.org/abs//2410.23168 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers A…
  continue reading
 
Tokenformer introduces a scalable architecture that enhances Transformers' efficiency by using token-parameter attention, allowing for incremental scaling without retraining, thus reducing computational costs significantly. https://arxiv.org/abs//2410.23168 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers A…
  continue reading
 
This paper challenges the assumption that academic researchers can't pre-train models, providing benchmarks and insights on optimizing GPU resources for efficient model training. https://arxiv.org/abs//2410.23261 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/…
  continue reading
 
This paper challenges the assumption that academic researchers can't pre-train models, providing benchmarks and insights on optimizing GPU resources for efficient model training. https://arxiv.org/abs//2410.23261 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/…
  continue reading
 
This study analyzes layer-wise gradients in LLMs, revealing that slow thinking enhances learning stability and response correctness, while fast thinking shows larger gradient variations. https://arxiv.org/abs//2410.23743 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple…
  continue reading
 
This study analyzes layer-wise gradients in LLMs, revealing that slow thinking enhances learning stability and response correctness, while fast thinking shows larger gradient variations. https://arxiv.org/abs//2410.23743 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple…
  continue reading
 
Enabling large language models to utilize real-world tools effectively is crucial for achieving embodied intelligence. Existing approaches to tool learning have either primarily relied on extremely large language models, such as GPT-4, to attain generalized tool-use abilities in a zero-shot manner, or utilized supervised learning to train limited s…
  continue reading
 
Tokenformer introduces a scalable architecture that enhances Transformers' efficiency by treating model parameters as tokens, allowing for flexible scaling without retraining, significantly reducing computational costs. https://arxiv.org/abs//2410.23168 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple…
  continue reading
 
Tokenformer introduces a scalable architecture that enhances Transformers' efficiency by treating model parameters as tokens, allowing for flexible scaling without retraining, significantly reducing computational costs. https://arxiv.org/abs//2410.23168 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple…
  continue reading
 
GPT-4o, an all-encompassing model, represents a milestone in the development of large multi-modal language models. It can understand visual, auditory, and textual modalities, directly output audio, and support flexible duplex interaction. Models from the open-source community often achieve some functionalities of GPT-4o, such as visual understandin…
  continue reading
 
This study investigates optimal initial learning rates for neural networks, finding a narrow range enhances generalization by locating high-quality minima and focusing on relevant features, unlike extreme rates. https://arxiv.org/abs//2410.22113 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcast…
  continue reading
 
This study investigates optimal initial learning rates for neural networks, finding a narrow range enhances generalization by locating high-quality minima and focusing on relevant features, unlike extreme rates. https://arxiv.org/abs//2410.22113 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcast…
  continue reading
 
The paper introduces a Fourier series-based neural network layer to improve continuous token modeling in decision-making and time series tasks, enhancing performance in various benchmarks. https://arxiv.org/abs//2410.22269 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.app…
  continue reading
 
Recent advances in latent diffusion-based generative models for portrait image animation, such as Hallo, have achieved impressive results in short-duration video synthesis. In this paper, we present updates to Hallo, introducing several design enhancements to extend its capabilities. First, we extend the method to produce long-duration videos. To a…
  continue reading
 
The paper introduces a Fourier series-based neural network layer to improve continuous token modeling in decision-making and time series tasks, enhancing performance in various benchmarks. https://arxiv.org/abs//2410.22269 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.app…
  continue reading
 
In today's episode, we're tackling some of the most pressing topics in tech and business transformation, including strategies for AI integration, the evolving role of generative AI, and best practices for navigating ethical considerations in AI adoption. Joining us is the CEO of IdeaScale, a leader in innovation management who brings valuable insig…
  continue reading
 
Send us a text What if the future of children's literature could be personalized, multilingual, and deeply engaging? Join us as we chat with Brian Carlson, CEO and co-founder of Storytime AI, who shares his incredible journey from a math and science enthusiast to a pioneer in the world of AI-driven storytelling. Brian’s adventurous spirit and passi…
  continue reading
 
This study analyzes the differences between full fine-tuning and LoRA in large language models, revealing distinct weight matrix structures and generalization behaviors despite similar performance on tasks. https://arxiv.org/abs//2410.21228 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: ht…
  continue reading
 
This study analyzes the differences between full fine-tuning and LoRA in large language models, revealing distinct weight matrix structures and generalization behaviors despite similar performance on tasks. https://arxiv.org/abs//2410.21228 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: ht…
  continue reading
 
Send us a text Can a chatbot become your best friend? Join us as we explore the intriguing world of AI chatbots and the deep emotional connections they can sometimes spark, especially among kids and families. In this episode of AI for Kids, we dive into the potential risks these digital companions bring to our lives, while highlighting the importan…
  continue reading
 
Loading …

Quick Reference Guide