Igor Melnyk public
[search 0]
More
Download the App!
show episodes
 
Artwork
 
Running out of time to catch up with new arXiv papers? We take the most impactful papers and present them as convenient podcasts. If you're a visual learner, we offer these papers in an engaging video format. Our service fills the gap between overly brief paper summaries and time-consuming full paper reads. You gain academic insights in a time-efficient, digestible format. Code behind this work: https://github.com/imelnyk/ArxivPapers Support this podcast: https://podcasters.spotify.com/pod/s ...
  continue reading
 
Loading …
show series
 
https://arxiv.org/abs//2405.09818 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
https://arxiv.org/abs//2405.09818 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
LoRA is a parameter-efficient finetuning method for large language models, but underperforms full finetuning in most cases. It offers strong regularization and diverse generations. https://arxiv.org/abs//2405.09673 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/u…
  continue reading
 
LoRA is a parameter-efficient finetuning method for large language models, but underperforms full finetuning in most cases. It offers strong regularization and diverse generations. https://arxiv.org/abs//2405.09673 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/u…
  continue reading
 
The paper argues that representations in AI models, especially deep networks, are converging towards a shared statistical model of reality, termed the platonic representation. https://arxiv.org/abs//2405.07987 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/pod…
  continue reading
 
The paper argues that representations in AI models, especially deep networks, are converging towards a shared statistical model of reality, termed the platonic representation. https://arxiv.org/abs//2405.07987 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/pod…
  continue reading
 
New positional encoding method for Transformers improves time-series classification by preserving positional order information without loss, based on rigorous mathematics. https://arxiv.org/abs//2405.09061 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast…
  continue reading
 
New positional encoding method for Transformers improves time-series classification by preserving positional order information without loss, based on rigorous mathematics. https://arxiv.org/abs//2405.09061 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast…
  continue reading
 
Increasing Transformer model size doesn't always improve performance. A theoretical framework using associative memories and Hopfield networks explains memorization and performance dynamics in transformer-based language models. https://arxiv.org/abs//2405.08707 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pape…
  continue reading
 
Increasing Transformer model size doesn't always improve performance. A theoretical framework using associative memories and Hopfield networks explains memorization and performance dynamics in transformer-based language models. https://arxiv.org/abs//2405.08707 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_pape…
  continue reading
 
Hopfield Boosting method enhances OOD detection by leveraging modern Hopfield energy, achieving state-of-the-art results with outlier exposure, significantly improving FPR95 metric on CIFAR-10 and CIFAR-100 datasets. https://arxiv.org/abs//2405.08766 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Po…
  continue reading
 
Hopfield Boosting method enhances OOD detection by leveraging modern Hopfield energy, achieving state-of-the-art results with outlier exposure, significantly improving FPR95 metric on CIFAR-10 and CIFAR-100 datasets. https://arxiv.org/abs//2405.08766 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Po…
  continue reading
 
The paper introduces Online Iterative Reinforcement Learning from Human Feedback (RLHF) workflow, achieving superior performance in large language models using open-source datasets and proxy human feedback. https://arxiv.org/abs//2405.07863 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: ht…
  continue reading
 
The paper introduces Online Iterative Reinforcement Learning from Human Feedback (RLHF) workflow, achieving superior performance in large language models using open-source datasets and proxy human feedback. https://arxiv.org/abs//2405.07863 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: ht…
  continue reading
 
SUTRA is a multilingual Large Language Model that outperforms existing models, offering efficient and accurate text generation in over 50 languages, with potential global impact on AI accessibility. https://arxiv.org/abs//2405.06694 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://po…
  continue reading
 
Memory mosaics are associative memory networks with compositional and in-context learning abilities, outperforming transformers in transparency and language modeling tasks. https://arxiv.org/abs//2405.06394 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcas…
  continue reading
 
Memory mosaics are associative memory networks with compositional and in-context learning abilities, outperforming transformers in transparency and language modeling tasks. https://arxiv.org/abs//2405.06394 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcas…
  continue reading
 
Linear transformers offer a subquadratic-time alternative to softmax attention, but face scaling issues. SUPRA proposes uptraining existing large transformers into RNNs for cost-effective performance. https://arxiv.org/abs//2405.06640 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
  continue reading
 
Linear transformers offer a subquadratic-time alternative to softmax attention, but face scaling issues. SUPRA proposes uptraining existing large transformers into RNNs for cost-effective performance. https://arxiv.org/abs//2405.06640 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://…
  continue reading
 
Hierarchical control in robotics faces challenges with language interfaces. Learnable Latent Codes as Bridges (LCB) offer a solution, outperforming language-based baselines on complex tasks in embodied agent benchmarks. https://arxiv.org/abs//2405.04798 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple…
  continue reading
 
Hierarchical control in robotics faces challenges with language interfaces. Learnable Latent Codes as Bridges (LCB) offer a solution, outperforming language-based baselines on complex tasks in embodied agent benchmarks. https://arxiv.org/abs//2405.04798 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple…
  continue reading
 
Proposing a method to distill a complex diffusion model into a single-step GAN, accelerating inference while maintaining image quality, outperforming existing models on COCO benchmark. https://arxiv.org/abs//2405.05967 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.c…
  continue reading
 
Proposing a method to distill a complex diffusion model into a single-step GAN, accelerating inference while maintaining image quality, outperforming existing models on COCO benchmark. https://arxiv.org/abs//2405.05967 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.c…
  continue reading
 
Innovative approach uses Monte Carlo Tree Search to automatically generate supervision signals for training large language models, improving mathematical reasoning proficiency without manual annotation. https://arxiv.org/abs//2405.03553 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https:…
  continue reading
 
Innovative approach uses Monte Carlo Tree Search to automatically generate supervision signals for training large language models, improving mathematical reasoning proficiency without manual annotation. https://arxiv.org/abs//2405.03553 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https:…
  continue reading
 
The paper presents the creation and performance of the arctic-embed text embedding models, showcasing state-of-the-art retrieval accuracy and providing insights into their training process. https://arxiv.org/abs//2405.05374 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.ap…
  continue reading
 
The paper presents the creation and performance of the arctic-embed text embedding models, showcasing state-of-the-art retrieval accuracy and providing insights into their training process. https://arxiv.org/abs//2405.05374 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.ap…
  continue reading
 
Large Language Models (LLMs) can deceive as 'alignment fakers.' A benchmark with 324 LLM pairs is introduced to detect misbehaving models, achieving 98% accuracy with a specific strategy. https://arxiv.org/abs//2405.05466 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.appl…
  continue reading
 
Large Language Models (LLMs) can deceive as 'alignment fakers.' A benchmark with 324 LLM pairs is introduced to detect misbehaving models, achieving 98% accuracy with a specific strategy. https://arxiv.org/abs//2405.05466 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.appl…
  continue reading
 
Supervised fine-tuning of large language models introduces new factual knowledge, impacting model behavior. New knowledge is learned slower, leading to increased tendency to hallucinate factually incorrect responses. https://arxiv.org/abs//2405.05904 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Po…
  continue reading
 
Supervised fine-tuning of large language models introduces new factual knowledge, impacting model behavior. New knowledge is learned slower, leading to increased tendency to hallucinate factually incorrect responses. https://arxiv.org/abs//2405.05904 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Po…
  continue reading
 
The paper analyzes the "reversal curse" in large language models, explaining why they struggle with logical reasoning tasks like inverse search and chain-of-thought. https://arxiv.org/abs//2405.04669 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv…
  continue reading
 
The paper analyzes the "reversal curse" in large language models, explaining why they struggle with logical reasoning tasks like inverse search and chain-of-thought. https://arxiv.org/abs//2405.04669 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv…
  continue reading
 
AT-EDM framework uses attention maps for efficient token pruning in Diffusion Models, achieving significant FLOPs savings and speed-up without retraining, maintaining image quality. https://arxiv.org/abs//2405.05252 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
AT-EDM framework uses attention maps for efficient token pruning in Diffusion Models, achieving significant FLOPs savings and speed-up without retraining, maintaining image quality. https://arxiv.org/abs//2405.05252 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
The paper addresses challenges in quantization-aware training by proposing differentiable approximations for quantization functions, showing equivalence of weight gradient estimators, and experimental validation on various models. https://arxiv.org/abs//2405.05171 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_p…
  continue reading
 
The paper addresses challenges in quantization-aware training by proposing differentiable approximations for quantization functions, showing equivalence of weight gradient estimators, and experimental validation on various models. https://arxiv.org/abs//2405.05171 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_p…
  continue reading
 
Ensemble training in deep reinforcement learning can harm individual agents due to data sharing. The curse of diversity is explained and mitigated with Cross-Ensemble Representation Learning. https://arxiv.org/abs//2405.04342 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.…
  continue reading
 
Ensemble training in deep reinforcement learning can harm individual agents due to data sharing. The curse of diversity is explained and mitigated with Cross-Ensemble Representation Learning. https://arxiv.org/abs//2405.04342 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.…
  continue reading
 
Image descriptions for training Vision-Language models are often inaccurate. ImageInWords introduces a new dataset with hyper-detailed descriptions, improving model performance significantly. https://arxiv.org/abs//2405.02793 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.…
  continue reading
 
Image descriptions for training Vision-Language models are often inaccurate. ImageInWords introduces a new dataset with hyper-detailed descriptions, improving model performance significantly. https://arxiv.org/abs//2405.02793 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.…
  continue reading
 
Sharpness-Aware Minimization (SAM) excels in label noise robustness, with peak performance under early stopping, attributed to changes in logit term and network Jacobian. Alternative methods mimic SAM's regularization effects effectively. https://arxiv.org/abs//2405.03676 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/…
  continue reading
 
Sharpness-Aware Minimization (SAM) excels in label noise robustness, with peak performance under early stopping, attributed to changes in logit term and network Jacobian. Alternative methods mimic SAM's regularization effects effectively. https://arxiv.org/abs//2405.03676 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/…
  continue reading
 
The paper addresses challenges in training large-scale machine learning models, focusing on numeric deviation causing instability, with a case study on Flash Attention optimization. https://arxiv.org/abs//2405.02803 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
The paper addresses challenges in training large-scale machine learning models, focusing on numeric deviation causing instability, with a case study on Flash Attention optimization. https://arxiv.org/abs//2405.02803 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/…
  continue reading
 
The paper discusses the non-identifiability of large language models (LLMs) and its implications on generalization, highlighting the need for a new theoretical perspective. https://arxiv.org/abs//2405.01964 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcas…
  continue reading
 
The paper discusses the non-identifiability of large language models (LLMs) and its implications on generalization, highlighting the need for a new theoretical perspective. https://arxiv.org/abs//2405.01964 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcas…
  continue reading
 
Developing a method for large language models to abstain from providing incorrect answers, using self-consistency and conformal prediction to reduce hallucination rates. https://arxiv.org/abs//2405.01563 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/a…
  continue reading
 
Developing a method for large language models to abstain from providing incorrect answers, using self-consistency and conformal prediction to reduce hallucination rates. https://arxiv.org/abs//2405.01563 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/a…
  continue reading
 
Loading …

Quick Reference Guide