Arx público
[search 0]
Más
Download the App!
show episodes
 
Artwork

1
Arxiv Papers

Igor Melnyk

Unsubscribe
Unsubscribe
Diariamente+
 
Running out of time to catch up with new arXiv papers? We take the most impactful papers and present them as convenient podcasts. If you're a visual learner, we offer these papers in an engaging video format. Our service fills the gap between overly brief paper summaries and time-consuming full paper reads. You gain academic insights in a time-efficient, digestible format. Code behind this work: https://github.com/imelnyk/ArxivPapers Support this podcast: https://podcasters.spotify.com/pod/s ...
  continue reading
 
Разнищваме заедно световната гейминг сцена в компанията на Борислав "Overneathe" Белев и Владислав "Deadset" Рашковски. На живо всяка сряда от 19:30 ч. на arx.bg/stream.
  continue reading
 
Des de la Plaça d’Osca, al 94.6 de la FM i també en streaming des d’onadesants.cat, cada dijous, de 10 a 11 de la nit, Sergi Pujol, Joel Díaz i Manel Vidal analitzen l’actualitat esportiva, en rigurós directe i sempre amb una cervesa a la mà. Tot presentat i moderat per l’Andreu Juanola. Els trobareu a lasotana.cat, a la cua de l’INEM, i també a twitter. A la producció i als piuets: Enric Gusó.
  continue reading
 
Magazin matinal de Vilassar Ràdio presentat per Jaume Cabot. L'actualitat local, comarcal i general i les entrevistes diàries a persones de tots els àmbits, centra l'atenció del programa. Compta amb una vintena de col·laboradors/es que parlen d'esports, teatre, cinema, gestió emocional, sexe, cuina, salut, consum, benestar femení, tarot, tertúlies d'avis i joves, etc.
  continue reading
 
Loading …
show series
 
The paper discusses the non-identifiability of large language models (LLMs) and its implications on generalization, highlighting the need for a new theoretical perspective. https://arxiv.org/abs//2405.01964 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcas…
  continue reading
 
The paper discusses the non-identifiability of large language models (LLMs) and its implications on generalization, highlighting the need for a new theoretical perspective. https://arxiv.org/abs//2405.01964 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcas…
  continue reading
 
Developing a method for large language models to abstain from providing incorrect answers, using self-consistency and conformal prediction to reduce hallucination rates. https://arxiv.org/abs//2405.01563 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/a…
  continue reading
 
Developing a method for large language models to abstain from providing incorrect answers, using self-consistency and conformal prediction to reduce hallucination rates. https://arxiv.org/abs//2405.01563 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/a…
  continue reading
 
Paper explores using neural architecture search (NAS) for structural pruning of pre-trained language models to optimize efficiency and generalization performance, utilizing two-stage weight-sharing NAS for accelerated search. https://arxiv.org/abs//2405.02267 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers…
  continue reading
 
Paper explores using neural architecture search (NAS) for structural pruning of pre-trained language models to optimize efficiency and generalization performance, utilizing two-stage weight-sharing NAS for accelerated search. https://arxiv.org/abs//2405.02267 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers…
  continue reading
 
The paper explores the impact of climate change on global food security, highlighting the need for sustainable agricultural practices to mitigate future risks. https://arxiv.org/abs//2404.18416 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-paper…
  continue reading
 
The paper explores the impact of climate change on global food security, highlighting the need for sustainable agricultural practices to mitigate future risks. https://arxiv.org/abs//2404.18416 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-paper…
  continue reading
 
The paper introduces Consistent Self-Attention and Semantic Motion Predictor to enhance content consistency in diffusion-based generative models for text-to-image and video generation, enabling rich visual story creation. https://arxiv.org/abs//2405.01434 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers App…
  continue reading
 
The paper introduces Consistent Self-Attention and Semantic Motion Predictor to enhance content consistency in diffusion-based generative models for text-to-image and video generation, enabling rich visual story creation. https://arxiv.org/abs//2405.01434 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers App…
  continue reading
 
The paper explores in-context learning (ICL) at extreme scales, showing performance improvements with hundreds or thousands of demonstrations, contrasting with example retrieval and finetuning. https://arxiv.org/abs//2405.00200 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcast…
  continue reading
 
The paper explores in-context learning (ICL) at extreme scales, showing performance improvements with hundreds or thousands of demonstrations, contrasting with example retrieval and finetuning. https://arxiv.org/abs//2405.00200 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcast…
  continue reading
 
WILDCHAT is a diverse dataset of 1 million user-ChatGPT conversations, offering rich insights into chatbot interactions and potential toxic use-cases for researchers. https://arxiv.org/abs//2405.01470 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxi…
  continue reading
 
WILDCHAT is a diverse dataset of 1 million user-ChatGPT conversations, offering rich insights into chatbot interactions and potential toxic use-cases for researchers. https://arxiv.org/abs//2405.01470 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxi…
  continue reading
 
NeMo-Aligner is a scalable toolkit for aligning Large Language Models with human values, supporting various alignment paradigms and designed for extensibility. https://arxiv.org/abs//2405.01481 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-paper…
  continue reading
 
NeMo-Aligner is a scalable toolkit for aligning Large Language Models with human values, supporting various alignment paradigms and designed for extensibility. https://arxiv.org/abs//2405.01481 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-paper…
  continue reading
 
Prometheus 2 is an open-source LM designed for evaluating responses, outperforming existing models in correlation with human and proprietary LM judgments. https://arxiv.org/abs//2405.01535 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1…
  continue reading
 
Prometheus 2 is an open-source LM designed for evaluating responses, outperforming existing models in correlation with human and proprietary LM judgments. https://arxiv.org/abs//2405.01535 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1…
  continue reading
 
Study investigates dataset contamination in large language models for mathematical reasoning using Grade School Math 1000 benchmark, finding evidence of overfitting and potential memorization of benchmark questions. https://arxiv.org/abs//2405.00332 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Pod…
  continue reading
 
Study investigates dataset contamination in large language models for mathematical reasoning using Grade School Math 1000 benchmark, finding evidence of overfitting and potential memorization of benchmark questions. https://arxiv.org/abs//2405.00332 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Pod…
  continue reading
 
The paper introduces SPPO, a self-play method for language model alignment, achieving state-of-the-art results without external supervision, outperforming DPO and IPO on various benchmarks. https://arxiv.org/abs//2405.00675 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.ap…
  continue reading
 
The paper introduces SPPO, a self-play method for language model alignment, achieving state-of-the-art results without external supervision, outperforming DPO and IPO on various benchmarks. https://arxiv.org/abs//2405.00675 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.ap…
  continue reading
 
Study evaluates model editing techniques on Llama-3, finding sequential editing more effective than batch editing. Suggests combining both methods for optimal performance. https://arxiv.org/abs//2405.00664 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast…
  continue reading
 
Study evaluates model editing techniques on Llama-3, finding sequential editing more effective than batch editing. Suggests combining both methods for optimal performance. https://arxiv.org/abs//2405.00664 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast…
  continue reading
 
Iterative preference optimization method enhances reasoning tasks by optimizing preference between generated Chain-of-Thought candidates, leading to improved accuracy on various datasets without additional sourcing. https://arxiv.org/abs//2404.19733 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Pod…
  continue reading
 
Iterative preference optimization method enhances reasoning tasks by optimizing preference between generated Chain-of-Thought candidates, leading to improved accuracy on various datasets without additional sourcing. https://arxiv.org/abs//2404.19733 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Pod…
  continue reading
 
https://arxiv.org/abs//2404.19708 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
https://arxiv.org/abs//2404.19708 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
Training language models to predict multiple future tokens at once improves sample efficiency, downstream capabilities, and inference speed without increasing training time, especially beneficial for larger models and generative tasks. https://arxiv.org/abs//2404.19737 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@ar…
  continue reading
 
Training language models to predict multiple future tokens at once improves sample efficiency, downstream capabilities, and inference speed without increasing training time, especially beneficial for larger models and generative tasks. https://arxiv.org/abs//2404.19737 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@ar…
  continue reading
 
https://arxiv.org/abs//2404.18796 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
https://arxiv.org/abs//2404.18796 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
The paper introduces Stylus, a method for efficiently selecting and composing task-specific adapters based on prompts' keywords, achieving high-quality image generation with improved efficiency and performance gains. https://arxiv.org/abs//2404.18928 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Po…
  continue reading
 
The paper introduces Stylus, a method for efficiently selecting and composing task-specific adapters based on prompts' keywords, achieving high-quality image generation with improved efficiency and performance gains. https://arxiv.org/abs//2404.18928 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Po…
  continue reading
 
Introducing Reinforced Token Optimization (RTO) framework for Reinforcement Learning from Human Feedback (RLHF) using Markov decision process (MDP) to improve token-wise reward learning and policy optimization. https://arxiv.org/abs//2404.18922 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts…
  continue reading
 
Introducing Reinforced Token Optimization (RTO) framework for Reinforcement Learning from Human Feedback (RLHF) using Markov decision process (MDP) to improve token-wise reward learning and policy optimization. https://arxiv.org/abs//2404.18922 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts…
  continue reading
 
Kangaroo introduces a self-speculative decoding framework for accelerating large language model inference, using a shallow sub-network and early exiting mechanisms to improve efficiency. https://arxiv.org/abs//2404.18911 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple…
  continue reading
 
Kangaroo introduces a self-speculative decoding framework for accelerating large language model inference, using a shallow sub-network and early exiting mechanisms to improve efficiency. https://arxiv.org/abs//2404.18911 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple…
  continue reading
 
https://arxiv.org/abs//2404.16873 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
https://arxiv.org/abs//2404.16873 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
Study explores if large language models understand their own language. Greedy Coordinate Gradient optimizer crafts prompts to compel coherent responses from nonsensical inputs, revealing efficiency and robustness differences. https://arxiv.org/abs//2404.17120 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers…
  continue reading
 
Study explores if large language models understand their own language. Greedy Coordinate Gradient optimizer crafts prompts to compel coherent responses from nonsensical inputs, revealing efficiency and robustness differences. https://arxiv.org/abs//2404.17120 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers…
  continue reading
 
Transformers can use meaningless filler tokens to solve tasks, but learning to use them is challenging. Additional tokens can provide computational benefits independently of token choice. https://arxiv.org/abs//2404.15758 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.appl…
  continue reading
 
Transformers can use meaningless filler tokens to solve tasks, but learning to use them is challenging. Additional tokens can provide computational benefits independently of token choice. https://arxiv.org/abs//2404.15758 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.appl…
  continue reading
 
This paper investigates how transformer-based language models retrieve information from long contexts, identifying special attention heads called retrieval heads as crucial for this task. https://arxiv.org/abs//2404.15574 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.appl…
  continue reading
 
This paper investigates how transformer-based language models retrieve information from long contexts, identifying special attention heads called retrieval heads as crucial for this task. https://arxiv.org/abs//2404.15574 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.appl…
  continue reading
 
AUTOCRAWLER combines large language models with crawlers to efficiently handle diverse web environments, improving adaptability and scalability compared to traditional methods. https://arxiv.org/abs//2404.12753 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/po…
  continue reading
 
AUTOCRAWLER combines large language models with crawlers to efficiently handle diverse web environments, improving adaptability and scalability compared to traditional methods. https://arxiv.org/abs//2404.12753 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/po…
  continue reading
 
https://arxiv.org/abs//2404.16820 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
https://arxiv.org/abs//2404.16820 YouTube: https://www.youtube.com/@ArxivPapers TikTok: https://www.tiktok.com/@arxiv_papers Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016 Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers --- Support this podcast: https://podcasters.spotify.com/pod/show/arxiv-papers/supp…
  continue reading
 
Loading …

Guia de referencia rapida