PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.
Live Feed

AI Research Paper Feed

Discover cutting-edge research papers in AI and machine learning. Stay ahead with the latest breakthroughs, insights, and discoveries from top researchers worldwide.

22,178 Research Papers
Multi-Domain Coverage
Real-time Updates
ArXivFeb 5, 2026

DFlash: Block Diffusion for Flash Speculative Decoding

Jian Chen, Yesheng Liang et al.

TLDR: DFlash is a speculative decoding framework that uses block diffusion for parallel drafting, achieving over 6x acceleration in language model inference compared to traditional methods.

0164
ArXivFeb 5, 2026

KV-CoRE: Benchmarking Data-Dependent Low-Rank Compressibility of KV-Caches in LLMs

Jian Chen, Zhuoran Wang et al.

TLDR: KV-CoRE is a method to evaluate the data-dependent compressibility of kv-caches in large language models, revealing patterns linked to model architecture and training data across multiple languages.

01,935
ArXivFeb 5, 2026

An Asymptotic Law of the Iterated Logarithm for $\mathrm{KL}_{\inf}$

Ashwin Ram, Aaditya Ramdas

TLDR: This paper establishes a tight law of the iterated logarithm for empirical KL-infinity statistics, applicable to very general data conditions.

0411
ArXivFeb 5, 2026

Copyright Detective: A Forensic System to Evidence LLMs Flickering Copyright Leakage Risks

Guangwei Zhang, Jianing Zhu et al.

TLDR: Copyright Detective is an interactive forensic system designed to detect and analyze potential copyright risks in the outputs of large language models (LLMs).

0335
ArXivFeb 5, 2026

Radon--Wasserstein Gradient Flows for Interacting-Particle Sampling in High Dimensions

Elias Hess-Childs, Dejan Slepčev et al.

TLDR: The paper introduces new Radon--Wasserstein gradient flows for efficient high-dimensional sampling using interacting particles with linear scaling costs.

03,810
ArXivFeb 5, 2026

LongR: Unleashing Long-Context Reasoning via Reinforcement Learning with Dense Utility Rewards

Bowen Ping, Zijun Chen et al.

TLDR: LongR is a framework that improves long-context reasoning in reinforcement learning by using a dynamic 'Think-and-Read' mechanism and dense utility rewards, achieving significant gains on benchmarks like LongBench v2.

01,806
ArXivFeb 5, 2026

Logarithmic-time Schedules for Scaling Language Models with Momentum

Damien Ferbach, Courtney Paquette et al.

TLDR: ADANA, an optimizer with time-varying schedules for hyperparameters, improves large-scale language model training efficiency by up to 40% compared to AdamW.

04,296
ArXivFeb 5, 2026

Path Sampling for Rare Events Boosted by Machine Learning

Porhouy Minh, Sapna Sarupria

TLDR: AIMMD is a new algorithm that uses machine learning to improve the efficiency of transition path sampling for studying molecular processes.

01,496
ArXivFeb 5, 2026

Private Prediction via Shrinkage

Chao Yan

TLDR: The paper presents a method to achieve differentially private prediction with reduced dependence on the number of queries, improving efficiency in streaming settings.

01,403
ArXivFeb 5, 2026

Multi-Field Tool Retrieval

Yichen Tang, Weihang Su et al.

TLDR: The paper introduces a Multi-Field Tool Retrieval framework to improve how Large Language Models select external tools by addressing challenges in tool documentation and user query alignment.

0412
ArXivFeb 5, 2026

Grammatical Error Correction Evaluation by Optimally Transporting Edit Representation

Takumi Goto, Yusuke Sakai et al.

TLDR: The paper introduces UOT-ERRANT, a new metric for evaluating grammatical error correction systems by optimally transporting edit vectors, showing improved performance and interpretability.

01,559
arXivFeb 5, 2026

Clifford Kolmogorov-Arnold Networks

Matthias Wolff, Francesco Alesiani et al.

TLDR: The Clifford Kolmogorov-Arnold Network (ClKAN) is a new architecture for approximating functions in Clifford algebra spaces, utilizing Randomized Quasi Monte Carlo methods and novel batch normalization strategies for improved scalability and efficiency.

01,530
arXivFeb 5, 2026

Diamond Maps: Efficient Reward Alignment via Stochastic Flow Maps

Peter Holderrieth, Douglas Chen et al.

TLDR: Diamond Maps are a new model for generative tasks that efficiently align with user preferences by enabling quick adaptation to rewards during inference.

0100
ArXivFeb 5, 2026

Faithful Bi-Directional Model Steering via Distribution Matching and Distributed Interchange Interventions

Yuntai Bao, Xuhong Zhang et al.

TLDR: The paper introduces Concept DAS (CDAS), a novel intervention-based model steering method that uses distribution matching to achieve more faithful and stable control compared to traditional preference-optimization methods.

01,456
ArXivFeb 5, 2026

Hinge Regression Tree: A Newton Method for Oblique Regression Tree Splitting

Hongyi Li, Han Lin et al.

TLDR: The Hinge Regression Tree (HRT) is a new method for creating oblique decision trees using a Newton method that improves split quality and convergence speed, outperforming traditional tree models.

01,466
ArXivFeb 5, 2026

Fairness Under Group-Conditional Prior Probability Shift: Invariance, Drift, and Target-Aware Post-Processing

Amir Asiaee, Kaveh Aryan

TLDR: The paper addresses fairness in machine learning under group-conditional prior probability shift and introduces a method to maintain fairness when label prevalences change across demographic groups between training and deployment.

01,374
ArXivFeb 5, 2026

PACE: Defying the Scaling Hypothesis of Exploration in Iterative Alignment for Mathematical Reasoning

Jun Rao, Zixiong Yu et al.

TLDR: PACE introduces a more efficient method for mathematical reasoning in language models by using minimal exploration, outperforming traditional methods with less computational cost.

0579
ArXivFeb 5, 2026

A Short and Unified Convergence Analysis of the SAG, SAGA, and IAG Algorithms

Feng Zhu, Robert W. Heath et al.

TLDR: This paper presents a unified convergence analysis for the SAG, SAGA, and IAG algorithms, providing a simpler and more comprehensive understanding of their performance.

02,341
arXivFeb 5, 2026

Parity, Sensitivity, and Transformers

Alexander Kozachinskiy, Tomasz Steifer et al.

TLDR: This paper presents a new construction of a transformer that can solve the PARITY problem using a single layer with practical features, and establishes a lower bound proving that a single-layer, single-head transformer cannot solve PARITY.

02,388
ArXivFeb 5, 2026

FedRandom: Sampling Consistent and Accurate Contribution Values in Federated Learning

Arno Geimer, Beltran Fiz Pontiveros et al.

TLDR: FedRandom is a novel technique that improves the stability and accuracy of participant contribution assessments in federated learning by generating more consistent samples, significantly reducing estimation errors.

0240
Showing 1-20 of 22178 papers
per page
…
1 / 1109