PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.
Live Feed

AI Research Paper Feed

Discover cutting-edge research papers in AI and machine learning. Stay ahead with the latest breakthroughs, insights, and discoveries from top researchers worldwide.

22,178 Research Papers
Multi-Domain Coverage
Real-time Updates
ArXivFeb 5, 2026

RRAttention: Dynamic Block Sparse Attention via Per-Head Round-Robin Shifts for Long-Context Inference

Siran Liu, Guoxia Wang et al.

TLDR: RRAttention is a novel attention mechanism that reduces computational complexity while maintaining performance in processing long contexts by using a dynamic block sparse attention method with a round-robin sampling strategy.

0190
arXivFeb 5, 2026

Shared LoRA Subspaces for almost Strict Continual Learning

Prakhar Kaushik, Ankit Vaidya et al.

TLDR: The Share method enables efficient continual learning by using a single, adaptable low-rank subspace, greatly reducing parameters and memory compared to traditional methods while maintaining performance.

02
ArXivFeb 5, 2026

DSB: Dynamic Sliding Block Scheduling for Diffusion LLMs

Lizhuo Luo, Shenggui Li et al.

TLDR: The paper introduces Dynamic Sliding Block (DSB), a training-free scheduling method for diffusion large language models that adapts to semantic difficulty, improving both quality and efficiency of text generation.

0283
ArXivFeb 5, 2026

KV-CoRE: Benchmarking Data-Dependent Low-Rank Compressibility of KV-Caches in LLMs

Jian Chen, Zhuoran Wang et al.

TLDR: KV-CoRE is a method to evaluate the data-dependent compressibility of kv-caches in large language models, revealing patterns linked to model architecture and training data across multiple languages.

0202
ArXivFeb 5, 2026

CASTLE: A Comprehensive Benchmark for Evaluating Student-Tailored Personalized Safety in Large Language Models

Rui Jia, Ruiyi Lan et al.

TLDR: CASTLE is a benchmark designed to evaluate the personalized safety of large language models in educational settings, revealing significant deficiencies in current models' abilities to tailor responses to individual student needs and risks.

0146
ArXivFeb 5, 2026

Logarithmic-time Schedules for Scaling Language Models with Momentum

Damien Ferbach, Courtney Paquette et al.

TLDR: ADANA, an optimizer with time-varying schedules for hyperparameters, improves large-scale language model training efficiency by up to 40% compared to AdamW.

0386
ArXivFeb 5, 2026

Extreme Weather Nowcasting via Local Precipitation Pattern Prediction

Changhoon Song, Teng Yuan Chang et al.

TLDR: The paper introduces exPreCast, a deterministic model for accurate and efficient nowcasting of both normal and extreme rainfall using a balanced dataset from the Korea Meteorological Administration.

0411
ArXivFeb 5, 2026

SAGE: Benchmarking and Improving Retrieval for Deep Research Agents

Tiansheng Hu, Yilun Zhao et al.

TLDR: The SAGE benchmark reveals that traditional BM25 outperforms LLM-based retrievers for scientific literature retrieval, with enhancements possible through document augmentation using LLMs.

0195
ArXivFeb 5, 2026

Disentangled Representation Learning via Flow Matching

Jinjin Chi, Taoping Liu et al.

TLDR: The paper introduces a flow matching-based framework for disentangled representation learning that improves semantic alignment and disentanglement scores by using a non-overlap regularizer to reduce factor interference.

0415
ArXivFeb 5, 2026

Polyglots or Multitudes? Multilingual LLM Answers to Value-laden Multiple-Choice Questions

Léo Labat, Etienne Ollion et al.

TLDR: This study examines how multilingual large language models (LLMs) respond to value-laden multiple-choice questions across different languages, revealing variability in consistency and language-specific behaviors.

0221
ArXivFeb 5, 2026

Different Time, Different Language: Revisiting the Bias Against Non-Native Speakers in GPT Detectors

Adnan Al Ali, Jindřich Helcl et al.

TLDR: The study finds no systematic bias against non-native Czech speakers in LLM-based text detectors and shows that modern detectors do not rely on perplexity to identify generated text.

0189
ArXivFeb 5, 2026

LongR: Unleashing Long-Context Reasoning via Reinforcement Learning with Dense Utility Rewards

Bowen Ping, Zijun Chen et al.

TLDR: LongR is a framework that improves long-context reasoning in reinforcement learning by using a dynamic 'Think-and-Read' mechanism and dense utility rewards, achieving significant gains on benchmarks like LongBench v2.

0185
ArXivFeb 5, 2026

Causal Front-Door Adjustment for Robust Jailbreak Attacks on LLMs

Yao Zhou, Zeen Song et al.

TLDR: The paper introduces a novel method using causal front-door adjustment to effectively bypass safety mechanisms in large language models for jailbreak attacks.

0143
ArXivFeb 5, 2026

MentorCollab: Selective Large-to-Small Inference-Time Guidance for Efficient Reasoning

Haojin Wang, Yike Wang et al.

TLDR: MentorCollab improves small model reasoning by selectively using guidance from a large model, enhancing performance with minimal additional cost.

04
ArXivFeb 5, 2026

Smoothness Errors in Dynamics Models and How to Avoid Them

Edward Berman, Luisa Li et al.

TLDR: This paper introduces relaxed unitary convolutions for graph neural networks to improve performance in dynamics modeling by balancing smoothness preservation with natural physical system requirements.

0369
arXivFeb 5, 2026

Learning Query-Aware Budget-Tier Routing for Runtime Agent Memory

Haozhen Zhang, Haodong Yue et al.

TLDR: BudgetMem is a runtime memory framework for LLMs that optimizes performance-cost trade-offs using query-aware budget-tier routing across memory modules.

03
ArXivFeb 5, 2026

PatchFlow: Leveraging a Flow-Based Model with Patch Features

Boxiang Zhang, Baijian Yang et al.

TLDR: PatchFlow improves defect detection in die casting using local patch features and a flow-based model, reducing error rates significantly on multiple datasets.

0368
ArXivFeb 5, 2026

Piecewise Deterministic Markov Processes for Bayesian Inference of PDE Coefficients

Leon Riccius, Iuri B. C. M. Rocha et al.

TLDR: The paper introduces a framework using piecewise deterministic Markov processes (PDMP) for efficient Bayesian inference in complex inverse problems, demonstrating improved accuracy and efficiency over traditional methods.

0159
ArXivFeb 5, 2026

Radon--Wasserstein Gradient Flows for Interacting-Particle Sampling in High Dimensions

Elias Hess-Childs, Dejan Slepčev et al.

TLDR: The paper introduces new Radon--Wasserstein gradient flows for efficient high-dimensional sampling using interacting particles with linear scaling costs.

0372
arXivFeb 5, 2026

Correctness-Optimized Residual Activation Lens (CORAL): Transferrable and Calibration-Aware Inference-Time Steering

Miranda Muqing Miao, Young-Min Cho et al.

TLDR: CORAL is a method that improves the accuracy and calibration of large language models during inference without retraining.

03
Showing 1-20 of 22178 papers
per page
…
1 / 1109