PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.
Live Feed

AI Research Paper Feed

Discover cutting-edge research papers in AI and machine learning. Stay ahead with the latest breakthroughs, insights, and discoveries from top researchers worldwide.

22,178 Research Papers
Multi-Domain Coverage
Real-time Updates
ArXivFeb 5, 2026

Codified Finite-state Machines for Role-playing

Letian Peng, Yupeng Hou et al.

TLDR: The paper introduces Codified Finite-State Machines (CFSMs) and their probabilistic extension (CPFSMs) to improve character state modeling in role-playing with large language models, enhancing consistency and engagement by automatically generating state transitions from character profiles.

025
ArXivFeb 5, 2026

A Short and Unified Convergence Analysis of the SAG, SAGA, and IAG Algorithms

Feng Zhu, Robert W. Heath et al.

TLDR: This paper presents a unified convergence analysis for the SAG, SAGA, and IAG algorithms, providing a simpler and more comprehensive understanding of their performance.

04
arXivFeb 5, 2026

Verification of the Implicit World Model in a Generative Model via Adversarial Sequences

András Balogh, Márk Jelasity

TLDR: The study evaluates the soundness of generative models in chess by using adversarial sequences to reveal their limitations and improve training techniques.

03
ArXivFeb 5, 2026

SAGE: Benchmarking and Improving Retrieval for Deep Research Agents

Tiansheng Hu, Yilun Zhao et al.

TLDR: The SAGE benchmark reveals that traditional BM25 outperforms LLM-based retrievers for scientific literature retrieval, with enhancements possible through document augmentation using LLMs.

01,543
arXivFeb 5, 2026

xList-Hate: A Checklist-Based Framework for Interpretable and Generalizable Hate Speech Detection

Adrián Girón, Pablo Miralles et al.

TLDR: xList-Hate is a new framework for hate speech detection that uses a checklist-based approach for improved interpretability and robustness across different datasets and conditions.

01
arXivFeb 5, 2026

DLM-Scope: Mechanistic Interpretability of Diffusion Language Models via Sparse Autoencoders

Xu Wang, Bingqing Jiang et al.

TLDR: DLM-Scope introduces a sparse autoencoder-based framework for interpreting diffusion language models, showing unique advantages over traditional autoregressive models.

02
ArXivFeb 5, 2026

Logarithmic-time Schedules for Scaling Language Models with Momentum

Damien Ferbach, Courtney Paquette et al.

TLDR: ADANA, an optimizer with time-varying schedules for hyperparameters, improves large-scale language model training efficiency by up to 40% compared to AdamW.

01,661
arXivFeb 5, 2026

EuroLLM-22B: Technical Report

Miguel Moura Ramos, Duarte M. Alves et al.

TLDR: EuroLLM-22B is a large language model designed to support all 24 EU languages and 11 additional languages, showing strong multilingual performance and being openly released for research use.

01
arXivFeb 5, 2026

Dr. Kernel: Reinforcement Learning Done Right for Triton Kernel Generations

Wei Liu, Jiawei Xu et al.

TLDR: The paper presents Dr.Kernel, a reinforcement learning approach for generating high-quality AI kernels using a new environment, KernelGYM, achieving significant speedup over existing models.

01
ArXivFeb 5, 2026

MerNav: A Highly Generalizable Memory-Execute-Review Framework for Zero-Shot Object Goal Navigation

Dekang Qi, Shuang Zeng et al.

TLDR: MerNav introduces a Memory-Execute-Review framework that significantly improves success rates and generalization in zero-shot object goal navigation tasks.

0163
ArXivFeb 5, 2026

Tight Long-Term Tail Decay of (Clipped) SGD in Non-Convex Optimization

Aleksandar Armacki, Dragana Bajović et al.

TLDR: This paper establishes tight long-term tail decay rates for SGD and clipped SGD in non-convex optimization, showing significantly faster decay than previously known results.

0139
ArXivFeb 5, 2026

Cost-Efficient RAG for Entity Matching with LLMs: A Blocking-based Exploration

Chuangtao Ma, Zeyu Zhang et al.

TLDR: CE-RAG4EM is a cost-efficient RAG system for entity matching that reduces computational overhead while maintaining or improving performance.

017
arXivFeb 5, 2026

DARWIN: Dynamic Agentically Rewriting Self-Improving Network

Henry Jiang

TLDR: DARWIN is an evolutionary GPT model that uses a genetic algorithm approach to iteratively improve its performance by allowing GPT agents to modify each other's training code, resulting in a 1.26% improvement in FLOPS utilization and a 2.07% improvement in perplexity over five iterations.

02
ArXivFeb 5, 2026

TADS: Task-Aware Data Selection for Multi-Task Multimodal Pre-Training

Guanjie Cheng, Boyi Li et al.

TLDR: TADS is a new framework for selecting high-quality, task-relevant data for multi-task multimodal pre-training, improving efficiency and performance using less data.

05
ArXivFeb 5, 2026

PatchFlow: Leveraging a Flow-Based Model with Patch Features

Boxiang Zhang, Baijian Yang et al.

TLDR: PatchFlow improves defect detection in die casting using local patch features and a flow-based model, reducing error rates significantly on multiple datasets.

01,529
ArXivFeb 5, 2026

Piecewise Deterministic Markov Processes for Bayesian Inference of PDE Coefficients

Leon Riccius, Iuri B. C. M. Rocha et al.

TLDR: The paper introduces a framework using piecewise deterministic Markov processes (PDMP) for efficient Bayesian inference in complex inverse problems, demonstrating improved accuracy and efficiency over traditional methods.

0532
ArXivFeb 5, 2026

Radon--Wasserstein Gradient Flows for Interacting-Particle Sampling in High Dimensions

Elias Hess-Childs, Dejan Slepčev et al.

TLDR: The paper introduces new Radon--Wasserstein gradient flows for efficient high-dimensional sampling using interacting particles with linear scaling costs.

01,508
ArXivFeb 5, 2026

Bayesian Neighborhood Adaptation for Graph Neural Networks

Paribesh Regmi, Rui Li et al.

TLDR: The paper introduces a Bayesian framework to adaptively determine the optimal neighborhood scope for graph neural networks, improving performance on node classification tasks across various datasets.

04
arXivFeb 5, 2026

Parity, Sensitivity, and Transformers

Alexander Kozachinskiy, Tomasz Steifer et al.

TLDR: This paper presents a new construction of a transformer that can solve the PARITY problem using a single layer with practical features, and establishes a lower bound proving that a single-layer, single-head transformer cannot solve PARITY.

01
ArXivFeb 5, 2026

Detecting Misbehaviors of Large Vision-Language Models by Evidential Uncertainty Quantification

Tao Huang, Rui Wang et al.

TLDR: The paper introduces Evidential Uncertainty Quantification (EUQ), a method to detect misbehaviors in large vision-language models by assessing internal conflicts and knowledge gaps, outperforming existing methods in identifying issues like hallucinations and adversarial vulnerabilities.

0255
Showing 1-20 of 22178 papers
per page
…
1 / 1109