PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.
Live Feed

AI Research Paper Feed

Discover cutting-edge research papers in AI and machine learning. Stay ahead with the latest breakthroughs, insights, and discoveries from top researchers worldwide.

21,679 Research Papers
Multi-Domain Coverage
Real-time Updates
ArXivFeb 3, 2026

Agentic Proposing: Enhancing Large Language Model Reasoning via Compositional Skill Synthesis

Zhengbo Jiao, Shaobo Wang et al.

TLDR: Agentic Proposing is a framework that improves reasoning in large language models by dynamically composing modular skills, achieving high accuracy with fewer data points.

07
ArXivFeb 3, 2026

CSR-Bench: A Benchmark for Evaluating the Cross-modal Safety and Reliability of MLLMs

Yuxuan Liu, Yuntian Shi et al.

TLDR: CSR-Bench is a benchmark for evaluating the cross-modal safety and reliability of multimodal large language models, revealing systematic alignment gaps and trade-offs between over-rejection and safety.

07
ArXivFeb 3, 2026

GFlowPO: Generative Flow Network as a Language Model Prompt Optimizer

Junmo Cho, Suhan Kim et al.

TLDR: GFlowPO is a new framework for optimizing language model prompts using a probabilistic approach and dynamic memory updates, leading to better performance in various language tasks.

070
ArXivFeb 3, 2026

Rejecting Arguments Based on Doubt in Structured Bipolar Argumentation

Michael A. Müller, Srdjan Vesic et al.

TLDR: The paper introduces structured bipolar argumentation frameworks (SBAFs) that allow agents to reject arguments based on doubt and to focus on individual sentences rather than whole arguments, providing new semantics that do not require accepting all defendable arguments.

066
ArXivFeb 3, 2026

R1-SyntheticVL: Is Synthetic Data from Generative Models Ready for Multimodal Large Language Model?

Jingyi Zhang, Tianyi Lin et al.

TLDR: The study introduces CADS, a method for generating high-quality synthetic multimodal data to improve multimodal large language models, resulting in the creation of MMSynthetic-20K and the high-performing R1-SyntheticVL model.

041
ArXivFeb 3, 2026

Unveiling Covert Toxicity in Multimodal Data via Toxicity Association Graphs: A Graph-Based Metric and Interpretable Detection Framework

Guanzong Wu, Zihao Zhu et al.

TLDR: The paper introduces a novel framework using Toxicity Association Graphs to detect covert toxicity in multimodal data, offering a new metric for measuring hidden toxicity and outperforming existing methods in interpretability and detection accuracy.

011
ArXivFeb 3, 2026

An Approximate Ascent Approach To Prove Convergence of PPO

Leif Doering, Daniel Schmidt et al.

TLDR: This paper provides a convergence proof for Proximal Policy Optimization (PPO) by interpreting its update scheme as approximate policy gradient ascent and addresses an issue in Generalized Advantage Estimation (GAE).

0390
ArXivFeb 3, 2026

Live or Lie: Action-Aware Capsule Multiple Instance Learning for Risk Assessment in Live Streaming Platforms

Yiran Qiao, Jing Chen et al.

TLDR: The study introduces AC-MIL, a novel framework for assessing risks in live streaming by analyzing user behaviors and coordination patterns, significantly improving detection accuracy and interpretability over existing methods.

0435
ArXivFeb 3, 2026

Beyond Variance: Prompt-Efficient RLVR via Rare-Event Amplification and Bidirectional Pairing

Xin Sheng, Jiaxin Li et al.

TLDR: The paper introduces a method called positive-negative pairing for prompt selection in reinforcement learning with verifiable rewards, leading to improved performance on deterministic reasoning tasks by amplifying rare event signals.

0333
ArXivFeb 3, 2026

Ontology-to-tools compilation for executable semantic constraint enforcement in LLM agents

Xiaochi Zhou, Patrick Bulter et al.

TLDR: This paper presents a method to integrate formal domain knowledge into large language models by compiling ontological specifications into tools that enforce semantic constraints during knowledge generation.

0318
ArXivFeb 3, 2026

Memora: A Harmonic Memory Representation Balancing Abstraction and Specificity

Menglin Xia, Xuchao Zhang et al.

TLDR: Memora is a memory representation system that balances abstraction and specificity to improve retrieval relevance and reasoning effectiveness in agent memory systems.

047
ArXivFeb 3, 2026

Entropy-Gated Selective Policy Optimization:Token-Level Gradient Allocation for Hybrid Training of Large Language Models

Yuelin Hu, Zhengxue Cheng et al.

TLDR: EGSPO improves large language model training by using token-level gradient modulation to enhance performance on mathematical reasoning tasks with minimal computational overhead.

049
ArXivFeb 3, 2026

MemCast: Memory-Driven Time Series Forecasting with Experience-Conditioned Reasoning

Xiaoyu Tao, Mingyue Cheng et al.

TLDR: MemCast introduces a memory-driven framework for time series forecasting that leverages past experiences to improve prediction accuracy and adaptability over time.

029
ArXivFeb 3, 2026

MeetBench-XL: Calibrated Multi-Dimensional Evaluation and Learned Dual-Policy Agents for Real-Time Meetings

Yuelin Hu, Jun Xu et al.

TLDR: MeetBench-XL introduces a comprehensive evaluation framework and a dual-policy AI agent for enhancing real-time meeting assistance in enterprise environments.

025
ArXivFeb 3, 2026

TodyComm: Task-Oriented Dynamic Communication for Multi-Round LLM-based Multi-Agent System

Wenzhe Fan, Tommaso Tognoli et al.

TLDR: TodyComm is a dynamic communication algorithm for multi-agent systems that adapts to changing conditions, improving task effectiveness and efficiency.

0778
ArXivFeb 3, 2026

RAGTurk: Best Practices for Retrieval Augmented Generation in Turkish

Süha Kağan Köse, Mehmet Can Baytekin et al.

TLDR: The study develops a Turkish-specific Retrieval-Augmented Generation (RAG) dataset and benchmarks various methods, finding that complex methods like HyDE significantly improve accuracy over simpler baselines.

0744
ArXivFeb 3, 2026

HySparse: A Hybrid Sparse Attention Architecture with Oracle Token Selection and KV Cache Sharing

Yizhao Gao, Jianyu Wei et al.

TLDR: HySparse is a new hybrid sparse attention model that improves performance and reduces memory usage by using full attention layers as oracles for token selection and sharing KV caches with sparse layers.

0391
ArXivFeb 3, 2026

ELIQ: A Label-Free Framework for Quality Assessment of Evolving AI-Generated Images

Xinyue Li, Zhiming Xu et al.

TLDR: ELIQ is a label-free framework designed to assess the quality of AI-generated images, adapting to evolving generative models without the need for human annotations.

0438
ArXivFeb 3, 2026

DeepDFA: Injecting Temporal Logic in Deep Learning for Sequential Subsymbolic Applications

Elena Umili, Francesco Argenziano et al.

TLDR: DeepDFA is a framework that integrates temporal logic into neural networks, outperforming traditional models in sequential tasks by bridging subsymbolic learning and symbolic reasoning.

0411
ArXivFeb 3, 2026

Accordion-Thinking: Self-Regulated Step Summaries for Efficient and Readable LLM Reasoning

Zhicheng Yang, Zhijiang Guo et al.

TLDR: Accordion-Thinking enables LLMs to self-regulate reasoning step granularity, achieving efficient and readable reasoning with reduced computational overhead.

00
Showing 1-20 of 21679 papers
per page
…
1 / 1084