PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.
Live Feed

AI Research Paper Feed

Discover cutting-edge research papers in AI and machine learning. Stay ahead with the latest breakthroughs, insights, and discoveries from top researchers worldwide.

22,578 Research Papers
Multi-Domain Coverage
Real-time Updates
ArXivFeb 19, 2026

Asymptotic Smoothing of the Lipschitz Loss Landscape in Overparameterized One-Hidden-Layer ReLU Networks

Saveliy Baturin

TLDR: This paper shows that in overparameterized one-hidden-layer ReLU networks, the loss landscape becomes smoother and flatter as the network width increases, resulting in smaller energy gaps between local and global minima.

0212
ArXivFeb 19, 2026

Optimal Unconstrained Self-Distillation in Ridge Regression: Strict Improvements, Precise Asymptotics, and One-Shot Tuning

Hien Dang, Pratik Patil et al.

TLDR: This paper demonstrates that self-distillation can significantly improve ridge regression performance by optimally mixing teacher predictions, providing precise asymptotic analyses and a practical one-shot tuning method.

03
ArXivFeb 19, 2026

Instructor-Aligned Knowledge Graphs for Personalized Learning

Abdulrahman AlRabah, Priyanka Kargupta et al.

TLDR: InstructKG is a framework that automatically constructs knowledge graphs from course materials to capture learning dependencies and aid personalized learning.

05,517
ArXivFeb 19, 2026

Catastrophic Forgetting Resilient One-Shot Incremental Federated Learning

Obaidullah Zaland, Zulfiqar Ahmad Khan et al.

TLDR: This paper introduces One-Shot Incremental Federated Learning (OSI-FL), a framework that addresses communication overhead and catastrophic forgetting in federated learning by using category-specific embeddings and selective sample retention.

0599
ArXivFeb 19, 2026

Simultaneous Blackwell Approachability and Applications to Multiclass Omniprediction

Lunjia Hu, Kevin Tian et al.

TLDR: This paper extends a binary omniprediction algorithm to handle multiclass prediction problems, achieving suboptimality bounds with a sample complexity or regret horizon of approximately ε^{-(k+1)} for k-class predictions.

03
ArXivFeb 19, 2026

Phase-Aware Mixture of Experts for Agentic Reinforcement Learning

Shengtian Yang, Yu Li et al.

TLDR: The paper introduces Phase-Aware Mixture of Experts (PA-MoE) to enhance reinforcement learning by allowing expert specialization for complex tasks without being dominated by simpler tasks.

03,620
ArXivFeb 19, 2026

A Privacy by Design Framework for Large Language Model-Based Applications for Children

Diana Addae, Diana Rogachova et al.

TLDR: This paper proposes a Privacy-by-Design framework for developing AI applications for children that integrates privacy regulations to ensure data protection and legal compliance.

07,163
ArXivFeb 19, 2026

KLong: Training LLM Agent for Extremely Long-horizon Tasks

Yue Liu, Zhiyuan Hu et al.

TLDR: KLong is a new LLM agent designed to tackle long-horizon tasks using a novel training method combining trajectory-splitting SFT and progressive RL, outperforming existing models on various benchmarks.

07,121
ArXivFeb 19, 2026

Deeper detection limits in astronomical imaging using self-supervised spatiotemporal denoising

Yuduo Guo, Hao Zhang et al.

TLDR: ASTERIS, a self-supervised denoising algorithm, enhances astronomical imaging detection limits by leveraging spatiotemporal data, improving detection by 1 magnitude and identifying previously undetectable features in deep space images.

05,328
ArXivFeb 19, 2026

Universal Fine-Grained Symmetry Inference and Enforcement for Rigorous Crystal Structure Prediction

Shi Yin, Jinming Mu et al.

TLDR: This paper presents a novel approach to crystal structure prediction using large language models and constrained optimization to improve symmetry inference and enforce physical validity, achieving state-of-the-art results without relying on existing databases.

04,578
ArXivFeb 19, 2026

Deep Reinforcement Learning for Optimal Portfolio Allocation: A Comparative Study with Mean-Variance Optimization

Srijan Sood, Kassiani Papasotiriou et al.

TLDR: This study compares Deep Reinforcement Learning (DRL) and Mean-Variance Optimization (MVO) for portfolio allocation, showing DRL's strong performance across various financial metrics.

03,904
ArXivFeb 19, 2026

Transforming Behavioral Neuroscience Discovery with In-Context Learning and AI-Enhanced Tensor Methods

Paimon Goulart, Jordan Steinhauser et al.

TLDR: This paper presents an AI-enhanced pipeline using In-Context Learning and tensor methods to improve data analysis in behavioral neuroscience, particularly for studying fear generalization in mice, which can help understand PTSD.

03,006
ArXivFeb 19, 2026

Sign Lock-In: Randomly Initialized Weight Signs Persist and Bottleneck Sub-Bit Model Compression

Akira Sakai, Yuma Ichikawa

TLDR: The paper identifies that weight sign persistence is a bottleneck in sub-bit model compression and proposes methods to reduce sign flips while maintaining performance.

03,663
ArXivFeb 19, 2026

Retaining Suboptimal Actions to Follow Shifting Optima in Multi-Agent Reinforcement Learning

Yonghyeon Jo, Sunwoo Lee et al.

TLDR: The paper introduces Successive Sub-value Q-learning (S2Q), a method that improves adaptability in multi-agent reinforcement learning by retaining multiple high-value actions, outperforming existing algorithms.

03,387
ArXivFeb 19, 2026

Evaluating Chain-of-Thought Reasoning through Reusability and Verifiability

Shashank Aggarwal, Ram Vikas Mishra et al.

TLDR: This paper introduces reusability and verifiability as new metrics to evaluate the quality of Chain-of-Thought reasoning in multi-agent IR pipelines, revealing that these metrics are not correlated with traditional accuracy measures.

06,217
ArXivFeb 19, 2026

The Anxiety of Influence: Bloom Filters in Transformer Attention Heads

Peter Balogh

TLDR: Certain transformer attention heads in language models act as membership testers, identifying repeated tokens with high precision, similar to Bloom filters.

06,120
ArXivFeb 19, 2026

LORA-CRAFT: Cross-layer Rank Adaptation via Frozen Tucker Decomposition of Pre-trained Attention Weights

Kasun Dewage, Marianna Pensky et al.

TLDR: CRAFT is a parameter-efficient fine-tuning method using Tucker decomposition on pre-trained attention weights, achieving competitive performance with minimal adaptation parameters.

05,977
ArXivFeb 19, 2026

Fine-Grained Uncertainty Quantification for Long-Form Language Model Outputs: A Comparative Study

Dylan Bouchard, Mohit Singh Chauhan et al.

TLDR: This study introduces a taxonomy for fine-grained uncertainty quantification in long-form language model outputs, revealing that claim-level scoring and uncertainty-aware decoding improve factuality in generated content.

06,113
ArXivFeb 19, 2026

From Labor to Collaboration: A Methodological Experiment Using AI Agents to Augment Research Perspectives in Taiwan's Humanities and Social Sciences

Yi-Chih Huang

TLDR: This study proposes a collaborative AI workflow for humanities and social sciences research, using Taiwan's Claude.ai data to validate its feasibility and effectiveness.

04,658
ArXivFeb 19, 2026

Revisiting Weight Regularization for Low-Rank Continual Learning

Yaoyue Zheng, Yin Zhang et al.

TLDR: The paper introduces EWC-LoRA, a method that uses weight regularization with low-rank updates to improve continual learning with pre-trained models, reducing task interference while maintaining efficiency.

00
Showing 1-20 of 22578 papers
per page
…
1 / 1129