Papers
Research papers from arXiv and related sources
A Persistent-State Dataflow Accelerator for Memory-Bound Linear Attention Decode on FPGA
Gated DeltaNet (GDN) is a linear attention mechanism that replaces the growing KV cache with a fixed-size recurrent state. Hybrid LLMs like Qwen3-Next use 75% GDN layers and achieve competitive acc...
Neelesh Gupta, Peter Wang, Rajgopal Kannan, Viktor K. Prasanna
CTAO Simulations for Potential PeVatron Candidates
This paper reports on the capabilities of the Cherenkov Telescope Array Observatory (CTAO) in detecting high-energy gamma-rays that show significant contributions of hadronic origin. We focus on fo...
P. Sharma, C. Dubos, S. R. Patel T. Suomijarvi
Weak-SIGReg: Covariance Regularization for Stable Deep Learning
Modern neural network optimization relies heavily on architectural priorssuch as Batch Normalization and Residual connectionsto stabilize training dynamics. Without these, or in low-data regimes wi...
Habibullah Akbar
Learning Next Action Predictors from Human-Computer Interaction
Truly proactive AI systems must anticipate what we will do next. This foresight demands far richer information than the sparse signals we type into our prompts -- it demands reasoning over the enti...
Omar Shaikh, Valentin Teutschbein, Kanishk Gandhi, Yikun Chi, Nick Haber, Thomas Robinson, Nilam ...
Design Experiments to Compare Multi-armed Bandit Algorithms
Online platforms routinely compare multi-armed bandit algorithms, such as UCB and Thompson Sampling, to select the best-performing policy. Unlike standard A/B tests for static treatments, each run ...
Huiling Meng, Ningyuan Chen, Xuefeng Gao
Stock Market Prediction Using Node Transformer Architecture Integrated with BERT Sentiment Analysis
Stock market prediction presents considerable challenges for investors, financial institutions, and policymakers operating in complex market environments characterized by noise, non-stationarity, a...
Mohammad Al Ridhawi, Mahtab Haj Ali, Hussein Al Osman
ThermoCAPTCHA: Privacy-Preserving Human Verification with Farm-Resistant Traceable Tokens
CAPTCHAs remain a critical defense against automated abuse, yet modern systems suffer from well-known limitations in usability, accessibility, and resistance to increasingly capable bots and low-co...
Shovon Paul, Md Imran Hossen, Xiali Hei
Multi-Shot Quantum Sensing for RF Signal Detection with MIMO Rydberg-Atom Receivers
Rydberg-atom quantum receivers (RAQRs) enable electric-field sensing with quantum-noise-limited performance, yet their optical readout provides only magnitude measurements whose fluctuations follow...
Saman Atapattu, Harini Hapuarachchi, Nathan Ross
DeepFact: Co-Evolving Benchmarks and Agents for Deep Research Factuality
Search-augmented LLM agents can produce deep research reports (DRRs), but verifying claim-level factuality remains challenging. Existing fact-checkers are primarily designed for general-domain, fac...
Yukun Huang, Leonardo F. R. Ribeiro, Momchil Hardalov, Bhuwan Dhingra, Markus Dreyer, Venkatesh S...
The World Won't Stay Still: Programmable Evolution for Agent Benchmarks
LLM-powered agents fulfill user requests by interacting with environments, querying data, and invoking tools in a multi-turn process. Yet, most existing benchmarks assume static environments with f...
Guangrui Li, Yaochen Xie, Yi Liu, Ziwei Dong, Xingyuan Pan, Tianqi Zheng, Jason Choi, Michael J. ...
InfoGatherer: Principled Information Seeking via Evidence Retrieval and Strategic Questioning
LLMs are increasingly deployed in high-stakes domains such as medical triage and legal assistance, often as document-grounded QA systems in which a user provides a description, relevant sources are...
Maksym Taranukhin, Shuyue Stella Li, Evangelos Milios, Geoff Pleiss, Yulia Tsvetkov, Vered Shwartz
LUMINA: LLM-Guided GPU Architecture Exploration via Bottleneck Analysis
GPU design space exploration (DSE) for modern AI workloads, such as Large-Language Model (LLM) inference, is challenging because of GPUs' vast, multi-modal design spaces, high simulation costs, and...
Tao Zhang, Rui Ma, Shuotao Xu, Peng Cheng, Yongqiang Xiong
Reference-guided Policy Optimization for Molecular Optimization via LLM Reasoning
Large language models (LLMs) benefit substantially from supervised fine-tuning (SFT) and reinforcement learning with verifiable rewards (RLVR) in reasoning tasks. However, these recipes perform poo...
Xuan Li, Zhanke Zhou, Zongze Li, Jiangchao Yao, Yu Rong, Lu Zhang, Bo Han
Building an Ensemble LLM Semantic Tagger for UN Security Council Resolutions
This paper introduces a new methodology for using LLM-based systems for accurate and efficient semantic tagging of UN Security Council resolutions. The main goal is to leverage LLM performance vari...
Hussein Ghaly
Lost in Stories: Consistency Bugs in Long Story Generation by LLMs
What happens when a storyteller forgets its own story? Large Language Models (LLMs) can now generate narratives spanning tens of thousands of words, but they often fail to maintain consistency thro...
Junjie Li, Xinrui Guo, Yuhao Wu, Roy Ka-Wei Lee, Hongzhi Li, Yutao Xie
Measuring Perceptions of Fairness in AI Systems: The Effects of Infra-marginality
Differences in data distributions between demographic groups, known as the problem of infra-marginality, complicate how people evaluate fairness in machine learning models. We present a user study ...
Schrasing Tong, Minseok Jung, Ilaria Liccardi, Lalana Kagal
Casimir-Polder potential on an excited atom near an atomic array
We develop a microscopic description of the fluctuation-mediated Casimir-Polder (CP) shifts on a 'test' two-level atom placed near a two-dimensional atomic array of two-level atoms. We derive the r...
Annyun Das, Kanu Sinha
Computational Pathology in the Era of Emerging Foundation and Agentic AI -- International Expert Perspectives on Clinical Integration and Translational Readiness
Recent breakthroughs in artificial intelligence through foundation models and agents have accelerated the evolution of computational pathology. Demonstrated performance gains reported across academ...
Qian Da, Yijiang Chen, Min Ju, Zheyi Ji, Albert Zhou, Wenwen Wang, Matthew A Abikenari, Philip Ch...
VerChol -- Grammar-First Tokenization for Agglutinative Languages
Tokenization is the foundational step in all large language model (LLM) pipelines, yet the dominant approach Byte Pair Encoding (BPE) and its variants is inherently script agnostic and optimized fo...
Prabhu Raja
Confidence Before Answering: A Paradigm Shift for Efficient LLM Uncertainty Estimation
Reliable deployment of large language models (LLMs) requires accurate uncertainty estimation. Existing methods are predominantly answer-first, producing confidence only after generating an answer, ...
Changcheng Li, Jiancan Wu, Hengheng Zhang, Zhengsu Chen, Guo An, Junxiang Qiu, Xiang Wang, Qi Tian