Research

Papers

Research papers from arXiv and related sources

Total: 4513 AI/LLM: 2483 Testing: 2030
AI LLM

Design-Specification Tiling for ICL-based CAD Code Generation

Large language models (LLMs) have demonstrated remarkable capabilities in code generation, yet they underperform on domain-specific tasks such as Computer-Aided Design (CAD) code generation due to ...

Yali Du, San-Zhuo Xi, Hui Sun, Ming Li

2603.12712 2026-03-13
AI LLM

AI Planning Framework for LLM-Based Web Agents

Developing autonomous agents for web-based tasks is a core challenge in AI. While Large Language Model (LLM) agents can interpret complex user requests, they often operate as black boxes, making it...

Orit Shahnovsky, Rotem Dror

2603.12710 2026-03-13
AI LLM

Cost-Efficient Multimodal LLM Inference via Cross-Tier GPU Heterogeneity

Multimodal large language model (MLLM) inference splits into two phases with opposing hardware demands: vision encoding is compute-bound, while language generation is memory-bandwidth-bound. We sho...

Donglin Yu

2603.12707 2026-03-13
AI LLM

FGTR: Fine-Grained Multi-Table Retrieval via Hierarchical LLM Reasoning

With the rapid advancement of large language models (LLMs), growing efforts have been made on LLM-based table retrieval. However, existing studies typically focus on single-table query, and impleme...

Chaojie Sun, Bin Cao, Tiantian Li, Chenyu Hou, Ruizhe Li, Qing Fan

2603.12702 2026-03-13
AI LLM

Seeing Eye to Eye: Enabling Cognitive Alignment Through Shared First-Person Perspective in Human-AI Collaboration

Despite advances in multimodal AI, current vision-based assistants often remain inefficient in collaborative tasks. We identify two key gulfs: a communication gulf, where users must translate rich ...

Zhuyu Teng, Pei Chen, Yichen Cai, Ruoqing Lu, Zhaoqu Jiang, Jiayang Li, Weitao You, Lingyun Sun

2603.12701 2026-03-13
AI LLM

Experimental evidence of progressive ChatGPT models self-convergence

Large Language Models (LLMs) that undergo recursive training on synthetically generated data are susceptible to model collapse, a phenomenon marked by the generation of meaningless output. Existing...

Konstantinos F. Xylogiannopoulos, Petros Xanthopoulos, Panagiotis Karampelas, Georgios A. Bakamitsos

2603.12683 2026-03-13
AI LLM

Colluding LoRA: A Composite Attack on LLM Safety Alignment

We introduce Colluding LoRA (CoLoRA), an attack in which each adapter appears benign and plausibly functional in isolation, yet their linear composition consistently compromises safety. Unlike atta...

Sihao Ding

2603.12681 2026-03-13
AI LLM

MetaKE: Meta-learning Aligned Knowledge Editing via Bi-level Optimization

Knowledge editing (KE) aims to precisely rectify specific knowledge in Large Language Models (LLMs) without disrupting general capabilities. State-of-the-art methods suffer from an open-loop contro...

Shuxin Liu, Ou Wu

2603.12677 2026-03-13
AI LLM

HyGra: Accelerating Network-State Simulation for LLM Training in DCNs via Adaptive Packet-Flow Granularity

In recent years, large language models (LLMs) have driven substantial intelligent transformation across diverse industries. Commercial LLM training is typically performed over data center networks ...

Wenyi Wang, Zheng Wu, Yanmeng Wang, Haolin Mao, Lei Han, Gaogang Xie, Fu Xiao

2603.12671 2026-03-13
AI LLM

RetroReasoner: A Reasoning LLM for Strategic Retrosynthesis Prediction

Retrosynthesis prediction is a core task in organic synthesis that aims to predict reactants for a given product molecule. Traditionally, chemists select a plausible bond disconnection and derive c...

Hanbum Ko, Chanhui Lee, Ye Rin Kim, Rodrigo Hormazabal, Sehui Han, Sungbin Lim, Sungwoong Kim

2603.12666 2026-03-13
AI LLM

From Text to Forecasts: Bridging Modality Gap with Temporal Evolution Semantic Space

Incorporating textual information into time-series forecasting holds promise for addressing event-driven non-stationarity; however, a fundamental modality gap hinders effective fusion: textual desc...

Lehui Li, Yuyao Wang, Jisheng Yan, Wei Zhang, Jinliang Deng, Haoliang Sun, Zhongyi Han, Yongshun ...

2603.12664 2026-03-13
AI LLM

Evaluation of TCP Congestion Control for Public High-Performance Wide-Area Networks

Practitioners of a growing number of scientific and artificial-intelligence (AI) applications use High-Performance Wide-Area Networks (HP-WANs) for moving massive data sets between remote facilitie...

Fatih Berkay Sarpkaya, Andrea Francini, Bilgehan Erman, Shivendra Panwar

2603.12660 2026-03-13
AI LLM

Continual Learning in Large Language Models: Methods, Challenges, and Opportunities

Continual learning (CL) has emerged as a pivotal paradigm to enable large language models (LLMs) to dynamically adapt to evolving knowledge and sequential tasks while mitigating catastrophic forget...

Hongyang Chen, Zhongwu Sun, Hongfei Ye, Kunchi Li, Xuemin Lin

2603.12658 2026-03-13
AI LLM

A Standards-Aligned Coordination Framework for Edge-Enhanced Collaborative Healthcare in 6G Networks

Mission-critical healthcare applications including real-time intensive care monitoring, ambulance-to-hospital orchestration, and distributed medical imaging inference require workflow-level, time-b...

Liuwang Kang, Fan Wang, Yuzhang Huang, Shang Yan, Jianbin Zheng, Wenbin Lei, Konstantin Yakovlev,...

2603.12653 2026-03-13
AI LLM

98$\times$ Faster LLM Routing Without a Dedicated GPU: Flash Attention, Prompt Compression, and Near-Streaming for the vLLM Semantic Router

System-level routers that intercept LLM requests for safety classification, domain routing, and PII detection must be both fast and operationally lightweight: they should add minimal latency to eve...

Xunzhuo Liu, Bowei He, Xue Liu, Andy Luo, Haichen Zhang, Huamin Chen

2603.12646 2026-03-13
AI LLM

LightMoE: Reducing Mixture-of-Experts Redundancy through Expert Replacing

Mixture-of-Experts (MoE) based Large Language Models (LLMs) have demonstrated impressive performance and computational efficiency. However, their deployment is often constrained by substantial memo...

Jiawei Hao, Zhiwei Hao, Jianyuan Guo, Li Shen, Yong Luo, Han Hu, Dan Zeng

2603.12645 2026-03-13
AI LLM

Uncovering Security Threats and Architecting Defenses in Autonomous Agents: A Case Study of OpenClaw

The rapid evolution of Large Language Models (LLMs) into autonomous, tool-calling agents has fundamentally altered the cybersecurity landscape. Frameworks like OpenClaw grant AI systems operating-s...

Zonghao Ying, Xiao Yang, Siyang Wu, Yumeng Song, Yang Qu, Hainan Li, Tianlin Li, Jiakai Wang, Ais...

2603.12644 2026-03-13
AI LLM

RoboStereo: Dual-Tower 4D Embodied World Models for Unified Policy Optimization

Scalable Embodied AI faces fundamental constraints due to prohibitive costs and safety risks of real-world interaction. While Embodied World Models (EWMs) offer promise through imagined rollouts, e...

Ruicheng Zhang, Guangyu Chen, Zunnan Xu, Zihao Liu, Zhizhou Zhong, Mingyang Zhang, Jun Zhou, Xiu Li

2603.12639 2026-03-13
AI LLM

Using a Human-AI Teaming Approach to Create and Curate Scientific Datasets with the SCILIRE System

The rapid growth of scientific literature has made manual extraction of structured knowledge increasingly impractical. To address this challenge, we introduce SCILIRE, a system for creating dataset...

Necva Bölücü, Jessica Irons, Changhyun Lee, Brian Jin, Maciej Rybinski, Huichen Yang, Andreas Due...

2603.12638 2026-03-13
AI LLM

Spend Less, Reason Better: Budget-Aware Value Tree Search for LLM Agents

Test-time scaling has become a dominant paradigm for improving LLM agent reliability, yet current approaches treat compute as an abundant resource, allowing agents to exhaust token and tool budgets...

Yushu Li, Wenlong Deng, Jiajin Li, Xiaoxiao Li

2603.12634 2026-03-13