Research

Paper

AI LLM March 05, 2026

SlideSparse: Fast and Flexible (2N-2):2N Structured Sparsity

Authors

Hanyong Shao, Yingbo Hao, Ting Song, Yan Xia, Di Zhang, Shaohan Huang, Xun Wu, Songchen Xu, Le Xu, Li Dong, Zewen Chi, Yi Zou, Furu Wei

Abstract

NVIDIA's 2:4 Sparse Tensor Cores deliver 2x throughput but demand strict 50% pruning -- a ratio that collapses LLM reasoning accuracy (Qwen3: 54% to 15%). Milder $(2N-2):2N$ patterns (e.g., 6:8, 25% pruning) preserve accuracy yet receive no hardware support, falling back to dense execution without any benefit from sparsity. We present SlideSparse, the first system to unlock Sparse Tensor Core acceleration for the $(2N-2):2N$ model family on commodity GPUs. Our Sliding Window Decomposition reconstructs any $(2N-2):2N$ weight block into $N-1$ overlapping 2:4-compliant windows without any accuracy loss; Activation Lifting fuses the corresponding activation rearrangement into per-token quantization at near-zero cost. Integrated into vLLM, SlideSparse is evaluated across various GPUs (A100, H100, B200, RTX 4090, RTX 5080, DGX-spark), precisions (FP4, INT8, FP8, BF16, FP16), and model families (Llama, Qwen, BitNet). On compute-bound workloads, the measured speedup ratio (1.33x) approaches the theoretical upper-bound $N/(N-1)=4/3$ at 6:8 weight sparsity in Qwen2.5-7B, establishing $(2N-2):2N$ as a practical path to accuracy-preserving LLM acceleration. Code available at https://github.com/bcacdwk/vllmbench.

Metadata

arXiv ID: 2603.05232
Provider: ARXIV
Primary Category: cs.LG
Published: 2026-03-05
Fetched: 2026-03-06 14:20

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.05232v1</id>\n    <title>SlideSparse: Fast and Flexible (2N-2):2N Structured Sparsity</title>\n    <updated>2026-03-05T14:49:16Z</updated>\n    <link href='https://arxiv.org/abs/2603.05232v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.05232v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>NVIDIA's 2:4 Sparse Tensor Cores deliver 2x throughput but demand strict 50% pruning -- a ratio that collapses LLM reasoning accuracy (Qwen3: 54% to 15%). Milder $(2N-2):2N$ patterns (e.g., 6:8, 25% pruning) preserve accuracy yet receive no hardware support, falling back to dense execution without any benefit from sparsity. We present SlideSparse, the first system to unlock Sparse Tensor Core acceleration for the $(2N-2):2N$ model family on commodity GPUs. Our Sliding Window Decomposition reconstructs any $(2N-2):2N$ weight block into $N-1$ overlapping 2:4-compliant windows without any accuracy loss; Activation Lifting fuses the corresponding activation rearrangement into per-token quantization at near-zero cost. Integrated into vLLM, SlideSparse is evaluated across various GPUs (A100, H100, B200, RTX 4090, RTX 5080, DGX-spark), precisions (FP4, INT8, FP8, BF16, FP16), and model families (Llama, Qwen, BitNet). On compute-bound workloads, the measured speedup ratio (1.33x) approaches the theoretical upper-bound $N/(N-1)=4/3$ at 6:8 weight sparsity in Qwen2.5-7B, establishing $(2N-2):2N$ as a practical path to accuracy-preserving LLM acceleration. Code available at https://github.com/bcacdwk/vllmbench.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.LG'/>\n    <published>2026-03-05T14:49:16Z</published>\n    <arxiv:primary_category term='cs.LG'/>\n    <author>\n      <name>Hanyong Shao</name>\n    </author>\n    <author>\n      <name>Yingbo Hao</name>\n    </author>\n    <author>\n      <name>Ting Song</name>\n    </author>\n    <author>\n      <name>Yan Xia</name>\n    </author>\n    <author>\n      <name>Di Zhang</name>\n    </author>\n    <author>\n      <name>Shaohan Huang</name>\n    </author>\n    <author>\n      <name>Xun Wu</name>\n    </author>\n    <author>\n      <name>Songchen Xu</name>\n    </author>\n    <author>\n      <name>Le Xu</name>\n    </author>\n    <author>\n      <name>Li Dong</name>\n    </author>\n    <author>\n      <name>Zewen Chi</name>\n    </author>\n    <author>\n      <name>Yi Zou</name>\n    </author>\n    <author>\n      <name>Furu Wei</name>\n    </author>\n  </entry>"
}