Research

Paper

TESTING March 24, 2026

MVPBench: A Multi-Video Perception Evaluation Benchmark for Multi-Modal Video Understanding

Authors

Purui Bai, Tao Wu, Jiayang Sun, Xinyue Liu, Huaibo Huang, Ran He

Abstract

The rapid progress of Large Language Models (LLMs) has spurred growing interest in Multi-modal LLMs (MLLMs) and motivated the development of benchmarks to evaluate their perceptual and comprehension abilities. Existing benchmarks, however, are limited to static images or single videos, overlooking the complex interactions across multiple videos. To address this gap, we introduce the Multi-Video Perception Evaluation Benchmark (MVPBench), a new benchmark featuring 14 subtasks across diverse visual domains designed to evaluate models on extracting relevant information from video sequences to make informed decisions. MVPBench includes 5K question-answering tests involving 2.7K video clips sourced from existing datasets and manually annotated clips. Extensive evaluations reveal that current models struggle to process multi-video inputs effectively, underscoring substantial limitations in their multi-video comprehension. We anticipate MVPBench will drive advancements in multi-video perception.

Metadata

arXiv ID: 2603.22756
Provider: ARXIV
Primary Category: cs.CV
Published: 2026-03-24
Fetched: 2026-03-25 06:02

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.22756v1</id>\n    <title>MVPBench: A Multi-Video Perception Evaluation Benchmark for Multi-Modal Video Understanding</title>\n    <updated>2026-03-24T03:33:07Z</updated>\n    <link href='https://arxiv.org/abs/2603.22756v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.22756v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>The rapid progress of Large Language Models (LLMs) has spurred growing interest in Multi-modal LLMs (MLLMs) and motivated the development of benchmarks to evaluate their perceptual and comprehension abilities. Existing benchmarks, however, are limited to static images or single videos, overlooking the complex interactions across multiple videos. To address this gap, we introduce the Multi-Video Perception Evaluation Benchmark (MVPBench), a new benchmark featuring 14 subtasks across diverse visual domains designed to evaluate models on extracting relevant information from video sequences to make informed decisions. MVPBench includes 5K question-answering tests involving 2.7K video clips sourced from existing datasets and manually annotated clips. Extensive evaluations reveal that current models struggle to process multi-video inputs effectively, underscoring substantial limitations in their multi-video comprehension. We anticipate MVPBench will drive advancements in multi-video perception.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.CV'/>\n    <published>2026-03-24T03:33:07Z</published>\n    <arxiv:comment>15 pages, 7 figures, accepted by IJCNN 2026, code and dataset available at https://github.com/MVPBench/MVPBench</arxiv:comment>\n    <arxiv:primary_category term='cs.CV'/>\n    <author>\n      <name>Purui Bai</name>\n    </author>\n    <author>\n      <name>Tao Wu</name>\n    </author>\n    <author>\n      <name>Jiayang Sun</name>\n    </author>\n    <author>\n      <name>Xinyue Liu</name>\n    </author>\n    <author>\n      <name>Huaibo Huang</name>\n    </author>\n    <author>\n      <name>Ran He</name>\n    </author>\n  </entry>"
}