Research

Paper

AI LLM March 11, 2026

AILS-NTUA at SemEval-2026 Task 8: Evaluating Multi-Turn RAG Conversations

Authors

Dimosthenis Athanasiou, Maria Lymperaiou, Giorgos Filandrianos, Athanasios Voulodimos, Giorgos Stamou

Abstract

We present the AILS-NTUA system for SemEval-2026 Task 8 (MTRAGEval), addressing all three subtasks of multi-turn retrieval-augmented generation: passage retrieval (A), reference-grounded response generation (B), and end-to-end RAG (C). Our unified architecture is built on two principles: (i) a query-diversity-over-retriever-diversity strategy, where five complementary LLM-based query reformulations are issued to a single corpus-aligned sparse retriever and fused via variance-aware nested Reciprocal Rank Fusion; and (ii) a multistage generation pipeline that decomposes grounded generation into evidence span extraction, dual-candidate drafting, and calibrated multi-judge selection. Our system ranks 1st in Task A (nDCG@5: 0.5776, +20.5% over the strongest baseline) and 2nd in Task B (HM: 0.7698). Empirical analysis shows that query diversity over a well-aligned retriever outperforms heterogeneous retriever ensembling, and that answerability calibration-rather than retrieval coverage-is the primary bottleneck in end-to-end performance.

Metadata

arXiv ID: 2603.10524
Provider: ARXIV
Primary Category: cs.CL
Published: 2026-03-11
Fetched: 2026-03-12 04:21

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.10524v1</id>\n    <title>AILS-NTUA at SemEval-2026 Task 8: Evaluating Multi-Turn RAG Conversations</title>\n    <updated>2026-03-11T08:28:31Z</updated>\n    <link href='https://arxiv.org/abs/2603.10524v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.10524v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>We present the AILS-NTUA system for SemEval-2026 Task 8 (MTRAGEval), addressing all three subtasks of multi-turn retrieval-augmented generation: passage retrieval (A), reference-grounded response generation (B), and end-to-end RAG (C). Our unified architecture is built on two principles: (i) a query-diversity-over-retriever-diversity strategy, where five complementary LLM-based query reformulations are issued to a single corpus-aligned sparse retriever and fused via variance-aware nested Reciprocal Rank Fusion; and (ii) a multistage generation pipeline that decomposes grounded generation into evidence span extraction, dual-candidate drafting, and calibrated multi-judge selection. Our system ranks 1st in Task A (nDCG@5: 0.5776, +20.5% over the strongest baseline) and 2nd in Task B (HM: 0.7698). Empirical analysis shows that query diversity over a well-aligned retriever outperforms heterogeneous retriever ensembling, and that answerability calibration-rather than retrieval coverage-is the primary bottleneck in end-to-end performance.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.CL'/>\n    <published>2026-03-11T08:28:31Z</published>\n    <arxiv:primary_category term='cs.CL'/>\n    <author>\n      <name>Dimosthenis Athanasiou</name>\n    </author>\n    <author>\n      <name>Maria Lymperaiou</name>\n    </author>\n    <author>\n      <name>Giorgos Filandrianos</name>\n    </author>\n    <author>\n      <name>Athanasios Voulodimos</name>\n    </author>\n    <author>\n      <name>Giorgos Stamou</name>\n    </author>\n  </entry>"
}