Research

Paper

AI LLM March 12, 2026

Locating Demographic Bias at the Attention-Head Level in CLIP's Vision Encoder

Authors

Alaa Yasser, Kittipat Phunjanna, Marcos Escudero Viñolo, Catarina Barata, Jenny Benois-Pineau

Abstract

Standard fairness audits of foundation models quantify that a model is biased, but not where inside the network the bias resides. We propose a mechanistic fairness audit that combines projected residual-stream decomposition, zero-shot Concept Activation Vectors, and bias-augmented TextSpan analysis to locate demographic bias at the level of individual attention heads in vision transformers. As a feasibility case study, we apply this pipeline to the CLIP ViT-L-14 encoder on 42 profession classes of the FACET benchmark, auditing both gender and age bias. For gender, the pipeline identifies four terminal-layer heads whose ablation reduces global bias (Cramer's V: 0.381 -> 0.362) while marginally improving accuracy (+0.42%); a layer-matched random control confirms that this effect is specific to the identified heads. A single head in the final layer contributes to the majority of the reduction in the most stereotyped classes, and class-level analysis shows that corrected predictions shift toward the correct occupation. For age, the same pipeline identifies candidate heads, but ablation produces weaker and less consistent effects, suggesting that age bias is encoded more diffusely than gender bias in this model. These results provide preliminary evidence that head-level bias localisation is feasible for discriminative vision encoders and that the degree of localisability may vary across protected attributes. keywords: Bias . CLIP . Mechanistic Interpretability . Vision Transformer . Fairness

Metadata

arXiv ID: 2603.11793
Provider: ARXIV
Primary Category: cs.CV
Published: 2026-03-12
Fetched: 2026-03-14 05:03

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.11793v1</id>\n    <title>Locating Demographic Bias at the Attention-Head Level in CLIP's Vision Encoder</title>\n    <updated>2026-03-12T10:54:26Z</updated>\n    <link href='https://arxiv.org/abs/2603.11793v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.11793v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Standard fairness audits of foundation models quantify that a model is biased, but not where inside the network the bias resides. We propose a mechanistic fairness audit that combines projected residual-stream decomposition, zero-shot Concept Activation Vectors, and bias-augmented TextSpan analysis to locate demographic bias at the level of individual attention heads in vision transformers. As a feasibility case study, we apply this pipeline to the CLIP ViT-L-14 encoder on 42 profession classes of the FACET benchmark, auditing both gender and age bias. For gender, the pipeline identifies four terminal-layer heads whose ablation reduces global bias (Cramer's V: 0.381 -&gt; 0.362) while marginally improving accuracy (+0.42%); a layer-matched random control confirms that this effect is specific to the identified heads. A single head in the final layer contributes to the majority of the reduction in the most stereotyped classes, and class-level analysis shows that corrected predictions shift toward the correct occupation. For age, the same pipeline identifies candidate heads, but ablation produces weaker and less consistent effects, suggesting that age bias is encoded more diffusely than gender bias in this model. These results provide preliminary evidence that head-level bias localisation is feasible for discriminative vision encoders and that the degree of localisability may vary across protected attributes.\n  keywords: Bias . CLIP . Mechanistic Interpretability . Vision Transformer . Fairness</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.CV'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.AI'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.CY'/>\n    <published>2026-03-12T10:54:26Z</published>\n    <arxiv:comment>14 pages, 6 tables, 2 figures. Work conducted during IPCV-AI Erasmus Mundus Master</arxiv:comment>\n    <arxiv:primary_category term='cs.CV'/>\n    <author>\n      <name>Alaa Yasser</name>\n    </author>\n    <author>\n      <name>Kittipat Phunjanna</name>\n    </author>\n    <author>\n      <name>Marcos Escudero Viñolo</name>\n    </author>\n    <author>\n      <name>Catarina Barata</name>\n    </author>\n    <author>\n      <name>Jenny Benois-Pineau</name>\n    </author>\n  </entry>"
}