<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Proceedings of Machine Learning Research</title>
    <description>Proceedings of AAAI 2026 Workshop on Bias in Multimodal AI
  Held in Singapore Expo, Tampines, Singapore on 25 January 2026

Published as Volume 332 by the Proceedings of Machine Learning Research on 09 April 2026.

Volume Edited by:
  Soyeon Caren Han
  Rina Carines Cabral

Series Editors:
  Neil D. Lawrence
</description>
    <link>https://proceedings.mlr.press/v332/</link>
    <atom:link href="https://proceedings.mlr.press/v332/feed.xml" rel="self" type="application/rss+xml"/>
    <pubDate>Tue, 05 May 2026 06:19:50 +0000</pubDate>
    <lastBuildDate>Tue, 05 May 2026 06:19:50 +0000</lastBuildDate>
    <generator>Jekyll v3.10.0</generator>
    
      <item>
        <title>Expert Collapse and Compositional Failure in Simple Multimodal MoE</title>
        <description>Mixture-of-Experts (MoE) is a technique that uses multiple MLPs at each transformer layer, rather than using a single MLP. MoE architectures are hypothesised to create specialised experts, but this is often inferred rather than quantitatively measured. Further, this specialisation is in conflict with standard load-balancing losses that promote more uniform load distribution, though can encourage expert redundancy. We construct a novel two-unimodal-expert (vision/text) MoE testbed and use a three-stage protocol to first force specialisation with uni-modal data (Stage 2), then test its stability during fine-tuning (Stage 3). Our findings demonstrate that while Stage 2 successfully creates specialised experts, this specialisation persists only at the object-level. In Stage 3, the standard multimodal loss actively overwrites this structure, causing the latent space to default to only clustering by &lt;i&gt;modality&lt;/i&gt;, rather than by &lt;i&gt;concept&lt;/i&gt;. We identify the mechanism as &lt;i&gt;layer-level expert collapse&lt;/i&gt;. Furthermore, a case study on compositional binding reveals that even when specialisation is present, it captures monolithic objects (e.g., ’car’) but fails to bind attributes like colour, perhaps highlighting a source of bias in multimodal representation.</description>
        <pubDate>Thu, 09 Apr 2026 00:00:00 +0000</pubDate>
        <link>https://proceedings.mlr.press/v332/ticinovic26a.html</link>
        <guid isPermaLink="true">https://proceedings.mlr.press/v332/ticinovic26a.html</guid>
        
        
      </item>
    
      <item>
        <title>Physics-based phenomenological characterization of cross-modal bias in multimodal models</title>
        <description>The term ’algorithmic fairness’ is used to evaluate whether AI models operate fairly in both comparative (where fairness is understood as formal equality, such as “treat like cases as like”) and non-comparative (where unfairness arises from the model’s inaccuracy, arbitrariness, or inscrutability) contexts. Recent advances in multimodal large language models (MLLMs) are breaking new ground in multimodal understanding, reasoning, and generation; however, we argue that inconspicuous distortions arising from complex multimodal interaction dynamics can lead to systematic bias. The purpose of this position paper is twofold: first, it is intended to acquaint AI researchers with phenomenological explainable approaches that rely on the physical entities that the machine experiences during training/inference, as opposed to the traditional cognitivist symbolic account or metaphysical approaches; second, it is to state that this phenomenological doctrine will be practically useful for tackling algorithmic fairness issues in MLLMs. We develop a surrogate physics-based model that describes transformer dynamics (i.e., semantic network structure and self-cross-attention) to analyze the dynamics of cross-modal bias in MLLM, which are not fully captured by conventional embedding- or representation-level analyses. We support this position through multi-input diagnostic experiments: 1) perturbation-based analyses of emotion classification using Qwen2.5-Omni and Gemma 3n, and 2) dynamical analysis of Lorenz chaotic time-series prediction through the physical surrogate. Across two architecturally distinct MLLMs, we show that multimodal inputs can reinforce modality dominance rather than mitigate it, as revealed by structured error-attractor patterns under systematic label perturbation, complemented by dynamical analysis.</description>
        <pubDate>Thu, 09 Apr 2026 00:00:00 +0000</pubDate>
        <link>https://proceedings.mlr.press/v332/kim26a.html</link>
        <guid isPermaLink="true">https://proceedings.mlr.press/v332/kim26a.html</guid>
        
        
      </item>
    
      <item>
        <title>Cultural Representation Bias and Alignment Divergence in Large Language Models</title>
        <description>Large Language Models (LLMs) are increasingly deployed as globally applicable tools, yet their internal mechanisms remain deeply conditioned by regional cultural schemas. Through a three-stage cultural audit comparing Western and Chinese LLMs, we identify a systematic divergence in how these models prioritize core social values. Quantitative results reveal a stark contrast: Western models consistently prioritize individualistic constructs like &quot;Autonomy&quot;, while Chinese models favor relational ethics such as &quot;Harmony&quot;. We attribute this divergence to a two-stage “cultural imprinting” process during large-scale pre-training and subsequent human-feedback refinement. This cumulative imprinting suggests that aligning AI to a single set of cultural standards may inadvertently impose a restrictive lens on the model, creating a risk where cultural differences are misconstrued as moral or behavioral deficits. Consequently, we advocate for the development of locally-aligned models and multidisciplinary fairness metrics to ensure global representation equity in the era of foundation AI.</description>
        <pubDate>Thu, 09 Apr 2026 00:00:00 +0000</pubDate>
        <link>https://proceedings.mlr.press/v332/kan26a.html</link>
        <guid isPermaLink="true">https://proceedings.mlr.press/v332/kan26a.html</guid>
        
        
      </item>
    
      <item>
        <title>Annotator Risk Preference as a Catalyst for Systemic Bias in Multimodal AI</title>
        <description>As artificial intelligence evolves toward multimodal cognition, systems are moving beyond unimodal dependencies to integrate visual, auditory, and linguistic dimensions, thereby simulating human perception of reality. However, this increased complexity not only enhances expressiveness but also opens more insidious channels for bias infiltration. Existing research largely focuses on the demographic attributes of annotators (e.g., race, gender) while overlooking critical variables within the dimension of decision psychology (Ferrara, 2024; Sap et al., 2022). Among these, risk preference acts as a core driver of individual decision-making, exerting a subtle anchoring effect during the multimodal annotation process. When annotators confront materials characterized by high ambiguity, fuzziness, or potential social sensitivity, their intrinsic risk tolerance directly dictates label polarity, the degree of neutralization, and sensitivity toward minority attributes.</description>
        <pubDate>Thu, 09 Apr 2026 00:00:00 +0000</pubDate>
        <link>https://proceedings.mlr.press/v332/hu26a.html</link>
        <guid isPermaLink="true">https://proceedings.mlr.press/v332/hu26a.html</guid>
        
        
      </item>
    
      <item>
        <title>Preface</title>
        <description>Preface</description>
        <pubDate>Thu, 09 Apr 2026 00:00:00 +0000</pubDate>
        <link>https://proceedings.mlr.press/v332/han26a.html</link>
        <guid isPermaLink="true">https://proceedings.mlr.press/v332/han26a.html</guid>
        
        
      </item>
    
  </channel>
</rss>
