Monash University, Monash
Abstract
Bias in large language models (LLMs) remains a persistent challenge,
manifesting in stereotyping and unfair treatment across social groups. While
prior research has primarily focused on individual models, the rise of
multi-agent systems (MAS), where multiple LLMs collaborate and communicate,
introduces new and largely unexplored dynamics in bias emergence and
propagation. In this work, we present a comprehensive study of stereotypical
bias in MAS, examining how internal specialization, underlying LLMs and
inter-agent communication protocols influence bias robustness, propagation, and
amplification. We simulate social contexts where agents represent different
social groups and evaluate system behavior under various interaction and
adversarial scenarios. Experiments on three bias benchmarks reveal that MAS are
generally less robust than single-agent systems, with bias often emerging early
through in-group favoritism. However, cooperative and debate-based
communication can mitigate bias amplification, while more robust underlying
LLMs improve overall system stability. Our findings highlight critical factors
shaping fairness and resilience in multi-agent LLM systems.
AI Insights - Cooperative, debate, and competitive protocols shape how agents negotiate, each offering a distinct path to reduce or amplify bias.
- Evaluating evidence-based reasoning across multiple viewpoints is key to preventing stereotype-driven conclusions.
- Safety guidelines for AI include passive safeguards and active countermeasures against jailbreak attempts and biased outputs.
- The Art of Reasoning by Kelley equips designers with tools to scrutinize logic and spot hidden biases.
- Thinking, Fast and Slow by Kahneman reveals why intuitive judgments often lean on stereotypes.
- Crowdsourced Data for Evaluating Social Bias in Language Models (Nangia et al., 2020) provides a real-world benchmark for bias detection.
- StereoSet (Nadeem et al., 2021) offers a balanced set of stereotype and anti-stereotype examples to test fairness.
Ontario Tech University
Abstract
Artificial Intelligence (AI) has emerged as both a continuation of historical
technological revolutions and a potential rupture with them. This paper argues
that AI must be viewed simultaneously through three lenses: \textit{risk},
where it resembles nuclear technology in its irreversible and global
externalities; \textit{transformation}, where it parallels the Industrial
Revolution as a general-purpose technology driving productivity and
reorganization of labor; and \textit{continuity}, where it extends the
fifty-year arc of computing revolutions from personal computing to the internet
to mobile. Drawing on historical analogies, we emphasize that no past
transition constituted a strict singularity: disruptive shifts eventually
became governable through new norms and institutions.
We examine recurring patterns across revolutions -- democratization at the
usage layer, concentration at the production layer, falling costs, and
deepening personalization -- and show how these dynamics are intensifying in
the AI era. Sectoral analysis illustrates how accounting, law, education,
translation, advertising, and software engineering are being reshaped as
routine cognition is commoditized and human value shifts to judgment, trust,
and ethical responsibility. At the frontier, the challenge of designing moral
AI agents highlights the need for robust guardrails, mechanisms for moral
generalization, and governance of emergent multi-agent dynamics.
We conclude that AI is neither a singular break nor merely incremental
progress. It is both evolutionary and revolutionary: predictable in its median
effects yet carrying singularity-class tail risks. Good outcomes are not
automatic; they require coupling pro-innovation strategies with safety
governance, ensuring equitable access, and embedding AI within a human order of
responsibility.
AI Insights - AI is reshaping law, education, translation, and software engineering by commodifying routine reasoning and shifting scarcity to judgment, trust, and ethical responsibility.
- Historical analogies show past tech revolutions became governable through new norms, standards, and institutions, dispelling the singularity myth.
- Moral AI demands interdisciplinary collaboration to engineer reliability, articulate values, and build accountability regimes for emergent multiâagent systems.
- Viewing AI as mathematics and infrastructureânot magicâhelps embed it in a human order of responsibility, balancing benefits and risks.
- Benigerâs âThe Control Revolutionâ traces how information societies reorganize economies, offering a useful lens for AIâs systemic effects.