Abstract
The widespread adoption of AI models, especially foundation models (FMs), has
made a profound impact on numerous domains. However, it also raises significant
ethical concerns, including bias issues. Although numerous efforts have been
made to quantify and mitigate social bias in AI models, geographic bias (in
short, geo-bias) receives much less attention, which presents unique
challenges. While previous work has explored ways to quantify geo-bias, these
measures are model-specific (e.g., mean absolute deviation of LLM ratings) or
spatially implicit (e.g., average fairness scores of all spatial partitions).
We lack a model-agnostic, universally applicable, and spatially explicit
geo-bias evaluation framework that allows researchers to fairly compare the
geo-bias of different AI models and to understand what spatial factors
contribute to the geo-bias. In this paper, we establish an
information-theoretic framework for geo-bias evaluation, called GeoBS (Geo-Bias
Scores). We demonstrate the generalizability of the proposed framework by
showing how to interpret and analyze existing geo-bias measures under this
framework. Then, we propose three novel geo-bias scores that explicitly take
intricate spatial factors (multi-scalability, distance decay, and anisotropy)
into consideration. Finally, we conduct extensive experiments on 3 tasks, 8
datasets, and 8 models to demonstrate that both task-specific GeoAI models and
general-purpose foundation models may suffer from various types of geo-bias.
This framework will not only advance the technical understanding of geographic
bias but will also establish a foundation for integrating spatial fairness into
the design, deployment, and evaluation of AI systems.
University of Hong Kong
Abstract
Large language models (LLMs) are increasingly central to many applications,
raising concerns about bias, fairness, and regulatory compliance. This paper
reviews risks of biased outputs and their societal impact, focusing on
frameworks like the EU's AI Act and the Digital Services Act. We argue that
beyond constant regulation, stronger attention to competition and design
governance is needed to ensure fair, trustworthy AI. This is a preprint of the
Communications of the ACM article of the same title.
AI Insights - Loopholes in the EU AI Act let high‑risk LLMs evade scrutiny.
- Competition policy must join AI regulation to curb gatekeeper dominance.
- Design governance should embed bias‑mitigation and explainability from the start.
- Transparency must include audit trails and model lineage, not just a compliance form.
- Content‑moderation literature shows LLMs can amplify polarizing narratives if unchecked.
- IP and liability directives may miss ownership issues of AI‑generated text.
- Platform‑regulation studies suggest legal tools can restore user agency in AI communication.