π― Top Personalized Recommendations
FIU
Why we think this paper is great for you:
This paper directly addresses the crucial challenge of achieving fairness in AI systems, especially when complete demographic data is unavailable. It explores methods to mitigate discriminatory outcomes, which is highly relevant to your focus.
Abstract
Fairness in artificial intelligence (AI) has become a growing concern due to discriminatory outcomes in AI-based decision-making systems. While various methods have been proposed to mitigate bias, most rely on complete demographic information, an assumption often impractical due to legal constraints and the risk of reinforcing discrimination. This survey examines fairness in AI when demographics are incomplete, addressing the gap between traditional approaches and real-world challenges. We introduce a novel taxonomy of fairness notions in this setting, clarifying their relationships and distinctions. Additionally, we summarize existing techniques that promote fairness beyond complete demographics and highlight open research questions to encourage further progress in the field.
AI Summary - The majority of existing AI fairness methods are impractical in real-world scenarios due to their reliance on complete demographic information, which is often unavailable due to privacy, legal, or individual choice constraints. [2]
- A novel taxonomy categorizes fairness notions for incomplete demographics into Rawlsian, Group, Counterfactual, Proxy, Individual, and Unawareness, distinguishing them by protection level (group/individual) and reliance on demographic data (explicit, latent proxy, or demographic-free). [2]
- Rawlsian fairness and certain adversarial learning approaches can address bias without explicit demographic information by focusing on worst-case performance or computationally identifiable error groups, though they may not guarantee statistical parity for specific demographic groups. [2]
- Proxy fairness methods, which infer or approximate demographic information from correlated features, are critical for practical applications but face challenges in ensuring the accuracy and representativeness of these proxies. [2]
- Individual fairness, based on similarity metrics, offers a demographic-free approach but struggles with defining appropriate similarity functions and can inadvertently create subgroup disparities. [2]
- Leveraging partial demographic information through uncertainty-aware attribute classifiers or disentanglement frameworks significantly enhances the accuracy of proxy demographics and fairness enforcement in demographic-scarce regimes. [2]
- Third-party involvement can facilitate fairness assessment or data preprocessing, but introduces significant challenges related to trust, data security, verification, and administrative costs. [2]
- Rawlsian Fairness: A fairness notion based on John Rawlsβs difference principle, aiming to improve the well-being of the least advantaged group by minimizing the variance of subgroup utilities, often without direct demographic information. [2]
- Proxy Fairness: A concept where fairness is measured or enforced using substitute or inferred demographic information (e.g., correlated features, predicted labels) when true demographic data is unavailable. [2]
- Individual Fairness (Lipschitz Condition): Ensures that similar individuals receive similar predictions, formalized as D(f(xi), f(xj)) β€ L D'(xi, xj), where D' is input space distance and D is output space distance. [2]
University of St Gallen
Why we think this paper is great for you:
This research delves into the complex interplay between human and AI biases, particularly how data imbalance affects decision-making. Understanding this interaction is key to developing more ethical and fair AI systems.
Abstract
Humans increasingly interact with artificial intelligence (AI) in decision-making. However, both AI and humans are prone to biases. While AI and human biases have been studied extensively in isolation, this paper examines their complex interaction. Specifically, we examined how class imbalance as an AI bias affects people's ability to appropriately rely on an AI-based decision-support system, and how it interacts with base rate neglect as a human bias. In a within-subject online study (N= 46), participants classified three diseases using an AI-based decision-support system trained on either a balanced or unbalanced dataset. We found that class imbalance disrupted participants' calibration of AI reliance. Moreover, we observed mutually reinforcing effects between class imbalance and base rate neglect, offering evidence of a compound human-AI bias. Based on these findings, we advocate for an interactionist perspective and further research into the mutually reinforcing effects of biases in human-AI interaction.
University of Notre Dame
Why we think this paper is great for you:
This paper presents a technical approach to ensuring model fairness and mitigating bias in deep learning applications. You will find its methods for achieving fairness while maintaining performance particularly insightful.
Abstract
As deep learning (DL) techniques become integral to various applications, ensuring model fairness while maintaining high performance has become increasingly critical, particularly in sensitive fields such as medical diagnosis. Although a variety of bias-mitigation methods have been proposed, many rely on computationally expensive debiasing strategies or suffer substantial drops in model accuracy, which limits their practicality in real-world, resource-constrained settings. To address this issue, we propose a fairness-oriented low rank factorization (LRF) framework that leverages singular value decomposition (SVD) to improve DL model fairness. Unlike traditional SVD, which is mainly used for model compression by decomposing and reducing weight matrices, our work shows that SVD can also serve as an effective tool for fairness enhancement. Specifically, we observed that elements in the unitary matrices obtained from SVD contribute unequally to model bias across groups defined by sensitive attributes. Motivated by this observation, we propose a method, named FairLRF, that selectively removes bias-inducing elements from unitary matrices to reduce group disparities, thus enhancing model fairness. Extensive experiments show that our method outperforms conventional LRF methods as well as state-of-the-art fairness-enhancing techniques. Additionally, an ablation study examines how major hyper-parameters may influence the performance of processed models. To the best of our knowledge, this is the first work utilizing SVD not primarily for compression but for fairness enhancement.
Memorial Sloan Kettering
Why we think this paper is great for you:
This work examines the ethical considerations and governance required for responsible data stewardship, especially in the sensitive domain of healthcare. It offers valuable insights into the evolving landscape of data ethics.
Abstract
Healthcare stands at a critical crossroads. Artificial Intelligence and modern computing are unlocking opportunities, yet their value lies in the data that fuels them. The value of healthcare data is no longer limited to individual patients. However, data stewardship and governance has not kept pace, and privacy-centric policies are hindering both innovation and patient protections. As healthcare moves toward a data-driven future, we must define reformed data stewardship that prioritizes patients' interests by proactively managing modern risks and opportunities while addressing key challenges in cost, efficacy, and accessibility.
Current healthcare data policies are rooted in 20th-century legislation shaped by outdated understandings of data-prioritizing perceived privacy over innovation and inclusion. While other industries thrive in a data-driven era, the evolution of medicine remains constrained by regulations that impose social rather than scientific boundaries. Large-scale aggregation is happening, but within opaque, closed systems. As we continue to uphold foundational ethical principles - autonomy, beneficence, nonmaleficence, and justice - there is a growing imperative to acknowledge they exist in evolving technological, social, and cultural realities.
Ethical principles should facilitate, rather than obstruct, dialogue on adapting to meet opportunities and address constraints in medical practice and healthcare delivery. The new ethics of data stewardship places patients first by defining governance that adapts to changing landscapes. It also rejects the legacy of treating perceived privacy as an unquestionable, guiding principle. By proactively redefining data stewardship norms, we can drive an era of medicine that promotes innovation, protects patients, and advances equity - ensuring future generations advance medical discovery and care.
UC Davis
Why we think this paper is great for you:
This study provides a thorough examination of biases present in large language models, highlighting their impact on fair outputs. It offers a comprehensive view of how biases manifest and the need for mitigation.
Abstract
Large Language Models (LLMs) inherit explicit and implicit biases from their training datasets. Identifying and mitigating biases in LLMs is crucial to ensure fair outputs, as they can perpetuate harmful stereotypes and misinformation. This study highlights the need to address biases in LLMs amid growing generative AI. We studied bias-specific benchmarks such as StereoSet and CrowSPairs to evaluate the existence of various biases in multiple generative models such as BERT and GPT 3.5. We proposed an automated Bias-Identification Framework to recognize various social biases in LLMs such as gender, race, profession, and religion. We adopted a two-pronged approach to detect explicit and implicit biases in text data. Results indicated fine-tuned models struggle with gender biases but excelled at identifying and avoiding racial biases. Our findings illustrated that despite having some success, LLMs often over-relied on keywords. To illuminate the capability of the analyzed LLMs in detecting implicit biases, we employed Bag-of-Words analysis and unveiled indications of implicit stereotyping within the vocabulary. To bolster the model performance, we applied an enhancement strategy involving fine-tuning models using prompting techniques and data augmentation of the bias benchmarks. The fine-tuned models exhibited promising adaptability during cross-dataset testing and significantly enhanced performance on implicit bias benchmarks, with performance gains of up to 20%.
UC Berkeley
Why we think this paper is great for you:
This paper offers a foundational perspective on the ethical and responsible deployment of artificial intelligence. It encourages a thoughtful approach to integrating AI into daily life.
Abstract
Artificial intelligence (AI) is no longer futuristic; it is a daily companion shaping our private and work lives. While AI simplifies our lives, its rise also invites us to rethink who we are - and who we wish to remain - as humans. Even if AI does not think, feel, or desire, it learns from our behavior, mirroring our collective values, biases, and aspirations. The question, then, is not what AI is, but what we are allowing it to become through data, computing power, and other parameters "teaching" it - and, even more importantly, who we are becoming through our relationship with AI.
As the EU AI Act and the Vienna Manifesto on Digital Humanism emphasize, technology must serve human dignity,social well-being, and democratic accountability. In our opinion, responsible use of AI is not only a matter of code nor law, but also of conscientious practice: how each of us engages and teaches others to use AI at home and at work. We propose Ten Commandments for the Wise and Responsible Use of AI are meant as guideline for this very engagement. They closely align with Floridi and Cowls' five guiding principles for AI in society - beneficence, non-maleficence, autonomy, justice, and explicability.
Colorado School of Mines
Why we think this paper is great for you:
This research investigates the practical implementation and effectiveness of privacy transparency mechanisms in real-world applications. It provides insights into how data practices are communicated to users.
Abstract
With the requirements and emphases on privacy transparency placed by regulations such as GDPR and CCPA, the Google Play Store requires Android developers to more responsibly communicate their apps' privacy practices to potential users by providing the proper information via the data safety, privacy policy, and permission manifest privacy transparency channels. However, it is unclear how effective those channels are in helping users make informed decisions in the app selection and installation process. In this article, we conducted a study for 190 participants to interact with our simulated privacy transparency channels of mobile apps. We quantitatively analyzed (supplemented by qualitative analysis) participants' responses to five sets of questions. We found that data safety provides the most intuitive user interfaces, privacy policy is most informative and effective, while permission manifest excels at raising participants' concerns about an app's overall privacy risks. These channels complement each other and should all be improved.