Papers from 15 to 19 September, 2025

Here are the personalized paper recommendations sorted by most relevant
LLMs for Compliance
👍 👎 ♄ Save
EPFL, ETH Zurich, CSCS, H
Abstract
We present Apertus, a fully open suite of large language models (LLMs) designed to address two systemic shortcomings in today's open model ecosystem: data compliance and multilingual representation. Unlike many prior models that release weights without reproducible data pipelines or regard for content-owner rights, Apertus models are pretrained exclusively on openly available data, retroactively respecting robots.txt exclusions and filtering for non-permissive, toxic, and personally identifiable content. To mitigate risks of memorization, we adopt the Goldfish objective during pretraining, strongly suppressing verbatim recall of data while retaining downstream task performance. The Apertus models also expand multilingual coverage, training on 15T tokens from over 1800 languages, with ~40% of pretraining data allocated to non-English content. Released at 8B and 70B scales, Apertus approaches state-of-the-art results among fully open models on multilingual benchmarks, rivalling or surpassing open-weight counterparts. Beyond model weights, we release all scientific artifacts from our development cycle with a permissive license, including data preparation scripts, checkpoints, evaluation suites, and training code, enabling transparent audit and extension.
AI Insights
  • Swiss AI Charter in Apertus’ prompt enforces transparency, accountability, and human‑value respect.
  • Prompt lists Swiss national languages, ensuring culturally relevant multilingual responses.
  • Goldfish objective cuts verbatim recall, balancing privacy with downstream performance.
  • All weights, code, and evaluation scripts are Apache 2.0‑licensed for full auditability.
  • 40 % non‑English tokens give Apertus an edge on low‑resource language benchmarks.
  • Instructions prioritize accuracy, transparently separate facts from speculation, and allow evidence‑based revision.
  • Read “The Swiss AI Charter” and “Apertus: A Multilingual AI Language Model for General Knowledge and Reasoning Tasks.”
👍 👎 ♄ Save
Cornell Tech
Abstract
Large language models equipped with Web search, information retrieval tools, and other agentic capabilities are beginning to supplant traditional search engines. As users start to rely on LLMs for information on many topics, including controversial and debatable issues, it is important to understand how the stances and opinions expressed in LLM outputs are influenced by the documents they use as their information sources. In this paper, we present MillStone, the first benchmark that aims to systematically measure the effect of external arguments on the stances that LLMs take on controversial issues (not all of them political). We apply MillStone to nine leading LLMs and measure how ``open-minded'' they are to arguments supporting opposite sides of these issues, whether different LLMs agree with each other, which arguments LLMs find most persuasive, and whether these arguments are the same for different LLMs. In general, we find that LLMs are open-minded on most issues. An authoritative source of information can easily sway an LLM's stance, highlighting the importance of source selection and the risk that LLM-based information retrieval and search systems can be manipulated.
AI Insights
  • MillStone probes nine leading LLMs, quantifying how authoritative sources shift stances.
  • Opus uniquely drops refusal when fed balanced arguments, revealing a rare neutrality.
  • The benchmark identifies which arguments most sway each model, exposing hidden biases.
  • Cross‑model agreement analysis shows divergent persuasive cues across architectures.
  • Findings warn that LLM‑powered search can be gamed by manipulating source credibility.
  • “Controversial topics” are formally defined as issues with documented public disagreement.
  • Recommended reading includes BERT and RoBERTa papers for foundational bias‑evaluation techniques.
AI Governance
👍 👎 ♄ Save
ADAPT Centre, Trinity Col
Paper visualization
Rate this image: 😍 👍 👎
Abstract
The upsurge of policies and guidelines that aim to ensure Artificial Intelligence (AI) systems are safe and trustworthy has led to a fragmented landscape of AI governance. The European Union (EU) is a key actor in the development of such policies and guidelines. Its High-Level Expert Group (HLEG) issued an influential set of guidelines for trustworthy AI, followed in 2024 by the adoption of the EU AI Act. While the EU policies and guidelines are expected to be aligned, they may differ in their scope, areas of emphasis, degrees of normativity, and priorities in relation to AI. To gain a broad understanding of AI governance from the EU perspective, we leverage qualitative thematic analysis approaches to uncover prevalent themes in key EU documents, including the AI Act and the HLEG Ethics Guidelines. We further employ quantitative topic modelling approaches, specifically through the use of the BERTopic model, to enhance the results and increase the document sample to include EU AI policy documents published post-2018. We present a novel perspective on EU policies, tracking the evolution of its approach to addressing AI governance.
AI Insights
  • The study blends thematic analysis with BERTopic, revealing how EU policy documents evolve post‑2018.
  • Sector‑specific AI, human autonomy, and regulatory enforcement emerge as top policy themes.
  • Human‑AI collaboration in document analysis is championed to boost accuracy and insight depth.
  • Researchers suggest refining topic models for greater interpretability and scalability.
  • Literature such as Post‑GPT Policy and When Politicians Talk AI illuminate AI framing in politics.
  • Papadopoulos et al. (2020) and Wang et al. (2025) provide foundational context for EU AI discourse.
  • Thematic Analysis and Topic Modeling are defined as pattern‑identification and statistical‑topic discovery, respectively.
👍 👎 ♄ Save
Universit de Montral
Abstract
Large Language Models (LLMs) are entering urban governance, yet their outputs are highly sensitive to prompts that carry value judgments. We propose Prompt Commons - a versioned, community-maintained repository of prompts with governance metadata, licensing, and moderation - to steer model behaviour toward pluralism. Using a Montreal dataset (443 human prompts; 3,317 after augmentation), we pilot three governance states (open, curated, veto-enabled). On a contested policy benchmark, a single-author prompt yields 24 percent neutral outcomes; commons-governed prompts raise neutrality to 48-52 percent while retaining decisiveness where appropriate. In a synthetic incident log, a veto-enabled regime reduces time-to-remediation for harmful outputs from 30.5 +/- 8.9 hours (open) to 5.6 +/- 1.5 hours. We outline licensing (CC BY/BY-SA for prompts with optional OpenRAIL-style restrictions for artefacts), auditable moderation, and safeguards against dominance capture. Prompt governance offers a practical lever for cities to align AI with local values and accountability.
Chat Designers
👍 👎 ♄ Save
Abstract
Despite being trained on vast amounts of data, most LLMs are unable to reliably generate well-designed UIs. Designer feedback is essential to improving performance on UI generation; however, we find that existing RLHF methods based on ratings or rankings are not well-aligned with designers' workflows and ignore the rich rationale used to critique and improve UI designs. In this paper, we investigate several approaches for designers to give feedback to UI generation models, using familiar interactions such as commenting, sketching and direct manipulation. We first perform a study with 21 designers where they gave feedback using these interactions, which resulted in ~1500 design annotations. We then use this data to finetune a series of LLMs to generate higher quality UIs. Finally, we evaluate these models with human judges, and we find that our designer-aligned approaches outperform models trained with traditional ranking feedback and all tested baselines, including GPT-5.
👍 👎 ♄ Save
Abstract
AI-driven chatbots are increasingly used to support community health workers (CHWs) in developing regions, yet little is known about how cultural framings in chatbot design shape trust in collectivist contexts where decisions are rarely made in isolation. This paper examines how CHWs in rural India responded to chatbots that delivered identical health content but varied in one specific cultural lever -- social norms. Through a mixed-methods study with 61 ASHAs who compared four normative framings -- neutral, descriptive, narrative identity, and injunctive authority -- we (1) analyze how framings influence preferences and trust, and (2) compare effects across low- and high-ambiguity scenarios. Results show that narrative framings were most preferred but encouraged uncritical overreliance, while authority framings were least preferred yet supported calibrated trust. We conclude with design recommendations for dynamic framing strategies that adapt to context and argue for calibrated trust -- following correct advice and resisting incorrect advice -- as a critical evaluation metric for safe, culturally-grounded AI.
AI for Compliance
👍 👎 ♄ Save
Copenhagen Business Sc hO
Paper visualization
Rate this image: 😍 👍 👎
Abstract
The cost and complexity of financial crime compliance (FCC) continue to rise, often without measurable improvements in effectiveness. While AI offers potential, most solutions remain opaque and poorly aligned with regulatory expectations. This paper presents the design and deployment of an agentic AI system for FCC in digitally native financial platforms. Developed through an Action Design Research (ADR) process with a fintech firm and regulatory stakeholders, the system automates onboarding, monitoring, investigation, and reporting, emphasizing explainability, traceability, and compliance-by-design. Using artifact-centric modeling, it assigns clearly bounded roles to autonomous agents and enables task-specific model routing and audit logging. The contribution includes a reference architecture, a real-world prototype, and insights into how Agentic AI can reconfigure FCC workflows under regulatory constraints. Our findings extend IS literature on AI-enabled compliance by demonstrating how automation, when embedded within accountable governance structures, can support transparency and institutional trust in high-stakes, regulated environments.
AI Insights
  • Prototype assigns bounded roles to agents via artifact‑centric modeling, enabling fine‑grained task routing and audit logging.
  • Reference architecture embeds explainability, traceability, and compliance‑by‑design into every agent’s decision loop.
  • ADR collaboration with fintech and regulators produced a governance framework balancing agentic autonomy and human oversight.
  • Early deployment shows agentic AI offloads routine checks, freeing officers for high‑value investigations.
  • Design shows FCC workflows can be reconfigured under regulatory constraints without sacrificing throughput or transparency.
  • Acharya et al. (2025) and Okpala et al. (2025) offer insights on model risk and autonomous goal‑setting in finance.
  • Further research must quantify long‑term impact on institutional trust and refine governance for evolving regulations.
👍 👎 ♄ Save
Microsoft Corporation
Abstract
Large language models (LLMs) are increasingly deployed in enterprise settings where they interact with multiple users and are trained or fine-tuned on sensitive internal data. While fine-tuning enhances performance by internalizing domain knowledge, it also introduces a critical security risk: leakage of confidential training data to unauthorized users. These risks are exacerbated when LLMs are combined with Retrieval-Augmented Generation (RAG) pipelines that dynamically fetch contextual documents at inference time. We demonstrate data exfiltration attacks on AI assistants where adversaries can exploit current fine-tuning and RAG architectures to leak sensitive information by leveraging the lack of access control enforcement. We show that existing defenses, including prompt sanitization, output filtering, system isolation, and training-level privacy mechanisms, are fundamentally probabilistic and fail to offer robust protection against such attacks. We take the position that only a deterministic and rigorous enforcement of fine-grained access control during both fine-tuning and RAG-based inference can reliably prevent the leakage of sensitive data to unauthorized recipients. We introduce a framework centered on the principle that any content used in training, retrieval, or generation by an LLM is explicitly authorized for \emph{all users involved in the interaction}. Our approach offers a simple yet powerful paradigm shift for building secure multi-user LLM systems that are grounded in classical access control but adapted to the unique challenges of modern AI workflows. Our solution has been deployed in Microsoft Copilot Tuning, a product offering that enables organizations to fine-tune models using their own enterprise-specific data.
Unsubscribe from these updates