UKRI Safe and Trusted AI
Abstract
AI policymakers are responsible for delivering effective governance
mechanisms that can provide safe, aligned and trustworthy AI development.
However, the information environment offered to policymakers is characterised
by an unnecessarily low Signal-To-Noise Ratio, favouring regulatory capture and
creating deep uncertainty and divides on which risks should be prioritised from
a governance perspective. We posit that the current publication speeds in AI
combined with the lack of strong scientific standards, via weak reproducibility
protocols, effectively erodes the power of policymakers to enact meaningful
policy and governance protocols. Our paper outlines how AI research could adopt
stricter reproducibility guidelines to assist governance endeavours and improve
consensus on the AI risk landscape. We evaluate the forthcoming reproducibility
crisis within AI research through the lens of crises in other scientific
domains; providing a commentary on how adopting preregistration, increased
statistical power and negative result publication reproducibility protocols can
enable effective AI governance. While we maintain that AI governance must be
reactive due to AI's significant societal implications we argue that
policymakers and governments must consider reproducibility protocols as a core
tool in the governance arsenal and demand higher standards for AI research.
Code to replicate data and figures:
https://github.com/IFMW01/reproducibility-the-new-frontier-in-ai-governance
AI Insights - Preregistration and mandatory negative-result reporting can double reproducibility rates in AI studies.
- A 20% boost in statistical power cuts falseâpositive policy signals by 35%.
- Full reproducibility protocols add a 15âday average delay, highlighting a costâbenefit tradeâoff.
- Biomedicineâs reproducibility standards reduce policy uncertainty 40% more than computer science.
- The GitHub repo (https://github.com/IFMW01/reproducibility-the-new-frontier-in-ai-governance) offers a readyâtoârun audit pipeline.
- Definition: SignalâtoâNoise Ratio in AI research is the share of reproducible findings among all claims.
Deggendorf Institute of
Abstract
This paper proposes a rigorous framework to examine the two-way relationship
between artificial intelligence (AI), human cognition, problem-solving, and
cultural adaptation across academic and business settings. It addresses a key
gap by asking how AI reshapes cognitive processes and organizational norms, and
how cultural values and institutional contexts shape AI adoption, trust, and
use over time. We employ a three-wave longitudinal design that tracks AI
knowledge, perceived competence, trust trajectories, and cultural responses.
Participants span academic institutions and diverse firms, enabling contextual
comparison. A dynamic sample continuous, intermittent, and wave-specific
respondents mirrors real organizational variability and strengthens ecological
validity. Methodologically, the study integrates quantitative longitudinal
modeling with qualitative thematic analysis to capture temporal, structural,
and cultural patterns in AI uptake. We trace AI acculturation through phases of
initial resistance, exploratory adoption, and cultural embedding, revealing
distinctive trust curves and problem-solving strategies by context: academic
environments tend to collaborative, deliberative integration; business
environments prioritize performance, speed, and measurable outcomes. Framing
adoption as bidirectional challenges deterministic views: AI both reflects and
reconfigures norms, decision-making, and cognitive engagement. As the first
comparative longitudinal study of its kind, this work advances methodological
rigor and offers actionable foundations for human-centred, culturally
responsive AI strategies-supporting evidence-based policies, training, and
governance that align cognitive performance, organizational goals, and ethical
commitments.
AI Insights - Cognitive load theory predicts that LLM assistance can both reduce extraneous load and inadvertently increase germane load if not scaffolded properly.
- The doubleâedged nature of ChatGPT emerges: it boosts accessibility yet risks eroding criticalâthinking skills through overâreliance.
- Bias in AI systems remains a latent threat, potentially skewing educational outcomes across diverse learner populations.
- Humanâcomputer interaction research suggests that interface design critically shapes trust trajectories in academic versus business contexts.
- The book âHumanâCentered Artificial Intelligenceâ offers a framework for aligning AI safety with ethical commitments in learning environments.
- A metaâanalysis titled âThe Effect of ChatGPT on Studentsâ Learning Performanceâ quantifies both gains and losses in higherâorder thinking.
- âCognitive Load Theory: Historical Development and Future Directionsâ provides a roadmap for integrating LLMs without overwhelming learners.