Abstract
As Artificial Intelligence (AI), particularly Large Language Models (LLMs),
becomes increasingly embedded in education systems worldwide, ensuring their
ethical, legal, and contextually appropriate deployment has become a critical
policy concern. This paper offers a comparative analysis of AI-related
regulatory and ethical frameworks across key global regions, including the
European Union, United Kingdom, United States, China, and Gulf Cooperation
Council (GCC) countries. It maps how core trustworthiness principles, such as
transparency, fairness, accountability, data privacy, and human oversight are
embedded in regional legislation and AI governance structures. Special emphasis
is placed on the evolving landscape in the GCC, where countries are rapidly
advancing national AI strategies and education-sector innovation. To support
this development, the paper introduces a Compliance-Centered AI Governance
Framework tailored to the GCC context. This includes a tiered typology and
institutional checklist designed to help regulators, educators, and developers
align AI adoption with both international norms and local values. By
synthesizing global best practices with region-specific challenges, the paper
contributes practical guidance for building legally sound, ethically grounded,
and culturally sensitive AI systems in education. These insights are intended
to inform future regulatory harmonization and promote responsible AI
integration across diverse educational environments.
Abstract
This paper introduces a novel self-consciousness defense mechanism for Large
Language Models (LLMs) to combat prompt injection attacks. Unlike traditional
approaches that rely on external classifiers, our method leverages the LLM's
inherent reasoning capabilities to perform self-protection. We propose a
framework that incorporates Meta-Cognitive and Arbitration Modules, enabling
LLMs to evaluate and regulate their own outputs autonomously. Our approach is
evaluated on seven state-of-the-art LLMs using two datasets: AdvBench and
Prompt-Injection-Mixed-Techniques-2024. Experiment results demonstrate
significant improvements in defense success rates across models and datasets,
with some achieving perfect and near-perfect defense in Enhanced Mode. We also
analyze the trade-off between defense success rate improvement and
computational overhead. This self-consciousness method offers a lightweight,
cost-effective solution for enhancing LLM ethics, particularly beneficial for
GenAI use cases across various platforms.