A global review of AI regulation covering the EU, US, China, and international efforts toward safe, ethical AI governance

Global AI Regulation: EU, US, China & the Race for Governance

The AI Governance Crossroads

Artificial Intelligence (AI) has rapidly evolved from niche research to a transformative force reshaping economies, societies, and political landscapes. As generative models, autonomous systems, and advanced analytics proliferate, governments worldwide face the formidable challenge of regulating AI in ways that protect human rights, ensure safety, and foster innovation.

Evidence of this regulatory acceleration is clear. The European Union (EU) has implemented the groundbreaking AI Act, making it the first region with comprehensive AI legislation. In the United States, federal and state-level initiatives like Executive Order 14179 and the California AI policy report are redefining national priorities. China emphasizes national standards ahead of its November 2025 deadlines. Meanwhile, global alliances—including a treaty under the Council of Europe and high-profile summits—are attempting to build consensus.

This article examines the current regulatory landscape, compares international approaches, analyzes the stakes for global cooperation and competition, and explores what comes next for technology, business, and society.


1. The EU AI Act: A New Global Standard

Timeline and Scope

The EU Artificial Intelligence Act (Regulation (EU) 2024/1689) entered into force on August 1, 2024, with full applicability staged across 6–36 months depending on risk class cimplifi.com+3en.wikipedia.org+3digital-strategy.ec.europa.eu+3. Built on a risk-based framework—unacceptable, high, limited, and minimal—plus a separate category for general-purpose AI systems, it is designed to balance safety and innovation across sectors en.wikipedia.org.

Key Provisions

  • Transparency and Disclosure: Providers of general-purpose AI (e.g., ChatGPT-like systems) must share model capabilities, training data patterns, and compliance reports.
  • High-Risk AI Compliance: Systems used in critical areas (healthcare, transport, employment) require rigorous assessment, logging, and third-party audits.
  • Discoverability: Users must be informed when content is AI-generated.
  • Penalties: Fines can reach €35 million or 7% of global annual revenue—whichever is higher newsfromthestates.com+4investopedia.com+4kslaw.com+4en.wikipedia.org+6cimplifi.com+6thetimes.co.uk+6.

Global Impact

The act’s extraterritorial design mandates compliance from any entity reaching EU users, which is prompting worldwide businesses to adopt its standards as global best practices . Some experts compare its rollout to GDPR: a spike in audits, followed by normalization—though smaller firms may find compliance financially burdensome newsfromthestates.com+4investopedia.com+4theguardian.com+4.

Internal Resources


2. US AI Strategy: From Federal Order to State Action

Executive Order 14179

On January 23, 2025, then-President Trump signed Executive Order 14179, titled “Removing Barriers to American Leadership in Artificial Intelligence” cimplifi.com+2en.wikipedia.org+2kslaw.com+2. This reversed previous Biden-era directives aimed at safe, ethical AI development, pivoting instead to:

  • Deregulation over caution: Prioritizing minimal barriers and avoiding ideological oversight.
  • Agency realignment: Encouraging federal agencies to repeal policies seen as restrictive.
  • Global leadership: Making the US a dominant global AI innovator.

Federal vs. State Policy

While Washington emphasizes innovation, states are stepping in:

  • California commissioned a 53-page policy report (June 17, 2025) warning of “irreversible harms,” particularly biosecurity risks. Led by figures such as Fei‑Fei Li, it highlights the need for transparency, whistleblower rights, and verification systems washingtonpost.com+1cimplifi.com+1time.com.
  • Colorado’s AI Act, effective 2026, requires developers to prevent bias and notify consumers—serving as a model for emerging laws in multiple states cimplifi.com+1whitecase.com+1.
  • Additional laws in Maryland, New Hampshire, and Tennessee introduce liability and content-usage caps holisticai.com+2cimplifi.com+2fairly.ai+2.

Industry Reaction

Big Tech likes federal preemption. Microsoft, Amazon, Meta, and Google support bipartisan bills banning divergent state regulations—but critics say this risks impeding adaptive policy. Cybersecurity experts warn delays in state oversight may intensify national vulnerabilities theguardian.com+3ft.com+3myjournalcourier.com+3.


3. China’s Standard-Driven Path

Timeline and Structure

China enacted its national AI Measures in August 2023, with sweeping new standards—especially for generative AI—due by November 1, 2025 whitecase.com+1investopedia.com+1.

Core Objectives

These regulations focus on:

  • User data protection
  • Dataset security
  • Guardrails against misinformation and AI-enabled harms

Other key regulations include the 2023 “Deep Synthesis” and “Recommendation Algorithms” rules cimplifi.comwhitecase.com.

Strategic Signals

China’s regime is centralized, relatively quick-moving, and coordinate-oriented—mirroring its broader industrial policy. This contrasts with Western decentralization and risk-tiering.


4. Global Treaties and Collaborative Frameworks

Council of Europe AI Convention

In May 2024, the “Framework Convention on Artificial Intelligence and Human Rights, Democracy, and the Rule of Law” was adopted, opening for signature in September 2024 en.wikipedia.org+1en.wikipedia.org+1. Over 50 countries—including the US, UK, and EU member states—have signed, committing to uphold democratic and ethical AI principles.

International Summits

  • AI Action Summit, Paris, February 10–11, 2025, co-hosted by Macron and Modi; featured 1,000+ participants from over 100 nations en.wikipedia.org.
  • Deliverables included €200 bn InvestAI package, EU AI gigafactories, and a Statement on Inclusive and Sustainable AI endorsed by 58 countries en.wikipedia.org.
  • Notably, the US and UK declined to sign that statement—reflecting regulatory divergence en.wikipedia.org.

5. Public Sentiment and Expert Opinion

Public Trust & Anxiety

Ipsos Mori surveys (23,000 respondents, 30 countries) reveal that English-speaking nations (UK, US, Canada, Australia) are most anxious about AI—with many fearing job displacement and data misuse. In contrast, large EU and Southeast Asian countries show more acceptance. Crucially, trust in government regulation plays a central role in public confidence theguardian.com.

Thought Leadership

  • Gaia Marcus (Ada Lovelace Institute): 72 % of UK respondents want stronger regulation; warns of power concentration and mental health risk ft.com.
  • Pope Leo XIV: Emphasizes that AI threatens human dignity and labor; urging global ethical oversight wsj.com+1nypost.com+1.

6. Comparative Analysis: Models & Tensions

RegionApproachStrengthsChallenges
EURisk-based regulation with binding rules, transparency, and penaltiesHigh trust, global standard-setterHigh compliance cost, innovation friction
US (Federal)Deregulation-driven, agency-guided policy, focus on scale & competitivenessPromotes rapid innovation, centralized economyFragmented state laws, public skepticism
US (States)Detailed, domain-specific legislationAdaptive, targeted enforcementInconsistent patchwork across states
ChinaNational standards, centralized enforcementCoherent, fast implementationLimited transparency, human rights concerns
Council of Europe & SummitsTreaty-based human-rights framework, public-private collaborationsEnhanced democratic safeguardsNon-binding, US/UK hesitancy
Scroll to Top