Page 11 - The Annual AI Governance Report 2025 Steering the Future of AI
P. 11

The Annual AI Governance Report 2025: Steering the Future of AI



                  2025 also saw increased attention to frontier model development outside the traditional
                  hubs. The release of DeepSeek-V2, China’s highest-performing open-weight model, marks
                  a significant milestone in the global development of frontier-scale AI systems. DeepSeek’s
                  capabilities prompted reflection on the minimum levels of compute, engineering expertise,
                  and energy resources needed to build frontier systems. At the same time, models like Mistral
                  and Falcon reinforced that open-source leadership is emerging across a range of geographies.

                  AI systems consume energy across two main stages: (a) Training, which requires high energy
                  input at a single location over a limited time. (b) Inference, which involves ongoing energy use
                  as models are deployed and used by millions daily. Current typical chatbot interactions have a
                  negligible per-use energy footprint (~0.2–0.34 Wh per query): assuming 100 interactions with
                  a chatbot (corresponding to 10 chats with 10 back-and-forth messages) every single day of a
                  year, corresponds to roughly 10 kWh for the annual chatbot use, or the equivalent of a 10 km car
                  drive, or 5 hot showers of 5 minutes, or 2 hot baths. However, future scenarios – where billions
                  of users each employ multiple AI agents – could dramatically increase global energy demand,
                  potentially reaching on the order of TWh/day or more. This would represent a significant share of
                  the world's total electricity consumption (~80 TWh/day, which itself is about 20% of the world’s
                  total energy consumption). The energy infrastructure for AI is rapidly expanding: Globally, AI
                  data centers may need 68 GW by 2027 and 327 GW by 2030, comparable to major U.S. states
                  like California.
                  The global  AI standards landscape is shaped by international Standards Development
                  Organizations – ITU, ISO, IEC – and practitioner-led bodies like IEEE. Together, they are building
                  a layered system of standards, spanning technical, managerial, and socio-technical domains.
                  The AI and Multimedia Authenticity Standards (AMAS) initiative, launched under the World
                  Standards Cooperation (the partnership of ITU, ISO and IEC), is developing tools and guidance
                  to support transparency, accountability, and human rights in AI. Its initial deliverables – mapping
                  the current landscape and identifying standardization gaps – are being published at the AI
                  for Good Summit 2025, with more to follow. A new AI Standards Exchange Database will be
                  introduced at the same summit, consolidating standards from ITU, ISO, IEC, IEEE, and IETF.
                  ITU, ISO, and IEC also launched the International AI Standards Summit series in 2024, with the
                  second edition planned for Seoul in December 2025, aligning with the goals of the UN Global
                  Digital Compact. Regionally, CEN-CENELEC’s JTC 21 in Europe is advancing sector-specific
                  AI standards, notably in healthcare and mobility, supporting implementation of the EU AI Act
                  and promoting cross-border interoperability through trustworthiness metrics and regulatory
                  alignment. Nevertheless, adoption of formal AI standards remains the exception rather than the
                  rule as firms are faced with a patchwork of faster – but unofficial – industry frameworks, resulting
                  in inconsistency and forum-shopping risks.

                  The Global AI Summit Series (UK 2023, Seoul 2024, France 2025, India 2026) has evolved
                  from ad hoc gatherings to structured international cooperation. Nevertheless, AI governance
                  efforts remain fragmented and politically uneven. While a number of countries have launched
                  AI safety institutes and initiated risk assessment frameworks, global coordination remains limited
                  to a handful of Track 2 dialogues – informal, non-government channels who meet to explore
                  ideas and issues that their governments (the “Track 1” diplomats) may be unable or unwilling
                  to discuss in public – and regional initiatives. Track 2 diplomacy between the US and China has
                  grown more substantive. The Ditchley Statement (2023) and the Beijing Dialogue (2024) reflect
                  expert consensus around compute thresholds, model registration, and dual-use risk evaluation.






                                                           2
   6   7   8   9   10   11   12   13   14   15   16