Page 47 - AI Standards for Global Impact: From Governance to Action
P. 47

AI Standards for Global Impact: From Governance to Action



                   step in the process of creating an AI system, from data-set collection through to deployment
                   (and beyond). A key aspect of this process involves developing methodologies for constructing
                   and training AI systems that are safe or ensuring their alignment with specific contexts and
                   scenarios.                                                                                       Part 2: Thematic AI

                   Berkeley University presented the challenges posed by AI security risks and the research
                   on trustworthiness and risk assessment of different Large Language Models (LLMs), which
                   identified different ways today’s LLMs can be attacked. It was noted that, with the enhanced
                   capabilities brought about by Agentic AI systems, there is a need for collaboration among
                   scientific communities to share lessons learned and best practices.

                   The Netherlands Organisation for Applied Scientific Research (TNO) has been conducting
                   research on AI security, working closely with various Dutch government bodies and NATO
                   partners. As AI becomes increasingly integrated into critical systems, ensuring its security is not
                   just a technical challenge but imperative for national safety. Despite the growing importance
                   of this field, they observed a significant gap: the lack of openly accessible tools to assess and
                   enhance the security of AI systems. TNO is working on the development of an AI Security
                   Assessment Framework tailored to the needs of the Dutch government. This framework aims
                   to provide a structured approach to evaluating the security and trustworthiness of AI systems.
                   Their findings underscore a clear call to action: the AI community must collaborate to create
                   and share more open, practical tools that support the secure deployment of AI.

                   Atlas Computing presented the research concept of Flexible Hardware Enabled Guarantees
                   (FlexHEG). As AI advances, so does the potential for catastrophic risks resulting from accidents,
                   misuse or loss of control over dangerous capabilities. For example, severe misuse in domains
                   such as disinformation or cyber attacks seems plausible within the next few years. As such,
                   governance of AI technology — whether by national governments, industry self-governance,
                   intergovernmental agreements, or all three — is a crucial capacity for humanity to develop, and
                   quickly.

                   Hardware-enabled governance has emerged as a promising pathway to help mitigate such
                   risks and increase product trustworthiness by implementing safety measures directly onto
                   high-performance computing chips. These hardware-enabled mechanisms could be added
                   to powerful AI accelerators used in datacentres.

                   However, it is not yet clear which compliance rules will be most appropriate in the future.
                   Therefore, these hardware-enabled governance mechanisms could be considered for the flexible
                   updating of compliance rules through a multilateral, cryptographically secure input channel,
                   without needing to retool the hardware.Concepts like FlexHEG could enable multilateral control
                   over AI technology, thus making it possible for a range of stakeholders to agree on a variety of
                   potential rules, from safety rules to robust benefit-sharing agreements. Mutually agreed-upon
                   rules could be set and updated through a multilateral and cryptographically secure mechanism
                   to guarantee that only agreed-upon rules are applied.


                   6�2  International collaboration on AI testing

                   The session discussed the gaps in testing AI systems, evaluation of trust in AI systems, and
                   opportunities for international collaboration to ensure the effective design, development, and
                   deployment of AI systems that integrate considerations for AI trust. There was general agreement
                   among participants on the need for better collaboration internationally on methodologies and




                                                            35
   42   43   44   45   46   47   48   49   50   51   52