Page 99 - AI Standards for Global Impact: From Governance to Action
P. 99

AI Standards for Global Impact: From Governance to Action                      Part 2: Thematic AI


































                   Figure 43: Evolution of NPU reference chips of IntelliFusion

                   13�4�7  Day 0 support for novel models enabling instantaneous deployment
                           of edge AI

                   AI compiler technology has become mission-critical for deploying AI models at scale on edge
                   devices. The usual trend is for solutions to build on a legacy software stack that interprets AI
                   models but does not allow for compilation. This limits the application of the technology for AI
                   models, such as language models, as they are not compatible with the existing technology stack.

                   AI compilation optimizes model execution at different levels of abstraction – known as
                   "intermediate representations" - that represent specific features of the execution of an AI
                   workload. The compiler translates the AI model through various intermediate representations
                   to a low level that is close to the target hardware. Using AI compilation instead of model
                   interpretation allows users to adapt to the constant stream of new AI models.

                   Roofline’s flexible SDK enables models to be imported from any AI framework such as
                   TensorFlow, PyTorch, or ONNX. It enables deployment across diverse hardware, covering
                   CPUs, MPUs, MCUs, GPUs, and dedicated AI hardware accelerators - all with just a single line
                   of Python. Roofline’s flexible SDK that enables day-0 support for PyTorch models found on
                   Hugging Face, which solves the challenge of the inability to support emerging new AI models
                   and layer types, due to specific software features not compatible across systems.

                   13�4�8  Defining the boundaries of AI from fundamental limitations to resource
                           constraints

                   CSEM mentioned the challenges and solutions in deploying AI at the edge, focusing on the
                   limitations posed by neural scaling laws and the memory wall. The need for hardware-aware
                   AI design, efficient feature selection, and adversarial training to overcome resource constraints
                   while maintaining performance was highlighted. The proposed strategies enable scalable,
                   private, and energy-efficient AI solutions tailored for edge computing environments.






                                                            87
   94   95   96   97   98   99   100   101   102   103   104