Author(s):
Oliveira, Afonso ; Moreira, Gonçalo ; Costa, Diogo Emanuel Carvalho ; Pinto, Sandro ; Gomes, Tiago Manuel Ribeiro
Date: 2025
Persistent ID: https://hdl.handle.net/1822/95509
Origin: RepositóriUM - Universidade do Minho
Subject(s): Arm Cortex-A; Artificial intelligence; Machine learning; Multi-Core Interference; Virtualization
Description
Significant advances in Artificial Intelligence (AI) over the past decade have opened new aisles of exploration for industries such as automotive and industrial robotics, leading to the widespread adoption of AI. To meet the demands of modern applications, embedded platforms have evolved into highly heterogeneous designs, transitioning from simple microcontroller-based systems to intricate platforms with multiple processor units and hardware accelerators. Driven by Size, Weight, Power, and Cost (SWAP-C) constraints, both industry and academia have focused on consolidating systems with different criticality levels, referred to as mixed-criticality systems (MCS), onto a single hardware platform. However, the co-existence of heterogeneous designs can still raise several safety and security issues, which can be addressed through secure computer architectures and hypervisor technologies that provide spatial and temporal isolation features. This paper discusses the impact of MCS consolidation on AI-based applications. The setup uses the Bao hypervisor to deploy two virtual machines (VMs): one VM hosting TensorFlow Lite to run AI workloads on a Linux system, and the other supporting a bare-metal memory-intensive application. The study involves running Convolutional Neural Networks (CNNs) and Deep Neural Networks (DNNs) machine learning models (ML), based on two widely used classification datasets: MNIST and CIFAR-10, while assessing the impact of sharing platform resources with a memory-intensive application. Empirical results show that contention on last-level cache and system bus can significantly impact the inference process of an ML model up to 6.39x.