7 research outputs found

    Edge computing and machine learning on embedded systems

    Get PDF
    Capstone Project submitted to the Department of Engineering, Ashesi University in partial fulfillment of the requirements for the award of Bachelor of Science degree in Computer Engineering, May 2022Running Machine Learning (ML) in embedded systems has fueled the rush for edge computing, where machine learning runs in edge devices. This approach to ML yields many results, such as lower latencies and reduction of network traffic and bandwidth. This project set out to explore machine learning in embedded systems. The Edge Impulse Platform was used to collect data and to a neural network. The neural network created was able to distinguish between five classes of motion. The Neural Network created was tested on two microcontrollers and a desktop. Inferencing on the Arduino Nano BLE took 24ms, and on the desktop, it took 271.8 μs.Ashesi Universit

    Optimizing Deep Learning Inference on Embedded Systems Through Adaptive Model Selection

    Get PDF
    Deep neural networks (DNNs) are becoming a key enabling technique for many application domains. However, on-device inference on battery-powered, resource-constrained embedding systems is often infeasible due to prohibitively long inferencing time and resource requirements of many DNNs. Offloading computation into the cloud is often unacceptable due to privacy concerns, high latency, or the lack of connectivity. Although compression algorithms often succeed in reducing inferencing times, they come at the cost of reduced accuracy. This article presents a new, alternative approach to enable efficient execution of DNNs on embedded devices. Our approach dynamically determines which DNN to use for a given input by considering the desired accuracy and inference time. It employs machine learning to develop a low-cost predictive model to quickly select a pre-trained DNN to use for a given input and the optimization constraint. We achieve this first by offline training a predictive model and then using the learned model to select a DNN model to use for new, unseen inputs. We apply our approach to two representative DNN domains: image classification and machine translation. We evaluate our approach on a Jetson TX2 embedded deep learning platform and consider a range of influential DNN models including convolutional and recurrent neural networks. For image classification, we achieve a 1.8x reduction in inference time with a 7.52% improvement in accuracy over the most capable single DNN model. For machine translation, we achieve a 1.34x reduction in inference time over the most capable single model with little impact on the quality of translation
    corecore