27 research outputs found
Introduction to Community Service-Learning (SRCL 1000)
Introduction to Community Service-Learning is a general elective open to first to fourth year international and domestic students from a variety of disciplines across campus. Every fall and winter semester each student volunteers at one of 30 local not-for-profit organizations for a full semester. Students are required to complete 24 hours of service as part of their course work. In this poster session, 16 not-for-profit organizations will be represented by 27 SRCL 1000 students. They will demonstrate personal reflections on their service experiences, how their experiences connect to the course work and their organizations, and what they will take back to their own communities after the course is over.
Students representing the following Kamloops not-for-profit organizations:
Active Care Services: Nolan Fenrich
St. John Ambulance: Damilola Abiyo and Ryuki Furuta
Overlander Residential Care: Glory Amukamara
Ponderosa Lodge: Rahab Kariuki
The Kamloops Food Bank: Yu Cao, Surkamal Singh Jhand, Xiangzhong Kong and Ruotong Shi
The ReStore – Habitat for Humanity: Dion Maborekhe, Fengyi Yang and Haonan Deng
Kamloops Immigrant Services: Dipak Parmar
Maple Leaf School: Qian Wang and Mengyao Zhu
BC SPCA: Dawei Xu
TRU Sustainability Office: Akash Ghosh, Takaya Hirose, Jihoon Kim and Kosuke Masunaga
TRU Horticulture: Ols Buta
TRU The X Radio: Marie Gabriela Jimenez and MD Majharul Islam Sabuj
Beattie School of the Arts: Makoto Iida
Gemstone Care Center: Tirth Panchal
Chartwell Ridgepointe: Sakina Shikama
Sikh Temple: Gurpreet Pua
EdgeNAS: discovering efficient neural architectures for edge systems
Edge systems integrated with deep neural networks (DNNs) are deemed to pave the way for future artificial intelligence (AI). However, designing accurate and efficient DNNs for resource-limited edge systems is challenging as well as requires a huge amount of engineering efforts from human experts since the design space is highly complex and diverse. Also, previous works mostly focus on designing DNNs with less floating-point operations (FLOPs), but indirect FLOPs count does not necessarily reflect the complexity of DNNs. To tackle these, we, in this paper, propose a novel neural architecture search (NAS) approach, namely EdgeNAS, to automatically discover efficient DNNs for less capable edge systems. To this end, we propose an end-to-end learning-based latency estimator, which is able to directly approximate the architecture latency on edge systems while incurring negligible computational overheads. Further, we effectively incorporate the latency estimator into EdgeNAS with a uniform sampling strategy, which guides the architecture search towards an edge-efficient direction. Moreover, a search space regularization approach is introduced to balance the trade-off between efficiency and accuracy. We evaluate EdgeNAS on the edge platform, Nvidia Jetson Xavier, with three popular datasets. Experimental results demonstrate the superiority of EdgeNAS over state-of-the-art approaches in terms of latency, accuracy, number of parameters, and the search cost.Ministry of Education (MOE)Nanyang Technological UniversitySubmitted/Accepted versionThis work is partially supported by the Ministry of Education, Singapore, under its Academic Research Fund Tier 2 (MOE2019-T2-1-071) and Tier 1 (MOE2019-T1-001-072), and partially supported by Nanyang Technological University, Singapore, under its NAP (M4082282) and SUG (M4082087)
Bringing AI to edge : from deep learning's perspective
Edge computing and artificial intelligence (AI), especially deep learning for
nowadays, are gradually intersecting to build a novel system, called edge
intelligence. However, the development of edge intelligence systems encounters
some challenges, and one of these challenges is the \textit{computational gap}
between computation-intensive deep learning algorithms and less-capable edge
systems. Due to the computational gap, many edge intelligence systems cannot
meet the expected performance requirements. To bridge the gap, a plethora of
deep learning techniques and optimization methods are proposed in the past
years: light-weight deep learning models, network compression, and efficient
neural architecture search. Although some reviews or surveys have partially
covered this large body of literature, we lack a systematic and comprehensive
review to discuss all aspects of these deep learning techniques which are
critical for edge intelligence implementation. As various and diverse methods
which are applicable to edge systems are proposed intensively, a holistic
review would enable edge computing engineers and community to know the
state-of-the-art deep learning techniques which are instrumental for edge
intelligence and to facilitate the development of edge intelligence systems.
This paper surveys the representative and latest deep learning techniques that
are useful for edge intelligence systems, including hand-crafted models, model
compression, hardware-aware neural architecture search and adaptive deep
learning models. Finally, based on observations and simple experiments we
conducted, we discuss some future directions.Submitted/Accepted versio
HACScale : hardware-aware compound scaling for resource-efficient DNNs
Model scaling is an effective way to improve the accuracy of deep neural networks (DNNs) by increasing the model capacity. However, existing approaches seldom consider the underlying hardware, causing inefficient utilization of hardware resources and consequently high inference latency. In this paper, we propose HACScale, a hardware-aware model scaling strategy to fully exploit hardware resources for higher accuracy. In HACScale, different dimensions of DNNs are jointly scaled with consideration of their contributions to hardware utilization and accuracy. To improve the efficiency of width scaling, we introduce importance-aware width scaling in HACScale, which computes the importance of each layer to the accuracy and scales each layer accordingly to optimize the trade-off between accuracy and model parameters. Experiments show that HACScale improves the hardware utilization by 1.92× on ImageNet, as a result, it achieves 2.41% accuracy improvement with a negligible latency increase of 0.6%. On CIFAR-10, HACScale improves the accuracy by 2.23% with only 6.5% latency growth.Nanyang Technological UniversityNational Research Foundation (NRF)Submitted/Accepted versionThis study is supported under the RIE2020 Industry Alignment Fund – Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contribution from the industry partner, HP Inc., through the HP-NTU Digital Manufacturing Corporate Lab. This work is also partially supported by NTU NAP M4082282 and SUG M4082087, Singapore
You only search once: on lightweight differentiable architecture search for resource-constrained embedded platforms
Benefiting from the search efficiency, differentiable neural architecture search (NAS) has evolved as the most dominant alternative to automatically design competitive deep neural networks (DNNs). We note that DNNs must be executed under strictly hard performance constraints in real-world scenarios, for example, the runtime latency on autonomous vehicles. However, to obtain the architecture that meets the given performance constraint, previous hardware-aware differentiable NAS methods have to repeat a plethora of search runs to manually tune the hyper-parameters by trial and error, and thus the total design cost increases proportionally. To resolve this, we introduce a lightweight hardware-aware differentiable NAS framework dubbed LightNAS, striving to find the required architecture that satisfies various performance constraints through a one-time search (i.e., \underline{\textit{you only search once}}). Extensive experiments are conducted to show the superiority of LightNAS over previous state-of-the-art methods.Ministry of Education (MOE)Nanyang Technological UniversitySubmitted/Accepted versionThis work is partially supported by the Ministry of Education, Singapore, under its Academic Research Fund Tier 2 (MOE2019-T2-1-071) and Tier 1 (MOE2019-T1-001-072), and partially supported by Nanyang Technological University, Singapore, under its NAP (M4082282) and SUG (M4082087)
SurgeNAS: a comprehensive surgery on hardware-aware differentiable neural architecture search
Differentiable neural architecture search (NAS) is an emerging paradigm to automate the design of top-performing convolutional neural networks (CNNs). However, previous differentiable NAS methods suffer from several crucial weaknesses, such as inaccurate gradient estimation, high memory consumption, search fairness, etc. More importantly, previous differentiable NAS works are mostly hardware-agnostic since they only search for CNNs in terms of accuracy, ignoring other critical performance metrics like latency. In this work, we introduce a novel hardware-aware differentiable NAS framework, namely SurgeNAS, in which we leverage the one-level optimization to avoid inaccuracy in gradient estimation. To this end, we propose an effective identity mapping regularization to alleviate the over-selecting issue. Besides, to mitigate the memory bottleneck, we propose an ordered differentiable sampling approach, which significantly reduces the search memory consumption to the single-path level, thereby allowing to directly search on target tasks instead of small proxy tasks. Meanwhile, it guarantees the strict search fairness. Moreover, we introduce a graph neural networks (GNNs) based predictor to approximate the on-device latency, which is further integrated into SurgeNAS to enable the latency-aware architecture search. Finally, we analyze the resource underutilization issue, in which we propose to scale up the searched SurgeNets within \textit{Comfort Zone} to balance the computation and memory access, which brings considerable accuracy improvement without deteriorating the execution efficiency. Extensive experiments are conducted on ImageNet with diverse hardware platforms, which clearly show the effectiveness of SurgeNAS in terms of accuracy, latency, and search efficiency.Ministry of Education (MOE)Nanyang Technological UniversitySubmitted/Accepted versionThis work is partially supported by the Ministry of Education, Singapore, under its Academic Research Fund Tier 2 (MOE2019-T2-1-071) and Tier 1 (MOE2019-T1-001-072), and partially supported by Nanyang Technological University, Singapore, under its NAP (M4082282) and SUG (M4082087)
Work-in-progress: what to expect of early training statistics? An investigation on hardware-aware neural architecture search
Neural architecture search (NAS) is an emerging paradigm to automate the design of top-performing deep neural networks (DNNs). Specifically, the increasing success of NAS is attributed to the reliable performance estimation of different architectures. Despite significant progress to date, previous relevant methods suffer from prohibitive computational overheads. To avoid this, we propose an effective yet computationally efficient proxy, namely Trained Batchwise Estimation (TBE), to reliably estimate the performance of different architectures using the early batchwise training statistics. We then integrate TBE into the hardware-aware NAS scenario to search for hardware-efficient architecture solutions. Experimental results clearly show the superiority of TBE over previous relevant state-of-the-art approaches.Nanyang Technological UniversitySubmitted/Accepted versionThis work is supported by Nanyang Technological University, Singapore, under its NAP (M4082282) and SUG (M4082087)
Designing efficient DNNs via hardware-aware neural architecture search and beyond
Hardware systems integrated with deep neural networks (DNNs) are deemed to pave the way for future artificial intelligence (AI). However, manually designing efficient DNNs involves non-trivial computation resources since significant trial-and-errors are required to finalize the network configuration. To this end, we, in this paper, introduce a novel hardware-aware neural architecture search (NAS) framework, namely GoldenNAS, to automate the design of efficient DNNs. To begin with, we present a novel technique, called dynamic channel scaling, to enable the channel-level search since the number of channels has non-negligible impacts on both accuracy and efficiency. Besides, we introduce an efficient progressive space shrinking method to raise the awareness of the search space towards target hardware and alleviate the search overheads as well. Moreover, we propose an effective hardware performance modeling method to approximate the runtime latency of DNNs upon target hardware, which is further integrated into GoldenNAS to avoid the tedious on-device measurements. Then, we employ the evolutionary algorithm (EA) to search for the optimal operator/channel configurations of DNNs, denoted as GoldenNets. Finally, to enable the depthwise adaptiveness of GoldenNets under dynamic environments, we propose the adaptive batch normalization (ABN) technique, followed by the self-knowledge distillation (SKD) approach to improve the accuracy of adaptive sub-networks. We conduct extensive experiments directly on ImageNet, which clearly demonstrate the advantages of GoldenNAS over existing state-of-the-art approaches.Ministry of Education (MOE)Nanyang Technological UniversitySubmitted/Accepted versionThis work is partially supported by the Ministry of Education, Singapore, under its Academic Research Fund Tier 2 (MOE2019-T2-1-071) and Tier 1 (MOE2019-T1-001-072), and partially supported by Nanyang Technological University, Singapore, under its NAP (M4082282) and SUG (M4082087)