36 research outputs found
Speeding up Deep Neural Networks on the Jetson TX1
In recent years, Deep Learning (DL) showed new top performances in almost all computer vision tasks that are important for automotive and robotic applications. In these applications both space and power are limited resources. Therefore, there is a need to apply DL approaches on a small and power ecient device, like the NVIDIA Jetson TX1 with a powerful GPU onboard. In this paper, we analyze the Jetson's suitability by benchmarking the run-time of DL operations in comparison to a high performance GPU. Exemplary, we port a topperforming DL-based person detector to this platform. We explain the steps necessary to signicantly speed up this approach on the device
DIANNE: a modular framework for designing, training and deploying deep neural networks on heterogeneous distributed infrastructure
Deep learning has shown tremendous results on various machine learning tasks, but the nature of the problems being tackled and the size of state-of-the-art deep neural networks often require training and deploying models on distributed infrastructure. DIANNE is a modular framework designed for dynamic (re)distribution of deep learning models and procedures. Besides providing elementary network building blocks as well as various training and evaluation routines, DIANNE focuses on dynamic deployment on heterogeneous distributed infrastructure, abstraction of Internet of Things (loT) sensors, integration with external systems and graphical user interfaces to build and deploy networks, while retaining the performance of similar deep learning frameworks. In this paper the DIANNE framework is proposed as an all-in-one solution for deep learning, enabling data and model parallelism though a modular design, offloading to local compute power, and the ability to abstract between simulation and real environment. (C) 2018 Elsevier Inc. All rights reserved
End-to-End Multiview Gesture Recognition for Autonomous Car Parking System
The use of hand gestures can be the most intuitive human-machine interaction medium.
The early approaches for hand gesture recognition used device-based methods. These
methods use mechanical or optical sensors attached to a glove or markers, which hinders
the natural human-machine communication. On the other hand, vision-based methods are
not restrictive and allow for a more spontaneous communication without the need of an
intermediary between human and machine. Therefore, vision gesture recognition has been
a popular area of research for the past thirty years.
Hand gesture recognition finds its application in many areas, particularly the automotive
industry where advanced automotive human-machine interface (HMI) designers are
using gesture recognition to improve driver and vehicle safety. However, technology advances
go beyond active/passive safety and into convenience and comfort. In this context,
one of America’s big three automakers has partnered with the Centre of Pattern Analysis
and Machine Intelligence (CPAMI) at the University of Waterloo to investigate expanding
their product segment through machine learning to provide an increased driver convenience
and comfort with the particular application of hand gesture recognition for autonomous
car parking.
In this thesis, we leverage the state-of-the-art deep learning and optimization techniques
to develop a vision-based multiview dynamic hand gesture recognizer for self-parking system.
We propose a 3DCNN gesture model architecture that we train on a publicly available
hand gesture database. We apply transfer learning methods to fine-tune the pre-trained
gesture model on a custom-made data, which significantly improved the proposed system
performance in real world environment. We adapt the architecture of the end-to-end solution
to expand the state of the art video classifier from a single image as input (fed by
monocular camera) to a multiview 360 feed, offered by a six cameras module. Finally, we
optimize the proposed solution to work on a limited resources embedded platform (Nvidia
Jetson TX2) that is used by automakers for vehicle-based features, without sacrificing the
accuracy robustness and real time functionality of the system
Design and Implementation of Sensing Methods on One-Tenth Scale of an Autonomous Race Car
Indiana University-Purdue University Indianapolis (IUPUI)Self-driving is simply the capacity of a vehicle to drive itself without human intervention. To accomplish this, the vehicle utilizes mechanical and electronic parts, sensors, actuators and an AI computer. The on-board PC runs advanced programming, which permits the vehicle to see and comprehend its current circumstance dependent on sensor input, limit itself in that climate and plan the ideal course from point A to point B. Independent driving is not an easy task, and to create self-sufficient driving arrangements is an exceptionally significant ability in the present programming designing field.
ROS is a robust and versatile communication middle ware (framework) tailored and widely used for robotics applications. This thesis work intends to show how ROS could be used to create independent driving programming by investigating self-governing driving issues, looking at existing arrangements and building up a model vehicle utilizing ROS.
The main focus of this thesis is to develop and implement a one-tenth scale of an autonomous RACECAR equipped with Jetson Nano board as the on-board computer, PCA9685 as PWM driver, sensors, and a ROS based software architecture.
Finally, by following the methods presented in this thesis, it is conceivable to build an autonomous RACECAR that runs on ROS.
By following the means portrayed in this theory of work, it is conceivable to build up a self-governing vehicle