20,896 research outputs found
Efficient Deep Learning in Network Compression and Acceleration
While deep learning delivers state-of-the-art accuracy on many artificial intelligence tasks, it comes at the cost of high computational complexity due to large parameters. It is important to design or develop efficient methods to support deep learning toward enabling its scalable deployment, particularly for embedded devices such as mobile, Internet of things (IOT), and drones. In this chapter, I will present a comprehensive survey of several advanced approaches for efficient deep learning in network compression and acceleration. I will describe the central ideas behind each approach and explore the similarities and differences between different methods. Finally, I will present some future directions in this field
A Smart System for Future Generation based on the Internet of Things Employing Machine Learning, Deep Learning, and Artificial Intelligence : Comprehensive Survey
The Internet of Things (IoT) is a networked system including interconnected things, devices, and networks that utilize the internet for communication and data exchange. The entity engages in interactions with both its internal and external surroundings. The IoT is capable of seeing the surrounding environment and responding in a way that is appropriate and adaptive. The utilization of advanced technology in this context enhances the environment and thus enhances the overall well-being of humanity. The IoT facilitates inter-device communication, whether through physical or virtual means. The IoT facilitates the enhancement of environmental intelligence, enabling seamless connectivity across many devices at any given moment. The concepts centred on the IoT, such as augmented reality, high-resolution video streaming, autonomous vehicles, intelligent environments, and electronic healthcare, have become pervasive in contemporary society. These applications have requirements for faster data rates, larger bandwidths, enhanced capacities, decreased latencies, and increased throughputs. IoT and Machine learning (ML) are among the fields of research that have shown significant potential for advancement. ML and IoT are used to build intelligent systems. Those networks will modify the ways in which worldwide entities exchange information. This article gives a comprehensive survey of the upcoming 5G-IoT situation, as well as a study of IoT smart system applications and usages. In addition to covering the latest developments in ML and deep learning (DL) and their impact on 5G-IoT, this article describes a comprehensive study of these important enabling technologies and the developing use cases of 5G-IoT
Engineering the application of machine learning in an IDS based on IoT traffic flow
Internet of Things (IoT) devices are now widely used, enabling intelligent services that, in association with
new communication technologies like the 5G and broadband internet, boost smart-city environments. Despite
their limited resources, IoT devices collect and share large amounts of data and are connected to the internet,
becoming an attractive target for malicious actors. This work uses machine learning combined with an Intrusion Detection System (IDS) to detect possible attacks. Due to the limitations of IoT devices and low latency services, the IDS must have a specialized architecture. Furthermore, although machine learning-based solutions have high potential, there are still challenges related to training and generalization, which may impose constraints on the architecture. Our proposal is an IDS with a distributed architecture that relies on Fog computing to run specialized modules
and use deep neural networks to identify malicious traffic inside IoT data flows. We compare our IoT-Flow
IDS with three other architectures. We assess model generalization using test data from different datasets and
evaluate their performance in terms of Recall, Precision, and F1-Score. Results confirm the feasibility of flowbased anomaly detection and the importance of network traffic segmentation and specialized models in the AI-based IDS for IoT.info:eu-repo/semantics/publishedVersio
Recommended from our members
Efficient Learning in Heterogeneous Internet of Things Ecosystems
The Internet of Things (IoT) is a growing network of heterogeneous devices, combining various sensing and computing nodes at different scales, which creates a large volume of data. Many IoT applications use machine learning (ML) algorithms to analyze the data. The high computational complexity of ML workloads poses significant computational challenges to IoT computing platforms, which tend to be less-powerful and resource-constrained devices. Transmitting such large volumes of data to the cloud also have various issues such as scalability, security and privacy. In this dissertation, we propose efficient solutions to perform the ML tasks while decreasing power consumption and improving performance. We first leverage the heterogeneous and interconnected nature of the IoT systems, where IoT applications run on many different architectures (e.g., X86 server or ARM-based edge device) while communicating with each other. We present a cross-platform power and performance prediction technique for intelligent task allocation. The proposed technique estimates the time-variant energy consumption with only 7% error across completely different architectures, enabling the intelligent task allocation that saves the energy consumption of 16.5% for state-of-the-art ML workloads.We next show how to further advance the learning procedures towards real-time and online processing by distributing such learning tasks onto the hierarchy of IoT devices. Our solution leverages brain-inspired high-dimensional (HD) computing to derive a new class oflearning algorithms that can easily run on IoT devices, while providing high accuracy comparable to the state-of-the-arts. We present that the HD-based learning algorithms can cover various real-world problems from conventional classification to other cognitive tasks beyond classical MLs such as DNA pattern matching. We demonstrate that the HD-based learning can enable secure, collaborative learning by efficiently distributing a large volume of learning tasks into heterogeneous computing nodes. We have implemented the proposed learning solution on various platforms while offering superior computing efficiency. For example, our solution achieves 486Ă—and 7Ă— performance improvements for each of the training and inference phases on a low-power ARM processor, as compared to state-of-the-art deep learning
Big Data and the Internet of Things
Advances in sensing and computing capabilities are making it possible to
embed increasing computing power in small devices. This has enabled the sensing
devices not just to passively capture data at very high resolution but also to
take sophisticated actions in response. Combined with advances in
communication, this is resulting in an ecosystem of highly interconnected
devices referred to as the Internet of Things - IoT. In conjunction, the
advances in machine learning have allowed building models on this ever
increasing amounts of data. Consequently, devices all the way from heavy assets
such as aircraft engines to wearables such as health monitors can all now not
only generate massive amounts of data but can draw back on aggregate analytics
to "improve" their performance over time. Big data analytics has been
identified as a key enabler for the IoT. In this chapter, we discuss various
avenues of the IoT where big data analytics either is already making a
significant impact or is on the cusp of doing so. We also discuss social
implications and areas of concern.Comment: 33 pages. draft of upcoming book chapter in Japkowicz and Stefanowski
(eds.) Big Data Analysis: New algorithms for a new society, Springer Series
on Studies in Big Data, to appea
- …