465 research outputs found
Clustering objectives in wireless sensor networks: A survey and research direction analysis
Wireless Sensor Networks (WSNs) typically include thousands of resource-constrained sensors to monitor their surroundings, collect data, and transfer it to remote servers for further processing. Although WSNs are considered highly flexible ad-hoc networks, network management has been a fundamental challenge in these types of net- works given the deployment size and the associated quality concerns such as resource management, scalability, and reliability. Topology management is considered a viable technique to address these concerns. Clustering is the most well-known topology management method in WSNs, grouping nodes to manage them and/or executing various tasks in a distributed manner, such as resource management. Although clustering techniques are mainly known to improve energy consumption, there are various quality-driven objectives that can be realized through clustering. In this paper, we review comprehensively existing WSN clustering techniques, their objectives and the network properties supported by those techniques. After refining more than 500 clustering techniques, we extract about 215 of them as the most important ones, which we further review, catergorize and classify based on clustering objectives and also the network properties such as mobility and heterogeneity. In addition, statistics are provided based on the chosen metrics, providing highly useful insights into the design of clustering techniques in WSNs.publishedVersio
A Survey and Future Directions on Clustering: From WSNs to IoT and Modern Networking Paradigms
Many Internet of Things (IoT) networks are created as an overlay over traditional ad-hoc networks such as Zigbee. Moreover, IoT networks can resemble ad-hoc networks over networks that support device-to-device (D2D) communication, e.g., D2D-enabled cellular networks and WiFi-Direct. In these ad-hoc types of IoT networks, efficient topology management is a crucial requirement, and in particular in massive scale deployments. Traditionally, clustering has been recognized as a common approach for topology management in ad-hoc networks, e.g., in Wireless Sensor Networks (WSNs). Topology management in WSNs and ad-hoc IoT networks has many design commonalities as both need to transfer data to the destination hop by hop. Thus, WSN clustering techniques can presumably be applied for topology management in ad-hoc IoT networks. This requires a comprehensive study on WSN clustering techniques and investigating their applicability to ad-hoc IoT networks. In this article, we conduct a survey of this field based on the objectives for clustering, such as reducing energy consumption and load balancing, as well as the network properties relevant for efficient clustering in IoT, such as network heterogeneity and mobility. Beyond that, we investigate the advantages and challenges of clustering when IoT is integrated with modern computing and communication technologies such as Blockchain, Fog/Edge computing, and 5G. This survey provides useful insights into research on IoT clustering, allows broader understanding of its design challenges for IoT networks, and sheds light on its future applications in modern technologies integrated with IoT.acceptedVersio
Unsupervised Heart-rate Estimation in Wearables With Liquid States and A Probabilistic Readout
Heart-rate estimation is a fundamental feature of modern wearable devices. In
this paper we propose a machine intelligent approach for heart-rate estimation
from electrocardiogram (ECG) data collected using wearable devices. The novelty
of our approach lies in (1) encoding spatio-temporal properties of ECG signals
directly into spike train and using this to excite recurrently connected
spiking neurons in a Liquid State Machine computation model; (2) a novel
learning algorithm; and (3) an intelligently designed unsupervised readout
based on Fuzzy c-Means clustering of spike responses from a subset of neurons
(Liquid states), selected using particle swarm optimization. Our approach
differs from existing works by learning directly from ECG signals (allowing
personalization), without requiring costly data annotations. Additionally, our
approach can be easily implemented on state-of-the-art spiking-based
neuromorphic systems, offering high accuracy, yet significantly low energy
footprint, leading to an extended battery life of wearable devices. We
validated our approach with CARLsim, a GPU accelerated spiking neural network
simulator modeling Izhikevich spiking neurons with Spike Timing Dependent
Plasticity (STDP) and homeostatic scaling. A range of subjects are considered
from in-house clinical trials and public ECG databases. Results show high
accuracy and low energy footprint in heart-rate estimation across subjects with
and without cardiac irregularities, signifying the strong potential of this
approach to be integrated in future wearable devices.Comment: 51 pages, 12 figures, 6 tables, 95 references. Under submission at
Elsevier Neural Network
Enhanced Differential Crossover and Quantum Particle Swarm Optimization for IoT Applications
An optimized design with real-time and multiple realistic constraints in complex engineering systems is a crucial challenge for designers. In the non-uniform Internet of Things (IoT) node deployments, the approximation accuracy is directly affected by the parameters like node density and coverage. We propose a novel enhanced differential crossover quantum particle swarm optimization algorithm for solving nonlinear numerical problems. The algorithm is based on hybrid optimization using quantum PSO. Differential evolution operator is used to circumvent group moves in small ranges and falling into the local optima and improves global searchability. The cross operator is employed to promote information interchange among individuals in a group, and exceptional genes can be continued moderately, accompanying the evolutionary process's continuance and adding proactive and reactive features. The proposed algorithm's performance is verified as well as compared with the other algorithms through 30 classic benchmark functions in IEEE CEC2017, with a basic PSO algorithm and improved versions. The results show the smaller values of fitness function and computational efficiency for the benchmark functions of IEEE CEC2019. The proposed algorithm outperforms the existing optimization algorithms and different PSO versions, and has a high precision and faster convergence speed. The average location error is substantially reduced for the smart parking IoT application
Particle Swarm Optimization
Particle swarm optimization (PSO) is a population based stochastic optimization technique influenced by the social behavior of bird flocking or fish schooling.PSO shares many similarities with evolutionary computation techniques such as Genetic Algorithms (GA). The system is initialized with a population of random solutions and searches for optima by updating generations. However, unlike GA, PSO has no evolution operators such as crossover and mutation. In PSO, the potential solutions, called particles, fly through the problem space by following the current optimum particles. This book represents the contributions of the top researchers in this field and will serve as a valuable tool for professionals in this interdisciplinary field
Non-Gaussian Hybrid Transfer Functions: Memorizing Mine Survivability Calculations
Hybrid algorithms and models have received significant interest in recent years and are increasingly used to solve real-world problems. Different from existing methods in radial basis transfer function construction, this study proposes a novel nonlinear-weight hybrid algorithm involving the non-Gaussian type radial basis transfer functions. The speed and simplicity of the non-Gaussian type with the accuracy and simplicity of radial basis function are used to produce fast and accurate on-the-fly model for survivability of emergency mine rescue operations, that is, the survivability under all conditions is precalculated and used to train the neural network. The proposed hybrid uses genetic algorithm as a learning method which performs parameter optimization within an integrated analytic framework, to improve network efficiency. Finally, the network parameters including mean iteration, standard variation, standard deviation, convergent time, and optimized error are evaluated using the mean squared error. The results demonstrate that the hybrid model is able to reduce the computation complexity, increase the robustness and optimize its parameters. This novel hybrid model shows outstanding performance and is competitive over other existing models
A Survey on Reservoir Computing and its Interdisciplinary Applications Beyond Traditional Machine Learning
Reservoir computing (RC), first applied to temporal signal processing, is a
recurrent neural network in which neurons are randomly connected. Once
initialized, the connection strengths remain unchanged. Such a simple structure
turns RC into a non-linear dynamical system that maps low-dimensional inputs
into a high-dimensional space. The model's rich dynamics, linear separability,
and memory capacity then enable a simple linear readout to generate adequate
responses for various applications. RC spans areas far beyond machine learning,
since it has been shown that the complex dynamics can be realized in various
physical hardware implementations and biological devices. This yields greater
flexibility and shorter computation time. Moreover, the neuronal responses
triggered by the model's dynamics shed light on understanding brain mechanisms
that also exploit similar dynamical processes. While the literature on RC is
vast and fragmented, here we conduct a unified review of RC's recent
developments from machine learning to physics, biology, and neuroscience. We
first review the early RC models, and then survey the state-of-the-art models
and their applications. We further introduce studies on modeling the brain's
mechanisms by RC. Finally, we offer new perspectives on RC development,
including reservoir design, coding frameworks unification, physical RC
implementations, and interaction between RC, cognitive neuroscience and
evolution.Comment: 51 pages, 19 figures, IEEE Acces
An energy-efficient and deadline-aware workflow scheduling algorithm in the fog and cloud environment
peer reviewedThe Internet of Things (IoT) is constantly evolving. The variety of IoT applications has caused new demands to emerge on users’ part and competition between computing service providers. On the one hand, an IoT application may exhibit several important criteria, such as deadline and runtime simultaneously, and it is confronted with resource limitations and high energy consumption on the other hand. This has turned to adopting a computing environment and scheduling as a fundamental challenge. To resolve the issue, IoT applications are considered in this paper as a workflow composed of a series of interdependent tasks. The tasks in the same workflow (at the same level) are subject to priorities and deadlines for execution, making the problem far more complex and closer to the real world. In this paper, a hybrid Particle Swarm Optimization and Simulated Annealing algorithm (PSO–SA) is used for prioritizing tasks and improving fitness function. Our proposed method managed the task allocation and optimized energy consumption and makespan at the fog-cloud environment nodes. The simulation results indicated that the PSO–SA enhanced energy and makespan by 5% and 9% respectively on average compared with the baseline algorithm (IKH-EFT)
- …