3,090 research outputs found

    Cloud computing resource scheduling and a survey of its evolutionary approaches

    Get PDF
    A disruptive technology fundamentally transforming the way that computing services are delivered, cloud computing offers information and communication technology users a new dimension of convenience of resources, as services via the Internet. Because cloud provides a finite pool of virtualized on-demand resources, optimally scheduling them has become an essential and rewarding topic, where a trend of using Evolutionary Computation (EC) algorithms is emerging rapidly. Through analyzing the cloud computing architecture, this survey first presents taxonomy at two levels of scheduling cloud resources. It then paints a landscape of the scheduling problem and solutions. According to the taxonomy, a comprehensive survey of state-of-the-art approaches is presented systematically. Looking forward, challenges and potential future research directions are investigated and invited, including real-time scheduling, adaptive dynamic scheduling, large-scale scheduling, multiobjective scheduling, and distributed and parallel scheduling. At the dawn of Industry 4.0, cloud computing scheduling for cyber-physical integration with the presence of big data is also discussed. Research in this area is only in its infancy, but with the rapid fusion of information and data technology, more exciting and agenda-setting topics are likely to emerge on the horizon

    A Review on Energy Consumption Optimization Techniques in IoT Based Smart Building Environments

    Get PDF
    In recent years, due to the unnecessary wastage of electrical energy in residential buildings, the requirement of energy optimization and user comfort has gained vital importance. In the literature, various techniques have been proposed addressing the energy optimization problem. The goal of each technique was to maintain a balance between user comfort and energy requirements such that the user can achieve the desired comfort level with the minimum amount of energy consumption. Researchers have addressed the issue with the help of different optimization algorithms and variations in the parameters to reduce energy consumption. To the best of our knowledge, this problem is not solved yet due to its challenging nature. The gap in the literature is due to the advancements in the technology and drawbacks of the optimization algorithms and the introduction of different new optimization algorithms. Further, many newly proposed optimization algorithms which have produced better accuracy on the benchmark instances but have not been applied yet for the optimization of energy consumption in smart homes. In this paper, we have carried out a detailed literature review of the techniques used for the optimization of energy consumption and scheduling in smart homes. The detailed discussion has been carried out on different factors contributing towards thermal comfort, visual comfort, and air quality comfort. We have also reviewed the fog and edge computing techniques used in smart homes

    A WOA-based optimization approach for task scheduling in cloud Computing systems

    Get PDF
    Task scheduling in cloud computing can directly affect the resource usage and operational cost of a system. To improve the efficiency of task executions in a cloud, various metaheuristic algorithms, as well as their variations, have been proposed to optimize the scheduling. In this work, for the first time, we apply the latest metaheuristics WOA (the whale optimization algorithm) for cloud task scheduling with a multiobjective optimization model, aiming at improving the performance of a cloud system with given computing resources. On that basis, we propose an advanced approach called IWC (Improved WOA for Cloud task scheduling) to further improve the optimal solution search capability of the WOA-based method. We present the detailed implementation of IWC and our simulation-based experiments show that the proposed IWC has better convergence speed and accuracy in searching for the optimal task scheduling plans, compared to the current metaheuristic algorithms. Moreover, it can also achieve better performance on system resource utilization, in the presence of both small and large-scale tasks

    Hybrid scheduling algorithms in cloud computing: a review

    Get PDF
    Cloud computing is one of the emerging fields in computer science due to its several advancements like on-demand processing, resource sharing, and pay per use. There are several cloud computing issues like security, quality of service (QoS) management, data center energy consumption, and scaling. Scheduling is one of the several challenging problems in cloud computing, where several tasks need to be assigned to resources to optimize the quality of service parameters. Scheduling is a well-known NP-hard problem in cloud computing. This will require a suitable scheduling algorithm. Several heuristics and meta-heuristics algorithms were proposed for scheduling the user's task to the resources available in cloud computing in an optimal way. Hybrid scheduling algorithms have become popular in cloud computing. In this paper, we reviewed the hybrid algorithms, which are the combinations of two or more algorithms, used for scheduling in cloud computing. The basic idea behind the hybridization of the algorithm is to take useful features of the used algorithms. This article also classifies the hybrid algorithms and analyzes their objectives, quality of service (QoS) parameters, and future directions for hybrid scheduling algorithms

    Dynamic Offloading Technique for Latency-Sensitive Internet of Things Applications using Fog Computing

    Get PDF
    Internet of Things (IoT) has evolved as a novel paradigm that provides com-putation power to different entities connected to it. IoT offers services to multiple sectors such as home automation, industrial automation, traffic management, healthcare sector, agriculture industry etc. IoT generally relies on cloud data centers for extended analytics, processing and storage support. The cloud offers highly scalable and robust platform for IoT applications. But latency sensitive IoT applications suffer delay issues as the cloud lies in remote location. Edge/fog computing was introduced to overcome the issues faced by delay-sensitive IoT applications. These platforms lie close to the IoT network, reducing the delay and response time. The fog nodes are usually distributed in nature. The data has to be properly offloaded to available fog nodes using efficient strategies to gain benefit from the integration. Differ-ent offloading schemes are available in the literature to overcome this prob-lem This paper proposes a novel offloading approach by combining two effi-cient metaheuristic algorithms, Honey Badger Algorithm (HBA) and Fla-mingo Search Algorithm (FSA) termed as HB-FS algorithm. The HB-FS is executed in an iterative manner optimizing the objective function in each it-eration. The performance evaluation of the proposed approach is done with different existing metaheuristic algorithms and the evaluations show that the proposed work outperforms the existing algorithms in terms of latency, response time and execution time. The methodology also offers better degree of imbalance with proper load balancing under different conditions

    Solving Task Scheduling Problem in Cloud Computing Environment Using Orthogonal Taguchi-Cat Algorithm

    Get PDF
    In cloud computing datacenter, task execution delay is no longer accidental. In recent times, a number of artificial intelligence scheduling techniques are proposed and applied to reduce task execution delay. In this study, we proposed an algorithm called Orthogonal Taguchi Based-Cat Swarm Optimization (OTB-CSO) to minimize total task execution time. In our proposed algorithm Taguchi Orthogonal approach was incorporated at CSO tracing mode for best task mapping on VMs with minimum execution time. The proposed algorithm was implemented on CloudSim tool and evaluated based on makespan metric. Experimental results showed for 20VMs used, proposed OTB-CSO was able to minimize makespan of total tasks scheduled across VMs with 42.86%, 34.57% and 2.58% improvement over Minimum and Maximum Job First (Min-Max), Particle Swarm Optimization with Linear Descending Inertia Weight (PSO-LDIW) and Hybrid Particle Swarm Optimization with Simulated Annealing (HPSO-SA) algorithms. Results obtained showed OTB-CSO is effective to optimize task scheduling and improve overall cloud computing performance with better system utilization

    Task Scheduling Based on Grey Wolf Optimizer Algorithm for Smart Meter Embedded Operating System

    Get PDF
    In recent years, with the rapid development of electric power informatization, smart meters are gradually developing towards intelligent IOT. Smart meters can not only measure user status, but also interconnect and communicate with cell phones, smart homes and other cloud devices, and these core functions are completed by the smart meter embedded operating system. Due to the dynamic heterogeneity of the user program side and the system processing side of the embedded system, resource allocation and task scheduling is a challenging problem for embedded operating systems of smart meters. Smart meters need to achieve fast response and shortest completion time for user program side requests, and also need to take into account the load balancing of each processing node to ensure the reliability of smart meter embedded systems. In this paper, based on the advanced Grey Wolf Optimizer, we study the scheduling principle of the service program nodes in the smart meter operating system, and analyze the problems of the traditional scheduling algorithm to find the optimal solution. Compared with traditional algorithms and classical swarm intelligence algorithms, the algorithm proposed in this paper avoids the dilemma of local optimization, can quickly allocate operating system tasks, effectively shorten the time consumption of task scheduling, ensure the real-time performance of multi task scheduling, and achieve the system tuning balance. Finally, the effectiveness of the algorithm is verified by simulation experiments
    corecore