14 research outputs found

    An Energy Aware Resource Utilization Framework to Control Traffic in Cloud Network and Overloads

    Get PDF
    Energy consumption in cloud computing occur due to the unreasonable way in which tasks are scheduled. So energy aware task scheduling is a major concern in cloud computing as energy consumption results into significant waste of energy, reduce the profit margin and also high carbon emissions which is not environmentally sustainable. Hence, energy efficient task scheduling solutions are required to attain variable resource management, live migration, minimal virtual machine design, overall system efficiency, reduction in operating costs, increasing system reliability, and prompting environmental protection with minimal performance overhead. This paper provides a comprehensive overview of the energy efficient techniques and approaches and proposes the energy aware resource utilization framework to control traffic in cloud networks and overloads

    STUDY ON CLOUD COMPUTING FOR EFFICIENT RESOURCE ALLOCATION AND SCHEDULING APPROACHES

    Get PDF
    The term "cloud storage" refers to a variety of online services that enable users to store and share digital media such as documents, data, photos, and videos. You may access these files from anywhere and on any device (laptop, mobile phone, tablet etc). With cloud computing, users are able to allocate their computing needs among a shared pool of powerful machines, increasing their access to resources like processing power, storage space, and software services. There is a growing population of internet users who rely on cloud-based resources. Large amounts of data are transferred from users to hosts and from hosts to users in the cloud environment, but as demand for cloud services grows, the associated cost and complexity for the cloud provider may become unsustainable. There may be times when two or more users make a request for the same item. Given these constraints, it's not easy to decide which host to use to gain access to the necessary resources and build a virtual machine (VM) in which to run the necessary applications in a way that maximises efficiency while minimising costs. Scheduling tasks in a cloud computing environment to maximise efficiency is one solution to this issue. An approach to job scheduling is provided by this project. In this research, we make an effort to suggest a host selection model based on shortest execution time to reduce overhead

    Joint Computation Offloading and Prioritized Scheduling in Mobile Edge Computing

    Get PDF
    With the rapid development of smart phones, enormous amounts of data are generated and usually require intensive and real-time computation. Nevertheless, quality of service (QoS) is hardly to be met due to the tension between resourcelimited (battery, CPU power) devices and computation-intensive applications. Mobileedge computing (MEC) emerging as a promising technique can be used to copy with stringent requirements from mobile applications. By offloading computationally intensive workloads to edge server and applying efficient task scheduling, energy cost of mobiles could be significantly reduced and therefore greatly improve QoS, e.g., latency. This paper proposes a joint computation offloading and prioritized task scheduling scheme in a multi-user mobile-edge computing system. We investigate an energy minimizing task offloading strategy in mobile devices and develop an effective priority-based task scheduling algorithm with edge server. The execution time, energy consumption, execution cost, and bonus score against both the task data sizes and latency requirement is adopted as the performance metric. Performance evaluation results show that, the proposed algorithm significantly reduce task completion time, edge server VM usage cost, and improve QoS in terms of bonus score. Moreover, dynamic prioritized task scheduling is also discussed herein, results show dynamic thresholds setting realizes the optimal task scheduling. We believe that this work is significant to the emerging mobile-edge computing paradigm, and can be applied to other Internet of Things (IoT)-Edge applications

    Framework for Virtualized Network Functions (VNFs) in Cloud of Things Based on Network Traffic Services

    Get PDF
    The cloud of things (CoT), which combines the Internet of Things (IoT) and cloud computing, may offer Virtualized Network Functions (VNFs) for IoT devices on a dynamic basis based on service-specific requirements. Although the provisioning of VNFs in CoT is described as an online decision-making problem, most widely used techniques primarily focus on defining the environment using simple models in order to discover the optimum solution. This leads to inefficient and coarse-grained provisioning since the Quality of Service (QoS) requirements for different types of CoT services are not considered, and important historical experience on how to provide for the best long-term benefits is disregarded. This paper suggests a methodology for providing VNFs intelligently in order to schedule adaptive CoT resources in line with the detection of traffic from diverse network services. The system makes decisions based on Deep Reinforcement Learning (DRL) based models that take into account the complexity of network configurations and traffic changes. To obtain stable performance in this model, a special surrogate objective function and a policy gradient DRL method known as Policy Optimisation using Kronecker-Factored Trust Region (POKTR) are utilised. The assertion that our strategy improves CoT QoS through real-time VNF provisioning is supported by experimental results. The POKTR algorithm-based DRL-based model maximises throughput while minimising network congestion compared to earlier DRL algorithms

    Hybridized Darts Game with Beluga Whale Optimization Strategy for Efficient Task Scheduling with Optimal Load Balancing in Cloud Computing

    Get PDF
    A cloud computing technology permits clients to use hardware and software technology virtually on a subscription basis. The task scheduling process is planned to effectively minimize implementation time and cost while simultaneously increasing resource utilization, and it is one of the most common problems in cloud computing systems. The Nondeterministic Polynomial (NP)-hard optimization problem occurs due to limitations like an insufficient make-span, excessive resource utilization, low implementation costs, and immediate response for scheduling. The task allocation is NP-hard because of the increase in the amount of combinations and computing resources. In this work, a hybrid heuristic optimization technique with load balancing is implemented for optimal task scheduling to increase the performance of service providers in the cloud infrastructure. Thus, the issues that occur in the scheduling process is greatly reduced. The load balancing problem is effectively solved with the help of the proposed task scheduling scheme. The allocation of tasks to the machines based on the workload is done with the help of the proposed Hybridized Darts Game-Based Beluga Whale Optimization Algorithm (HDG-BWOA). The objective functions like higher Cloud Data Center (CDC) resource consumption, increased task assurance ratio, minimized mean reaction time, and reduced energy utilization are considered while allocating the tasks to the virtual machines. This task scheduling approach ensures flexibility among virtual machines, preventing them from overloading or underloading. Also, using this technique, more tasks is efficiently completed within the deadline. The efficacy of the offered arrangement is ensured with the conventional heuristic-based task scheduling approaches in accordance with various evaluation measures

    A latency-aware max-min algorithm for resource allocation in cloud

    Get PDF
    Cloud computing is an emerging distributed computing paradigm. However, it requires certain initiatives that need to be tailored for the cloud environment such as the provision of an on-the-fly mechanism for providing resource availability based on the rapidly changing demands of the customers. Although, resource allocation is an important problem and has been widely studied, there are certain criteria that need to be considered. These criteria include meeting user’s quality of service (QoS) requirements. High QoS can be guaranteed only if resources are allocated in an optimal manner. This paper proposes a latency-aware max-min algorithm (LAM) for allocation of resources in cloud infrastructures. The proposed algorithm was designed to address challenges associated with resource allocation such as variations in user demands and on-demand access to unlimited resources. It is capable of allocating resources in a cloud-based environment with the target of enhancing infrastructure-level performance and maximization of profits with the optimum allocation of resources. A priority value is also associated with each user, which is calculated by analytic hierarchy process (AHP). The results validate the superiority for LAM due to better performance in comparison to other state-of-the-art algorithms with flexibility in resource allocation for fluctuating resource demand patterns

    Does firm size matter? Evidence on the impact of the green innovation strategy on corporate financial performance in the automotive sector

    Get PDF
    In the past few years, there has been increasing awareness regarding the significance of the Green Innovation Strategy (GIS) in the academic and practical fields. Hence, it becomes important to determine the correlation between the GIS and the Corporate Financial Performance (CFP). This study attempted to determine the dynamic correlation between the GIS and the CFP, with regards to the firm size. For this purpose, this study has collected data for 163 international automotive firms, from the CSRHub database, for the period ranging between 2011 and 2017. Furthermore, we also used the dynamic panel data system, i.e., the Generalised Method of Moment (GMM) method, for estimating this relationship. The empirical results indicated that the GIS positively affected the CFP. Interestingly, we also uncovered that the firm size moderated the negative correlation between the GIS and the CFP. The small-sized firms showed higher green innovation investments return than the larger-sized firms, which indicated that these smaller firms were more prone to seek variation and visibility, for accessing better resources. Furthermore, due to the extensive scrutiny of the stakeholders, these small firms could generate higher profits. The implications for managers and the theories in this regard are then discussed
    corecore