1,103 research outputs found

    Increasing the Efficiency of Rule-Based Expert Systems Applied on Heterogeneous Data Sources

    Get PDF
    Nowadays, the proliferation of heterogeneous data sources provided by different research and innovation projects and initiatives is proliferating more and more and presents huge opportunities. These developments create an increase in the number of different data sources, which could be involved in the process of decisionmaking for a specific purpose, but this huge heterogeneity makes this task difficult. Traditionally, the expert systems try to integrate all information into a main database, but, sometimes, this information is not easily available, or its integration with other databases is very problematic. In this case, it is essential to establish procedures that make a metadata distributed integration for them. This process provides a “mapping” of available information, but it is only at logic level. Thus, on a physical level, the data is still distributed into several resources. In this sense, this chapter proposes a distributed rule engine extension (DREE) based on edge computing that makes an integration of metadata provided by different heterogeneous data sources, applying then a mathematical decomposition over the antecedent of rules. The use of the proposed rule engine increases the efficiency and the capability of rule-based expert systems, providing the possibility of applying these rules over distributed and heterogeneous data sources, increasing the size of data sets that could be involved in the decision-making process

    Scientific Workflow Scheduling for Cloud Computing Environments

    Get PDF
    The scheduling of workflow applications consists of assigning their tasks to computer resources to fulfill a final goal such as minimizing total workflow execution time. For this reason, workflow scheduling plays a crucial role in efficiently running experiments. Workflows often have many discrete tasks and the number of different task distributions possible and consequent time required to evaluate each configuration quickly becomes prohibitively large. A proper solution to the scheduling problem requires the analysis of tasks and resources, production of an accurate environment model and, most importantly, the adaptation of optimization techniques. This study is a major step toward solving the scheduling problem by not only addressing these issues but also optimizing the runtime and reducing monetary cost, two of the most important variables. This study proposes three scheduling algorithms capable of answering key issues to solve the scheduling problem. Firstly, it unveils BaRRS, a scheduling solution that exploits parallelism and optimizes runtime and monetary cost. Secondly, it proposes GA-ETI, a scheduler capable of returning the number of resources that a given workflow requires for execution. Finally, it describes PSO-DS, a scheduler based on particle swarm optimization to efficiently schedule large workflows. To test the algorithms, five well-known benchmarks are selected that represent different scientific applications. The experiments found the novel algorithms solutions substantially improve efficiency, reducing makespan by 11% to 78%. The proposed frameworks open a path for building a complete system that encompasses the capabilities of a workflow manager, scheduler, and a cloud resource broker in order to offer scientists a single tool to run computationally intensive applications

    Anticipatory models of load balancing in cloud computing

    Full text link
    University of Technology Sydney. Faculty of Engineering and Information Technology.Cloud Computing is a recent arrival to the world of IT infrastructure. The concept allows companies to maximise utilisation of their potentials and consequently boost their performance. One of the main benefits of Cloud Computing is the significant increase in efficiency of executing business plans. Additionally, Cloud Computing provides large-scale applications with powerful computing power across global locations. Yet Cloud users are able to share their data easily by using replication methodologies. Cloud Computing structure has been developed based on a multi-tenancy concept. Therefore, availability and efficiency of the resources are important factors in the Cloud architecture. However, as the numbers of users are increasing rapidly, the load will have a significant impact on performance and operation of the Cloud systems. Accordingly, optimised load balancing algorithms that can manage the Cloud load in a time- and cost-efficient manner are required. Much research in recent years has been dedicated to optimising load balancing in Cloud Computing. This optimisation is demonstrated through a balanced network of interacting resources. The goal of this network is to minimise the wait time and maximise utilisation of the throughput. This thesis provides a set of solutions which mitigate the problem of load balancing in the Cloud. The dissertation investigates a novel class of heuristic scheduling algorithms that improves load balancing in workflow scheduling applications. Furthermore, it proposes a new anticipatory replication methodology with the objective of improving data availability to enhance the load balancing between the Cloud sites. In summary, this research innovation implicates the design of optimised load balancing algorithms that consider the magnitude and direction of the load in workflow applications. Furthermore, by architecting the anticipatory replication algorithm, it minimises the numbers of the replicas and enhances the effective network usage in Cloud-based systems

    DATA REPLICATION IN DISTRIBUTED SYSTEMS USING OLYMPIAD OPTIMIZATION ALGORITHM

    Get PDF
    Achieving timely access to data objects is a major challenge in big distributed systems like the Internet of Things (IoT) platforms. Therefore, minimizing the data read and write operation time in distributed systems has elevated to a higher priority for system designers and mechanical engineers. Replication and the appropriate placement of the replicas on the most accessible data servers is a problem of NP-complete optimization. The key objectives of the current study are minimizing the data access time, reducing the quantity of replicas, and improving the data availability. The current paper employs the Olympiad Optimization Algorithm (OOA) as a novel population-based and discrete heuristic algorithm to solve the replica placement problem which is also applicable to other fields such as mechanical and computer engineering design problems. This discrete algorithm was inspired by the learning process of student groups who are preparing for the Olympiad exams. The proposed algorithm, which is divide-and-conquer-based with local and global search strategies, was used in solving the replica placement problem in a standard simulated distributed system. The 'European Union Database' (EUData) was employed to evaluate the proposed algorithm, which contains 28 nodes as servers and a network architecture in the format of a complete graph. It was revealed that the proposed technique reduces data access time by 39% with around six replicas, which is vastly superior to the earlier methods. Moreover, the standard deviation of the results of the algorithm's different executions is approximately 0.0062, which is lower than the other techniques' standard deviation within the same experiments

    A Time-driven Data Placement Strategy for a Scientific Workflow Combining Edge Computing and Cloud Computing

    Full text link
    Compared to traditional distributed computing environments such as grids, cloud computing provides a more cost-effective way to deploy scientific workflows. Each task of a scientific workflow requires several large datasets that are located in different datacenters from the cloud computing environment, resulting in serious data transmission delays. Edge computing reduces the data transmission delays and supports the fixed storing manner for scientific workflow private datasets, but there is a bottleneck in its storage capacity. It is a challenge to combine the advantages of both edge computing and cloud computing to rationalize the data placement of scientific workflow, and optimize the data transmission time across different datacenters. Traditional data placement strategies maintain load balancing with a given number of datacenters, which results in a large data transmission time. In this study, a self-adaptive discrete particle swarm optimization algorithm with genetic algorithm operators (GA-DPSO) was proposed to optimize the data transmission time when placing data for a scientific workflow. This approach considered the characteristics of data placement combining edge computing and cloud computing. In addition, it considered the impact factors impacting transmission delay, such as the band-width between datacenters, the number of edge datacenters, and the storage capacity of edge datacenters. The crossover operator and mutation operator of the genetic algorithm were adopted to avoid the premature convergence of the traditional particle swarm optimization algorithm, which enhanced the diversity of population evolution and effectively reduced the data transmission time. The experimental results show that the data placement strategy based on GA-DPSO can effectively reduce the data transmission time during workflow execution combining edge computing and cloud computing

    Optimization for Integration of Plug-in Hybrid Electric Vehicles into Distribution Grid

    Get PDF
    Plug-in hybrid electric vehicles (PHEVs) feature combined electric and gasoline powertrains with internal combustion engine and electric motors powered by battery packs. The battery packs of PHEVs are mostly charged during off-peaks hours at lower prices and meanwhile absorb the excess power from the grid. Similarly, the power stored in the batteries can also flow back to the electric grid in response to ease the peak load demands. With the increasing penetration and integration of PHEVs, the reliability of PHEVs is essential to overall power system reliability since the charging mechanisms of PHEVs can influence the reliability of power system. Furthermore, due to the direct integration of PHEVs into the residential distribution network, the PHEVs can work as backup batteries for power systems in case of power outage. Therefore, the reliability analysis of power systems and the ancillary services provided by PHEVs are also proposed in this thesis study. For the driving pattern of each PHEV, there are three basic elements modeled, which are the departure time, the arrival time and the daily mileage, all represented by probability density functions. Based on these basic concepts, the methodology for modeling the load demand of PHEVs is introduced. In the proposed system, both the Differential Evolution and the Particle Swarm Optimization are proposed to optimize the control strategies for power systems with integration of PHEVs. Aside from using the minimum cost as a figure of merit when designing and determining the best PHEV charging mechanism, the reliability improvement brought to the power systems by PHEVs and the extra earnings obtained by performing frequency regulation services are also quantified and taken into account. Although the reliability of power systems with PHEV penetrations has been well-studied, the adoption of the Differential Evolution algorithm for minimizing the cost of overall system is not exercised, not to mention a thorough comparative study with other common optimization algorithms. To sum up, the Differential Evolution can not only achieve multiple goals by improving the power quality, reducing the peak load, providing regulation services and minimizing the total virtual cost in this system, it can also offer better results compared with the Particle Swarm Optimization in terms of minimizing the cost

    Optimized Load Balancing based Task Scheduling in Cloud Environment

    Get PDF
    The fundamental issue of Task scheduling is one important factor to load balance between the virtual machines in a Cloud Computing network. However, the optimal broadcast methods which have been proposed so far focus only on cluster or grid environment. In this paper, task scheduling strategy based on load balancing Quantum Particles Swarm algorithm (BLQPSO) was proposed. The fitness function based minimizing the makespan and data transmission cost. In addition, the salient feature of this algorithm is to optimize node available throughput dynamically using MatLab10A software. Furthermore, the performance of proposed algorithm had been compared with existing PSO and shows their effectiveness in balancing the load

    A Kriging Method for Modeling Cycle Time-Throughput Profiles in Manufacturing

    Get PDF
    In semiconductor manufacturing, the steady-state behavior of a wafer fab system can be characterized by its cycle time-throughput profiles. These profiles quantify the relationship between the cycle time of a product and the system throughput and product mix. The objective of this work is to efficiently generate such cycle time-throughput profiles in manufacturing which can further assist decision makings in production planning.;In this research, a metamodeling approach based on Stochastic Kriging model with Qualitative factors (SKQ) has been adopted to quantify the target relationship of interest. Furthermore, a sequential experimental design procedure is developed to improve the efficiency of simulation experiments. For the initial design, a Sequential Conditional Maximin algorithm is utilized. Regarding the follow-up designs, batches of design points are determined using a Particle Swarm Optimization algorithm.;The procedure is applied to a Jackson network, as well as a scale-down wafer fab system. In both examples, the prediction performance of the SKQ model is promising. It is also shown that the SKQ model provides narrower confidence intervals compared to the Stochastic Kriging model (SK) by pooling the information of the qualitative variables
    corecore