2,849 research outputs found

    Generalized Spring Tensor Algorithms: with Workflow Scheduling Applications in Cloud Computing

    Full text link
    In Cloud Computing, designing an efficient workflow scheduling algorithm is considered as a main goal. Load balancing is one of the most sophisticated methodologies, which can optimize workflow scheduling by distributing the load evenly among available resources. A well-designed load balancing algorithm has significant impact on performance and output in Cloud Computing. Therefore, designing robust load balancing techniques to manage the networks' load has always been a priority. Researchers have proposed and examined different load balancing methods; there is, however, a large knowledge gap in adopting an efficient load balancing algorithm in the Cloud system. This paper describes how a generalized spring tensor, an evolutionary algorithm with mathematical apparatus, can be utilized for a more efficient and effective load management in Cloud Computing. Considering the fluctuation and magnitude of the load, a novel application of workflow scheduling is investigated in the context of various mathematical patterns. The preliminary results of the research show that defining the dependency ratio between workflow tasks in Cloud Computing, results in better resource management, maximized performance and minimized response time while dealing with customer's requests

    Toward Customizable Multi-tenant SaaS Applications

    Get PDF
    abstract: Nowadays, Computing is so pervasive that it has become indeed the 5th utility (after water, electricity, gas, telephony) as Leonard Kleinrock once envisioned. Evolved from utility computing, cloud computing has emerged as a computing infrastructure that enables rapid delivery of computing resources as a utility in a dynamically scalable, virtualized manner. However, the current industrial cloud computing implementations promote segregation among different cloud providers, which leads to user lockdown because of prohibitive migration cost. On the other hand, Service-Orented Computing (SOC) including service-oriented architecture (SOA) and Web Services (WS) promote standardization and openness with its enabling standards and communication protocols. This thesis proposes a Service-Oriented Cloud Computing Architecture by combining the best attributes of the two paradigms to promote an open, interoperable environment for cloud computing development. Mutil-tenancy SaaS applicantions built on top of SOCCA have more flexibility and are not locked down by a certain platform. Tenants residing on a multi-tenant application appear to be the sole owner of the application and not aware of the existence of others. A multi-tenant SaaS application accommodates each tenant’s unique requirements by allowing tenant-level customization. A complex SaaS application that supports hundreds, even thousands of tenants could have hundreds of customization points with each of them providing multiple options, and this could result in a huge number of ways to customize the application. This dissertation also proposes innovative customization approaches, which studies similar tenants’ customization choices and each individual users behaviors, then provides guided semi-automated customization process for the future tenants. A semi-automated customization process could enable tenants to quickly implement the customization that best suits their business needs.Dissertation/ThesisDoctoral Dissertation Computer Science 201

    Towards An Efficient Cloud Computing System: Data Management, Resource Allocation and Job Scheduling

    Get PDF
    Cloud computing is an emerging technology in distributed computing, and it has proved to be an effective infrastructure to provide services to users. Cloud is developing day by day and faces many challenges. One of challenges is to build cost-effective data management system that can ensure high data availability while maintaining consistency. Another challenge in cloud is efficient resource allocation which ensures high resource utilization and high SLO availability. Scheduling, referring to a set of policies to control the order of the work to be performed by a computer system, for high throughput is another challenge. In this dissertation, we study how to manage data and improve data availability while reducing cost (i.e., consistency maintenance cost and storage cost); how to efficiently manage the resource for processing jobs and increase the resource utilization with high SLO availability; how to design an efficient scheduling algorithm which provides high throughput, low overhead while satisfying the demands on completion time of jobs. Replication is a common approach to enhance data availability in cloud storage systems. Previously proposed replication schemes cannot effectively handle both correlated and non-correlated machine failures while increasing the data availability with the limited resource. The schemes for correlated machine failures must create a constant number of replicas for each data object, which neglects diverse data popularities and cannot utilize the resource to maximize the expected data availability. Also, the previous schemes neglect the consistency maintenance cost and the storage cost caused by replication. It is critical for cloud providers to maximize data availability hence minimize SLA (Service Level Agreement) violations while minimize cost caused by replication in order to maximize the revenue. In this dissertation, we build a nonlinear programming model to maximize data availability in both types of failures and minimize the cost caused by replication. Based on the model\u27s solution for the replication degree of each data object, we propose a low-cost multi-failure resilient replication scheme (MRR). MRR can effectively handle both correlated and non-correlated machine failures, considers data popularities to enhance data availability, and also tries to minimize consistency maintenance and storage cost. In current cloud, providers still need to reserve resources to allow users to scale on demand. The capacity offered by cloud offerings is in the form of pre-defined virtual machine (VM) configurations. This incurs resource wastage and results in low resource utilization when the users actually consume much less resource than the VM capacity. Existing works either reallocate the unused resources with no Service Level Objectives (SLOs) for availability\footnote{Availability refers to the probability of an allocated resource being remain operational and accessible during the validity of the contract~\cite{CarvalhoCirne14}.} or consider SLOs to reallocate the unused resources for long-running service jobs. This approach increases the allocated resource whenever it detects that SLO is violated in order to achieve SLO in the long term, neglecting the frequent fluctuations of jobs\u27 resource requirements in real-time application especially for short-term jobs that require fast responses and decision making for resource allocation. Thus, this approach cannot fully utilize the resources to process data because they cannot quickly adjust the resource allocation strategy dealing with the fluctuations of jobs\u27 resource requirements. What\u27s more, the previous opportunistic based resource allocation approach aims at providing long-term availability SLOs with good QoS for long-running jobs, which ensures that the jobs can be finished within weeks or months by providing slighted degraded resources with moderate availability guarantees, but it ignores deadline constraints in defining Quality of Service (QoS) for short-lived jobs requiring online responses in real-time application, thus it cannot truly guarantee the QoS and long-term availability SLOs. To overcome the drawbacks of previous works, we adequately consider the fluctuations of unused resource caused by bursts of jobs\u27 resource demands, and present a cooperative opportunistic resource provisioning (CORP) scheme to dynamically allocate the resource to jobs. CORP leverages complementarity of jobs\u27 requirements on different resource types and utilizes the job packing to reduce the resource wastage and increase the resource utilization. An increasing number of large-scale data analytics frameworks move towards larger degrees of parallelism aiming at high throughput. Scheduling that assigns tasks to workers and preemption that suspends low-priority tasks and runs high-priority tasks are two important functions in such frameworks. There are many existing works on scheduling and preemption in literature to provide high throughput. However, previous works do not substantially consider dependency in increasing throughput in scheduling or preemption. Considering dependency is crucial to increase the overall throughput. Besides, extensive task evictions for preemption increase context switches, which may decrease the throughput. To address the above problems, we propose an efficient scheduling system Dependency-aware Scheduling and Preemption (DSP) to achieve high throughput in scheduling and preemption. First, we build a mathematical model to minimize the makespan with the consideration of task dependency, and derive the target workers for tasks which can minimize the makespan; second, we utilize task dependency information to determine tasks\u27 priorities for preemption; finally, we present a probabilistic based preemption to reduce the numerous preemptions, while satisfying the demands on completion time of jobs. We conduct trace driven simulations on a real-cluster and real-world experiments on Amazon S3/EC2 to demonstrate the efficiency and effectiveness of our proposed system in comparison with other systems. The experimental results show the superior performance of our proposed system. In the future, we will further consider data update frequency to reduce consistency maintenance cost, and we will consider the effects of node joining and node leaving. Also we will consider energy consumption of machines and design an optimal replication scheme to improve data availability while saving power. For resource allocation, we will consider using the greedy approach for deep learning to reduce the computation overhead caused by the deep neural network. Also, we will additionally consider the heterogeneity of jobs (i.e., short jobs and long jobs), and use a hybrid resource allocation strategy to provide SLO availability customization for different job types while increasing the resource utilization. For scheduling, we will aim to handle scheduling tasks with partial dependency, worker failures in scheduling and make our DSP fully distributed to increase its scalability. Finally, we plan to use different workloads and real-world experiment to fully test the performance of our methods and make our preliminary system design more mature

    Balancing Demand and Supply in Complex Manufacturing Operations: Tactical-Level Planning Processes

    Get PDF
    By balancing medium-term demand and supply, tactical planning enables manufacturing firms to realize strategic, long-term business objectives. However, such balancing in engineer-to-order (ETO) and configured-to-order (CTO) operations, due to the constant pressure of substantial complexity (e.g., volatility, uncertainty, and ambiguity), induces frequent swings between over- and undercapacity and thus considerable financial losses. Manufacturers respond to such complexity by using planning processes that address the business’s needs and risks at various medium-term horizons, ranging from 3 months to 3 years. Because the importance of decision-making increases exponentially as the horizon shrinks, understanding the interaction between complexity and demand-supply balancing requires extending findings reported in the literature on operations and supply chain planning and control. Therefore, this thesis addresses complexity’s impact on planning medium-term demand-supply balancing on three horizons: the strategic– tactical interface, the tactical level, and the tactical–operational interface.To explore complexity’s impact on demand–supply balancing in planning processes, the thesis draws on five studies, the first two of which addressed customer order fulfillment in ETO operations. Whereas Study I, an in-depth single-case study, examined relevant tactical-level decisions, planning activities, and their interface with the complexity affecting demand–supply balancing at the strategic–tactical interface, Study II, an in-depth multiple-case study, revealed the cross-functional mechanisms of integration affecting those decisions and activities and their impact on complexity. Next, Study III, also an in-depth multiple-case study, investigated areas of uncertainty, information-processing needs (IPNs), and information-processing mechanisms (IPMs) within sales and operations planning in ETO operations. By contrast, Studies IV and V addressed material delivery schedules (MDSs) in CTO operations; whereas Study IV, another in-depth multiple-case study, identified complexity interactions causing MDS instability at the tactical–operational interface, Study V, a case study, quantitatively explained how several factors affect MDS instability.Compiling six papers based on those five studies, the thesis contributes to theory and practice by extending knowledge about relationships between complexity and demand–supply balancing within a medium-term horizon. Its theoretical contributions, in building upon and supporting the limited knowledge on tactical planning in complex manufacturing operations, consist of a detailed tactical-level planning framework, identifying IPNs generated by uncertainty, pinpointing causal and moderating factors of MDS instability, and balancing complexity-reducing and complexity-absorbing strategies, cross-functional integrative mechanisms, IPMs, and dimensions of planning process quality. Meanwhile, its practical contributions consist of concise yet holistic descriptions of relationships between complexity in context and in demand– supply balancing. Manufacturers can readily capitalize on those descriptions to develop and implement context-appropriate tactical-level planning processes that enable efficient, informed, and effective decision-making

    Definition and evaluation of model-free coordination of electrical vehicle charging with reinforcement learning

    Get PDF
    Demand response (DR) becomes critical to manage the charging load of a growing electric vehicle (EV) deployment. Initial DR studies mainly adopt model predictive control, but models are largely uncertain for the EV scenario (e.g., customer behavior). Model-free approaches, based on reinforcement learning (RL), are an attractive alternative. We propose a new Markov decision process (MDP) formulation in the RL framework, to jointly coordinate a set of charging stations. State-of-the-art algorithms either focus on a single EV, or control an aggregate of EVs in multiple steps (e.g., 1) make aggregate load decisions and 2) translate the aggregate decision to individual EVs). In contrast, our RL approach jointly controls the whole set of EVs at once. We contribute a new MDP formulation with a scalable state representation independent of the number of charging stations. Using a batch RL algorithm, fitted QQ -iteration, we learn an optimal charging policy. With simulations using real-world data, we: 1) differentiate settings in training the RL policy (e.g., the time span covered by training data); 2) compare its performance to an oracle all-knowing benchmark (providing an upper performance bound); 3) analyze performance fluctuations throughout a full year; and 4) demonstrate generalization capacity to larger sets of charging stations

    Integrated Rapid Prototyping: Efficient Development of Custom Orthotics

    Get PDF
    Integrated Rapid Prototyping (IRP) is a systematic approach of optimizing product development from conception to realization, a process which we defined by the combination of 3D Digitization, 3D Modeling, Finite Element Methods, Additive Manufacturing, and Non-Destructive Testing. IRP has numerous applications, from consumer accessibility to industry level manufacturing. Our process was applied to the medical field through the creation of a custom orthotic device. A process done by using leg scans from a portable scanner, designing an orthotic model, analysis, fabrication via additive manufacturing, and product testing. Through this application, the development process was validated and analyzed by considering material characteristics, surface metrology, and full field optical techniques

    The Family of MapReduce and Large Scale Data Processing Systems

    Full text link
    In the last two decades, the continuous increase of computational power has produced an overwhelming flow of data which has called for a paradigm shift in the computing architecture and large scale data processing mechanisms. MapReduce is a simple and powerful programming model that enables easy development of scalable parallel applications to process vast amounts of data on large clusters of commodity machines. It isolates the application from the details of running a distributed program such as issues on data distribution, scheduling and fault tolerance. However, the original implementation of the MapReduce framework had some limitations that have been tackled by many research efforts in several followup works after its introduction. This article provides a comprehensive survey for a family of approaches and mechanisms of large scale data processing mechanisms that have been implemented based on the original idea of the MapReduce framework and are currently gaining a lot of momentum in both research and industrial communities. We also cover a set of introduced systems that have been implemented to provide declarative programming interfaces on top of the MapReduce framework. In addition, we review several large scale data processing systems that resemble some of the ideas of the MapReduce framework for different purposes and application scenarios. Finally, we discuss some of the future research directions for implementing the next generation of MapReduce-like solutions.Comment: arXiv admin note: text overlap with arXiv:1105.4252 by other author

    09061 Abstracts Collection -- Combinatorial Scientific Computing

    Get PDF
    From 01.02.2009 to 06.02.2009, the Dagstuhl Seminar 09061 ``Combinatorial Scientific Computing \u27\u27 was held in Schloss Dagstuhl -- Leibniz Center for Informatics. During the seminar, several participants presented their current research, and ongoing work and open problems were discussed. Abstracts of the presentations given during the seminar as well as abstracts of seminar results and ideas are put together in this paper. The first section describes the seminar topics and goals in general. Links to extended abstracts or full papers are provided, if available
    • …
    corecore