33,935 research outputs found

    Quality-aware model-driven service engineering

    Get PDF
    Service engineering and service-oriented architecture as an integration and platform technology is a recent approach to software systems integration. Quality aspects ranging from interoperability to maintainability to performance are of central importance for the integration of heterogeneous, distributed service-based systems. Architecture models can substantially influence quality attributes of the implemented software systems. Besides the benefits of explicit architectures on maintainability and reuse, architectural constraints such as styles, reference architectures and architectural patterns can influence observable software properties such as performance. Empirical performance evaluation is a process of measuring and evaluating the performance of implemented software. We present an approach for addressing the quality of services and service-based systems at the model-level in the context of model-driven service engineering. The focus on architecture-level models is a consequence of the black-box character of services

    Self-* overload control for distributed web systems

    Full text link
    Unexpected increases in demand and most of all flash crowds are considered the bane of every web application as they may cause intolerable delays or even service unavailability. Proper quality of service policies must guarantee rapid reactivity and responsiveness even in such critical situations. Previous solutions fail to meet common performance requirements when the system has to face sudden and unpredictable surges of traffic. Indeed they often rely on a proper setting of key parameters which requires laborious manual tuning, preventing a fast adaptation of the control policies. We contribute an original Self-* Overload Control (SOC) policy. This allows the system to self-configure a dynamic constraint on the rate of admitted sessions in order to respect service level agreements and maximize the resource utilization at the same time. Our policy does not require any prior information on the incoming traffic or manual configuration of key parameters. We ran extensive simulations under a wide range of operating conditions, showing that SOC rapidly adapts to time varying traffic and self-optimizes the resource utilization. It admits as many new sessions as possible in observance of the agreements, even under intense workload variations. We compared our algorithm to previously proposed approaches highlighting a more stable behavior and a better performance.Comment: The full version of this paper, titled "Self-* through self-learning: overload control for distributed web systems", has been published on Computer Networks, Elsevier. The simulator used for the evaluation of the proposed algorithm is available for download at the address: http://www.dsi.uniroma1.it/~novella/qos_web

    Regional Data Archiving and Management for Northeast Illinois

    Get PDF
    This project studies the feasibility and implementation options for establishing a regional data archiving system to help monitor and manage traffic operations and planning for the northeastern Illinois region. It aims to provide a clear guidance to the regional transportation agencies, from both technical and business perspectives, about building such a comprehensive transportation information system. Several implementation alternatives are identified and analyzed. This research is carried out in three phases. In the first phase, existing documents related to ITS deployments in the broader Chicago area are summarized, and a thorough review is conducted of similar systems across the country. Various stakeholders are interviewed to collect information on all data elements that they store, including the format, system, and granularity. Their perception of a data archive system, such as potential benefits and costs, is also surveyed. In the second phase, a conceptual design of the database is developed. This conceptual design includes system architecture, functional modules, user interfaces, and examples of usage. In the last phase, the possible business models for the archive system to sustain itself are reviewed. We estimate initial capital and recurring operational/maintenance costs for the system based on realistic information on the hardware, software, labor, and resource requirements. We also identify possible revenue opportunities. A few implementation options for the archive system are summarized in this report; namely: 1. System hosted by a partnering agency 2. System contracted to a university 3. System contracted to a national laboratory 4. System outsourced to a service provider The costs, advantages and disadvantages for each of these recommended options are also provided.ICT-R27-22published or submitted for publicationis peer reviewe

    The MINERν\nuA Data Acquisition System and Infrastructure

    Full text link
    MINERν\nuA (Main INjector ExpeRiment ν\nu-A) is a new few-GeV neutrino cross section experiment that began taking data in the FNAL NuMI (Fermi National Accelerator Laboratory Neutrinos at the Main Injector) beam-line in March of 2010. MINERν\nuA employs a fine-grained scintillator detector capable of complete kinematic characterization of neutrino interactions. This paper describes the MINERν\nuA data acquisition system (DAQ) including the read-out electronics, software, and computing architecture.Comment: 34 pages, 16 figure

    South Pole Telescope Software Systems: Control, Monitoring, and Data Acquisition

    Full text link
    We present the software system used to control and operate the South Pole Telescope. The South Pole Telescope is a 10-meter millimeter-wavelength telescope designed to measure anisotropies in the cosmic microwave background (CMB) at arcminute angular resolution. In the austral summer of 2011/12, the SPT was equipped with a new polarization-sensitive camera, which consists of 1536 transition-edge sensor bolometers. The bolometers are read out using 36 independent digital frequency multiplexing (\dfmux) readout boards, each with its own embedded processors. These autonomous boards control and read out data from the focal plane with on-board software and firmware. An overall control software system running on a separate control computer controls the \dfmux boards, the cryostat and all other aspects of telescope operation. This control software collects and monitors data in real-time, and stores the data to disk for transfer to the United States for analysis
    • …
    corecore