99,177 research outputs found
The CoLiS Platform for the Analysis of Maintainer Scripts in Debian Software Packages
International audienceThe software packages of the Debian distribution include more than twenty-seven thousand maintainer scripts in total, almost all of them being written in the Posix shell language. These scripts are executed with root privileges at installation, update, and removal of a package, which makes them critical for system maintenance. While the Debian policy provides guidance for package maintainers producing the scripts, only few tools exist to check the compliance of a script to that policy. We present CoLiS, a software platform for discovering violations of non-trivial properties required by the Debian policy in maintainer scripts. We describe our methodology which is based on symbolic execution and feature tree constraints, and we give an overview of the toolchain. We obtain promising results: our toolchain is effective in analysing a large set of Debian maintainer scripts, and it has already detected over 150 policy violations that have lead to bug reports, more than two-third of them now being fixed
Installing, Running and Maintaining Large Linux Clusters at CERN
Having built up Linux clusters to more than 1000 nodes over the past five
years, we already have practical experience confronting some of the LHC scale
computing challenges: scalability, automation, hardware diversity, security,
and rolling OS upgrades. This paper describes the tools and processes we have
implemented, working in close collaboration with the EDG project [1],
especially with the WP4 subtask, to improve the manageability of our clusters,
in particular in the areas of system installation, configuration, and
monitoring. In addition to the purely technical issues, providing shared
interactive and batch services which can adapt to meet the diverse and changing
requirements of our users is a significant challenge. We describe the
developments and tuning that we have introduced on our LSF based systems to
maximise both responsiveness to users and overall system utilisation. Finally,
this paper will describe the problems we are facing in enlarging our
heterogeneous Linux clusters, the progress we have made in dealing with the
current issues and the steps we are taking to gridify the clustersComment: 5 pages, Proceedings for the CHEP 2003 conference, La Jolla,
California, March 24 - 28, 200
BOA: Framework for Automated Builds
Managing large-scale software products is a complex software engineering
task. The automation of the software development, release and distribution
process is most beneficial in the large collaborations, where the big number of
developers, multiple platforms and distributed environment are typical factors.
This paper describes Build and Output Analyzer framework and its components
that have been developed in CMS to facilitate software maintenance and improve
software quality. The system allows to generate, control and analyze various
types of automated software builds and tests, such as regular rebuilds of the
development code, software integration for releases and installation of the
existing versions.Comment: 3 pages, 2 figure
Eight Things you should Know about Open Source Integrated Library Systems.
Open source library management systems are free alternative to costly commercial library systems. It helps to automate library functions and give a tremendous savings on library automation expenses. User ’s participation in all stages of software project ensures the development of the features that the library really wants. Lack of awareness and knowledge in open source technology among library professionals restrict wide adoption of open source library management system. This article gives an insight into the use and maintenance of open source library management systems
Optimization of the long-term planning of supply chains with decaying performance
This master's thesis addresses the optimization of supply and distribution chains considering the effect that equipment aging may cause over the performance of facilities involved in the process. The decaying performance of the facilities is modeled as an exponential equation and can be either physical or economic, thus giving rise to a novel mixed integer non-linear programming (MINLP) formulation. The optimization model has been developed based on a typical chemical supply chain. Thus, the best long-term investment plan has to be determined given production nodes, their production capacity and expected evolution; aggregated consumption nodes (urban or industrial districts) and their lumped demand (and expected evolution); actual and potential distribution nodes; distances between the nodes of the network; and a time horizon. The model includes the balances in each node, a general decaying performance function, and a cost function, as well as constraints to be satisfied. Hence, the investment plan (decision variables) consists not only on the start-up and shutdown of alternative distribution facilities, but also on the sizing of the lines satisfying the flows. The model has been implemented using GAMS optimization software. Results considering a variety of scenarios have been discussed. In addition, different approaches to the starting point for the model have been compared, showing the importance of initializing the optimization algorithm. The capabilities of the proposed approach have been tested through its application to two case studies: a natural gas network with physical decaying performance and an electricity distribution network with economic decaying performance. Each case study is solved with a different procedure to obtain results. Results demonstrate that overlooking the effect of equipment aging can lead to infeasible (for physical decaying performance) or unrealistic (for economic decaying performance) solutions in practice and show how the proposed model allows overcoming such limitations thus becoming a practical tool to support the decision-making process in the distribution secto
Recommended from our members
UC Berkeley's Cory Hall: Evaluation of Challenges and Potential Applications of Building-to-Grid Implementation
From September 2009 through June 2010, a team of researchers developed, installed, and tested instrumentation on the energy flows in Cory Hall on the UC Berkeley campus to create a Building-to-Grid testbed. The UC Berkeley team was headed by Professor David Culler, and assisted by members from EnerNex, Lawrence Berkeley National Laboratory, California State University Sacramento, and the California Institute for Energy & Environment. While the Berkeley team mapped the load tree of the building, EnerNex researched types of meters, submeters, monitors, and sensors to be used (Task 1). Next the UC Berkeley team analyzed building needs and designed the network of metering components and data storage/visualization software (Task 2). After meeting with vendors in January, the UCB team procured and installed the components starting in late March (Task 3). Next, the UCB team tested and demonstrated the system (Task 4). Meanwhile, the CSUS team documented the methodology and steps necessary to implement a testbed (Task 5) and Harold Galicer developed a roadmap for the CSUS Smart Grid Center with results from the testbed (Task 5a) and evaluated the Cory Hall implementation process (Task 5b). The CSUS team also worked with local utilities to develop an approach to the energy information communication link between buildings and the utility (Task 6). The UC Berkeley team then prepared a roadmap to outline necessary technology development for Building-to-Grid, and presented the results of the project in early July (Task 7). Finally, CIEE evaluated the implementation, noting challenges and potential applications of Building-to-Grid (Task 8). These deliverables are available at the i4Energy site: http://i4energy.org/
- …