17,415 research outputs found
Many-Task Computing and Blue Waters
This report discusses many-task computing (MTC) generically and in the
context of the proposed Blue Waters systems, which is planned to be the largest
NSF-funded supercomputer when it begins production use in 2012. The aim of this
report is to inform the BW project about MTC, including understanding aspects
of MTC applications that can be used to characterize the domain and
understanding the implications of these aspects to middleware and policies.
Many MTC applications do not neatly fit the stereotypes of high-performance
computing (HPC) or high-throughput computing (HTC) applications. Like HTC
applications, by definition MTC applications are structured as graphs of
discrete tasks, with explicit input and output dependencies forming the graph
edges. However, MTC applications have significant features that distinguish
them from typical HTC applications. In particular, different engineering
constraints for hardware and software must be met in order to support these
applications. HTC applications have traditionally run on platforms such as
grids and clusters, through either workflow systems or parallel programming
systems. MTC applications, in contrast, will often demand a short time to
solution, may be communication intensive or data intensive, and may comprise
very short tasks. Therefore, hardware and software for MTC must be engineered
to support the additional communication and I/O and must minimize task dispatch
overheads. The hardware of large-scale HPC systems, with its high degree of
parallelism and support for intensive communication, is well suited for MTC
applications. However, HPC systems often lack a dynamic resource-provisioning
feature, are not ideal for task communication via the file system, and have an
I/O system that is not optimized for MTC-style applications. Hence, additional
software support is likely to be required to gain full benefit from the HPC
hardware
Software Effort Estimation Accuracy Prediction of Machine Learning Techniques: A Systematic Performance Evaluation
Software effort estimation accuracy is a key factor in effective planning,
controlling and to deliver a successful software project within budget and
schedule. The overestimation and underestimation both are the key challenges
for future software development, henceforth there is a continuous need for
accuracy in software effort estimation (SEE). The researchers and practitioners
are striving to identify which machine learning estimation technique gives more
accurate results based on evaluation measures, datasets and the other relevant
attributes. The authors of related research are generally not aware of
previously published results of machine learning effort estimation techniques.
The main aim of this study is to assist the researchers to know which machine
learning technique yields the promising effort estimation accuracy prediction
in the software development. In this paper, the performance of the machine
learning ensemble technique is investigated with the solo technique based on
two most commonly used accuracy evaluation metrics. We used the systematic
literature review methodology proposed by Kitchenham and Charters. This
includes searching for the most relevant papers, applying quality assessment
criteria, extracting data and drawing results. We have evaluated a
state-of-the-art accuracy performance of 28 selected studies (14 ensemble, 14
solo) using Mean Magnitude of Relative Error (MMRE) and PRED (25) as a set of
reliable accuracy metrics for performance evaluation of accuracy among two
techniques to report the research questions stated in this study. We found that
machine learning techniques are the most frequently implemented in the
construction of ensemble effort estimation (EEE) techniques. The results of
this study revealed that the EEE techniques usually yield a promising
estimation accuracy than the solo techniques.Comment: Pages: 27 Figures: 15 Tables:
Status and Future Perspectives for Lattice Gauge Theory Calculations to the Exascale and Beyond
In this and a set of companion whitepapers, the USQCD Collaboration lays out
a program of science and computing for lattice gauge theory. These whitepapers
describe how calculation using lattice QCD (and other gauge theories) can aid
the interpretation of ongoing and upcoming experiments in particle and nuclear
physics, as well as inspire new ones.Comment: 44 pages. 1 of USQCD whitepapers
Predicting and Evaluating Software Model Growth in the Automotive Industry
The size of a software artifact influences the software quality and impacts
the development process. In industry, when software size exceeds certain
thresholds, memory errors accumulate and development tools might not be able to
cope anymore, resulting in a lengthy program start up times, failing builds, or
memory problems at unpredictable times. Thus, foreseeing critical growth in
software modules meets a high demand in industrial practice. Predicting the
time when the size grows to the level where maintenance is needed prevents
unexpected efforts and helps to spot problematic artifacts before they become
critical.
Although the amount of prediction approaches in literature is vast, it is
unclear how well they fit with prerequisites and expectations from practice. In
this paper, we perform an industrial case study at an automotive manufacturer
to explore applicability and usability of prediction approaches in practice. In
a first step, we collect the most relevant prediction approaches from
literature, including both, approaches using statistics and machine learning.
Furthermore, we elicit expectations towards predictions from practitioners
using a survey and stakeholder workshops. At the same time, we measure software
size of 48 software artifacts by mining four years of revision history,
resulting in 4,547 data points. In the last step, we assess the applicability
of state-of-the-art prediction approaches using the collected data by
systematically analyzing how well they fulfill the practitioners' expectations.
Our main contribution is a comparison of commonly used prediction approaches
in a real world industrial setting while considering stakeholder expectations.
We show that the approaches provide significantly different results regarding
prediction accuracy and that the statistical approaches fit our data best
The institutional character of computerized information systems
We examine how important social and technical choices become part of the history of a computer-based information system (CB/SJ and embedded in the social structure which supports its development and use. These elements of a CBIS can be organized in specific ways to enhance its usability and performance. Paradoxically, they can also constrain future implementations and post-implementations.We argue that CBIS developed from complex, interdependent social and technical choices should be conceptualized in terms of their institutional characteristics, as well as their information-processing characteristics. The social system which supports the development and operation of a CBIS is one major element whose institutional characteristics can effectively support routine activities while impeding substantial innovation. Characterizing CBIS as institutions is important for several reasons: (1) the usability of CBIS is more critical than the abstract information-processing capabilities of the underlying technology; (2) CBIS that are well-used and have stable social structures are more difficult to replace than those with less developed social structures and fewer participants; (3) CBIS vary from one social setting to another according to the ways in which they are organized and embedded in organized social systems. These ideas are illustrated with the case study of a failed attempt to convert a complex inventory control system in a medium-sized manufacturing firm
An Intelligent Framework for Estimating Software Development Projects using Machine Learning
The IT industry has faced many challenges related to software effort and cost estimation. A cost assessment is conducted after software effort estimation, which benefits customers as well as developers. The purpose of this paper is to discuss various methods for the estimation of software effort and cost in the context of software engineering, such as algorithmic methods, expert judgment methods, analogy-based estimation methods, and machine learning methods, as well as their different aspects. In spite of this, estimation of the effort involved in software development are subject to uncertainty. Several methods have been developed in the literature for improving estimation accuracy, many of which involve the use of machine learning techniques. A machine learning framework is proposed in this paper to address this challenging problem. In addition to being completely independent of algorithmic models and estimation problems, this framework also features a modular architecture. It has high interpretability, learning capability, and robustness to imprecise and uncertain inputs
- …