21,667 research outputs found
Asynchronous iterative computations with Web information retrieval structures: The PageRank case
There are several ideas being used today for Web information retrieval, and
specifically in Web search engines. The PageRank algorithm is one of those that
introduce a content-neutral ranking function over Web pages. This ranking is
applied to the set of pages returned by the Google search engine in response to
posting a search query. PageRank is based in part on two simple common sense
concepts: (i)A page is important if many important pages include links to it.
(ii)A page containing many links has reduced impact on the importance of the
pages it links to. In this paper we focus on asynchronous iterative schemes to
compute PageRank over large sets of Web pages. The elimination of the
synchronizing phases is expected to be advantageous on heterogeneous platforms.
The motivation for a possible move to such large scale distributed platforms
lies in the size of matrices representing Web structure. In orders of
magnitude: pages with nonzero elements and bytes
just to store a small percentage of the Web (the already crawled); distributed
memory machines are necessary for such computations. The present research is
part of our general objective, to explore the potential of asynchronous
computational models as an underlying framework for very large scale
computations over the Grid. The area of ``internet algorithmics'' appears to
offer many occasions for computations of unprecedent dimensionality that would
be good candidates for this framework.Comment: 8 pages to appear at ParCo2005 Conference Proceeding
Extensions to the Estimation Calculus
Walther’s estimation calculus was designed to prove the termination of functional programs, and can also be used to solve the similar problem of proving the well-foundedness of induction rules. However, there are certain features of the goal formulae which are more common to the problem of induction rule well-foundedness than the problem of termination, and which the calculus cannot handle. We present a sound extension of the calculus that is capable of dealing with these features. The extension develops Walther’s concept of an argument bounded function in two ways: firstly, so that the function may be bounded below by its argument, and secondly, so that a bound may exist between two arguments of a predicate. Our calculus enables automatic proofs of the well-foundedness of a large class of induction rules not captured by the original calculus
Fitness Landscape-Based Characterisation of Nature-Inspired Algorithms
A significant challenge in nature-inspired algorithmics is the identification
of specific characteristics of problems that make them harder (or easier) to
solve using specific methods. The hope is that, by identifying these
characteristics, we may more easily predict which algorithms are best-suited to
problems sharing certain features. Here, we approach this problem using fitness
landscape analysis. Techniques already exist for measuring the "difficulty" of
specific landscapes, but these are often designed solely with evolutionary
algorithms in mind, and are generally specific to discrete optimisation. In
this paper we develop an approach for comparing a wide range of continuous
optimisation algorithms. Using a fitness landscape generation technique, we
compare six different nature-inspired algorithms and identify which methods
perform best on landscapes exhibiting specific features.Comment: 10 pages, 1 figure, submitted to the 11th International Conference on
Adaptive and Natural Computing Algorithm
Building fault detection data to aid diagnostic algorithm creation and performance testing.
It is estimated that approximately 4-5% of national energy consumption can be saved through corrections to existing commercial building controls infrastructure and resulting improvements to efficiency. Correspondingly, automated fault detection and diagnostics (FDD) algorithms are designed to identify the presence of operational faults and their root causes. A diversity of techniques is used for FDD spanning physical models, black box, and rule-based approaches. A persistent challenge has been the lack of common datasets and test methods to benchmark their performance accuracy. This article presents a first of its kind public dataset with ground-truth data on the presence and absence of building faults. This dataset spans a range of seasons and operational conditions and encompasses multiple building system types. It contains information on fault severity, as well as data points reflective of the measurements in building control systems that FDD algorithms typically have access to. The data were created using simulation models as well as experimental test facilities, and will be expanded over time
Optimal dynamic operations scheduling for small-scale satellites
A satellite's operations schedule is crafted based on each subsystem/payload operational needs, while taking into account the available resources on-board. A number of operating modes are carefully designed, each one with a different operations plan that can serve emergency cases, reduced functionality cases, the nominal case, the end of mission case and so on. During the mission span, should any operations planning amendments arise, a new schedule needs to be manually developed and uplinked to the satellite during a communications' window. The current operations planning techniques over a reduced number of solutions while approaching operations scheduling in a rigid manner. Given the complexity of a satellite as a system as well as the numerous restrictions and uncertainties imposed by both environmental and technical parameters, optimising the operations scheduling in an automated fashion can over a flexible approach while enhancing the mission robustness. In this paper we present Opt-OS (Optimised Operations Scheduler), a tool loosely based on the Ant Colony System algorithm, which can solve the Dynamic Operations Scheduling Problem (DOSP). The DOSP is treated as a single-objective multiple constraint discrete optimisation problem, where the objective is to maximise the useful operation time per subsystem on-board while respecting a set of constraints such as the feasible operation timeslot per payload or maintaining the power consumption below a specific threshold. Given basic mission inputs such as the Keplerian elements of the satellite's orbit, its launch date as well as the individual subsystems' power consumption and useful operation periods, Opt-OS outputs the optimal ON/OFF state per subsystem per orbital time step, keeping each subsystem's useful operation time to a maximum while ensuring that constraints such as the power availability threshold are never violated. Opt-OS can provide the flexibility needed for designing an optimal operations schedule on the spot throughout any mission phase as well as the ability to automatically schedule operations in case of emergency. Furthermore, Opt-OS can be used in conjunction with multi-objective optimisation tools for performing full system optimisation. Based on the optimal operations schedule, subsystem design parameters are being optimised in order to achieve the maximal usage of the satellite while keeping its mass minimal
12th International Workshop on Termination (WST 2012) : WST 2012, February 19–23, 2012, Obergurgl, Austria / ed. by Georg Moser
This volume contains the proceedings of the 12th International Workshop on Termination (WST 2012), to be held February 19–23, 2012 in Obergurgl, Austria. The goal of the Workshop on Termination is to be a venue for presentation and discussion of all topics in and around termination. In this way, the workshop tries to bridge the gaps between different communities interested and active in research in and around termination. The 12th International Workshop on Termination in Obergurgl continues the successful workshops held in St. Andrews (1993), La Bresse (1995), Ede (1997), Dagstuhl (1999), Utrecht (2001), Valencia (2003), Aachen (2004), Seattle (2006), Paris (2007), Leipzig (2009), and Edinburgh (2010). The 12th International Workshop on Termination did welcome contributions on all aspects of termination and complexity analysis. Contributions from the imperative, constraint, functional, and logic programming communities, and papers investigating applications of complexity or termination (for example in program transformation or theorem proving) were particularly welcome. We did receive 18 submissions which all were accepted. Each paper was assigned two reviewers. In addition to these 18 contributed talks, WST 2012, hosts three invited talks by Alexander Krauss, Martin Hofmann, and Fausto Spoto
A Big Data Analyzer for Large Trace Logs
Current generation of Internet-based services are typically hosted on large
data centers that take the form of warehouse-size structures housing tens of
thousands of servers. Continued availability of a modern data center is the
result of a complex orchestration among many internal and external actors
including computing hardware, multiple layers of intricate software, networking
and storage devices, electrical power and cooling plants. During the course of
their operation, many of these components produce large amounts of data in the
form of event and error logs that are essential not only for identifying and
resolving problems but also for improving data center efficiency and
management. Most of these activities would benefit significantly from data
analytics techniques to exploit hidden statistical patterns and correlations
that may be present in the data. The sheer volume of data to be analyzed makes
uncovering these correlations and patterns a challenging task. This paper
presents BiDAl, a prototype Java tool for log-data analysis that incorporates
several Big Data technologies in order to simplify the task of extracting
information from data traces produced by large clusters and server farms. BiDAl
provides the user with several analysis languages (SQL, R and Hadoop MapReduce)
and storage backends (HDFS and SQLite) that can be freely mixed and matched so
that a custom tool for a specific task can be easily constructed. BiDAl has a
modular architecture so that it can be extended with other backends and
analysis languages in the future. In this paper we present the design of BiDAl
and describe our experience using it to analyze publicly-available traces from
Google data clusters, with the goal of building a realistic model of a complex
data center.Comment: 26 pages, 10 figure
Recommended from our members
Local search: A guide for the information retrieval practitioner
There are a number of combinatorial optimisation problems in information retrieval in which the use of local search methods are worthwhile. The purpose of this paper is to show how local search can be used to solve some well known tasks in information retrieval (IR), how previous research in the field is piecemeal, bereft of a structure and methodologically flawed, and to suggest more rigorous ways of applying local search methods to solve IR problems. We provide a query based taxonomy for analysing the use of local search in IR tasks and an overview of issues such as fitness functions, statistical significance and test collections when conducting experiments on combinatorial optimisation problems. The paper gives a guide on the pitfalls and problems for IR practitioners who wish to use local search to solve their research issues, and gives practical advice on the use of such methods. The query based taxonomy is a novel structure which can be used by the IR practitioner in order to examine the use of local search in IR
- …