440 research outputs found

    A programming-language extension for distributed real-time systems

    Get PDF
    In this paper we propose a method for extending programming languages that enables the specification of timing properties of systems. The way time is treated is not language specific and the extension can therefore be included in many existing programming languages. The presented method includes a view on the system development process. An essential feature is that it enables the construction of (hard) real-time programs that may be proven correct independently of the properties of the machines that are used for their execution. It therefore provides a similar abstraction from the execution platform as is normal for non-real-time languages. The aim of this paper is to illustrate the method and demonstrate its applicability to actual real-time problems. To this end we define a simple programming language that includes the timing extension. We present a formal semantics for a characteristic part of the language constructs and apply formal methods to prove the correctness of a small example program. We consider in detail a larger example, namely the mine-pump problem known from the literature. We construct a real-time program for this problem and describe various ways to map the program to an implementation for different platforms

    Flexible Scheduling in Middleware for Distributed rate-based real-time applications - Doctoral Dissertation, May 2002

    Get PDF
    Distributed rate-based real-time systems, such as process control and avionics mission computing systems, have traditionally been scheduled statically. Static scheduling provides assurance of schedulability prior to run-time overhead. However, static scheduling is brittle in the face of unanticipated overload, and treats invocation-to-invocation variations in resource requirements inflexibly. As a consequence, processing resources are often under-utilized in the average case, and the resulting systems are hard to adapt to meet new real-time processing requirements. Dynamic scheduling offers relief from the limitations of static scheduling. However, dynamic scheduling offers relief from the limitations of static scheduling. However, dynamic scheduling often has a high run-time cost because certain decisions are enforced on-line. Furthermore, under conditions of overload tasks can be scheduled dynamically that may never be dispatched, or that upon dispatch would miss their deadlines. We review the implications of these factors on rate-based distributed systems, and posits the necessity to combine static and dynamic approaches to exploit the strengths and compensate for the weakness of either approach in isolation. We present a general hybrid approach to real-time scheduling and dispatching in middleware, that can employ both static and dynamic components. This approach provides (1) feasibility assurance for the most critical tasks, (2) the ability to extend this assurance incrementally to operations in successively lower criticality equivalence classes, (3) the ability to trade off bounds on feasible utilization and dispatching over-head in cases where, for example, execution jitter is a factor or rates are not harmonically related, and (4) overall flexibility to make more optimal use of scarce computing resources and to enforce a wider range of application-specified execution requirements. This approach also meets additional constraints of an increasingly important class of rate-based systems, those with requirements for robust management of real-time performance in the face of rapidly and widely changing operating conditions. To support these requirements, we present a middleware framework that implements the hybrid scheduling and dispatching approach described above, and also provides support for (1) adaptive re-scheduling of operations at run-time and (2) reflective alternation among several scheduling strategies to improve real-time performance in the face of changing operating conditions. Adaptive re-scheduling must be performed whenever operating conditions exceed the ability of the scheduling and dispatching infrastructure to meet the critical real-time requirements of the system under the currently specified rates and execution times of operations. Adaptive re-scheduling relies on the ability to change the rates of execution of at least some operations, and may occur under the control of a higher-level middleware resource manager. Different rates of execution may be specified under different operating conditions, and the number of such possible combinations may be arbitrarily large. Furthermore, adaptive rescheduling may in turn require notification of rate-sensitive application components. It is therefore desirable to handle variations in operating conditions entirely within the scheduling and dispatching infrastructure when possible. A rate-based distributed real-time application, or a higher-level resource manager, could thus fall back on adaptive re-scheduling only when it cannot achieve acceptable real-time performance through self-adaptation. Reflective alternation among scheduling heuristics offers a way to tune real-time performance internally, and we offer foundational support for this approach. In particular, run-time observable information such as that provided by our metrics-feedback framework makes it possible to detect that a given current scheduling heuristic is underperforming the level of service another could provide. Furthermore we present empirical results for our framework in a realistic avionics mission computing environment. This forms the basis for guided adaption. This dissertation makes five contributions in support of flexible and adaptive scheduling and dispatching in middleware. First, we provide a middle scheduling framework that supports arbitrary and fine-grained composition of static/dynamic scheduling, to assure critical timeliness constraints while improving noncritical performance under a range of conditions. Second, we provide a flexible dispatching infrastructure framework composed of fine-grained primitives, and describe how appropriate configurations can be generated automatically based on the output of the scheduling framework. Third, we describe algorithms to reduce the overhead and duration of adaptive rescheduling, based on sorting for rate selection and priority assignment. Fourth, we provide timely and efficient performance information through an optimized metrics-feedback framework, to support higher-level reflection and adaptation decisions. Fifth, we present the results of empirical studies to quantify and evaluate the performance of alternative canonical scheduling heuristics, across a range of load and load jitter conditions. These studies were conducted within an avionics mission computing applications framework running on realistic middleware and embedded hardware. The results obtained from these studies (1) demonstrate the potential benefits of reflective alternation among distinct scheduling heuristics at run-time, and (2) suggest performance factors of interest for future work on adaptive control policies and mechanisms using this framework

    PC tools for project management: Programs and the state-of-the-practice

    Get PDF
    The use of microcomputer tools for NASA project management; which features are the most useful; the impact of these tools on job performance and individual style; and the prospects for new features in project management tools and related tools are addressed. High, mid, and low end PM tools are examined. The pro's and con's of the tools are assessed relative to various tasks. The strengths and weaknesses of the tools are presented through cases and demonstrations

    Tele-Autonomous control involving contact

    Get PDF
    Object localization and its application in tele-autonomous systems are studied. Two object localization algorithms are presented together with the methods of extracting several important types of object features. The first algorithm is based on line-segment to line-segment matching. Line range sensors are used to extract line-segment features from an object. The extracted features are matched to corresponding model features to compute the location of the object. The inputs of the second algorithm are not limited only to the line features. Featured points (point to point matching) and featured unit direction vectors (vector to vector matching) can also be used as the inputs of the algorithm, and there is no upper limit on the number of the features inputed. The algorithm will allow the use of redundant features to find a better solution. The algorithm uses dual number quaternions to represent the position and orientation of an object and uses the least squares optimization method to find an optimal solution for the object's location. The advantage of using this representation is that the method solves for the location estimation by minimizing a single cost function associated with the sum of the orientation and position errors and thus has a better performance on the estimation, both in accuracy and speed, than that of other similar algorithms. The difficulties when the operator is controlling a remote robot to perform manipulation tasks are also discussed. The main problems facing the operator are time delays on the signal transmission and the uncertainties of the remote environment. How object localization techniques can be used together with other techniques such as predictor display and time desynchronization to help to overcome these difficulties are then discussed

    Regional Moment Tensors of the 2009 L'Aquila Earthquake Sequence

    Get PDF
    Broadband waveform inversion of ground velocities in the 0.02 0.10 Hz frequency band is successfully applied to 181 earthquakes with ML ≥ 3 of the April, 2009, L'Aquila, Italy, earthquake sequence. This was made possible by the development of a new regional crustal velocity model constrained by deep crustal profiles, surfacewave dispersion and teleseismic Pwave receiver functions and tested through waveform fit. Although all earthquakes exhibit normal faulting, with the fault plane dipping southwest at about 55º for the majority of events, a subset of events had much shallower dips. The issue of confidence in the derived parameters was investigated by applying the same inversion procedure by two groups who subjectively selected different traces for inversion. The unexpected difficulty in modeling the regional broadband waveforms of the mainshock as a point source was investigated through an extensive finitefault modeling of broadband velocity and accelerometer data, which placed the location of major moment release updip and about 47 seconds after the initial firstarrival hypocentral parameters

    Proceedings of 31st Annual ARCOM Conference, vol 2

    Get PDF

    Robust optimization over time : a critical review

    Get PDF
    Robust optimization over time (ROOT) is the combination of robust optimization and dynamic optimization. In ROOT, frequent changes to deployed solutions are undesirable, which can be due to the high cost of switching between deployed solutions, limitations on the resources required to deploy new solutions, and/or the system’s inability to tolerate frequent changes in the deployed solutions. ROOT is dedicated to the study and development of algorithms capable of dealing with the implications of deploying or maintaining solutions over longer time horizons involving multiple environmental changes. This paper presents an in-depth review of the research on ROOT. The overarching aim of this survey is to help researchers gain a broad perspective on the current state of the field, what has been achieved so far, and the existing challenges and pitfalls. This survey also aims to improve accessibility and clarity by standardizing terminology and unifying mathematical notions used across the field, providing explicit mathematical formulations of definitions, and improving many existing mathematical descriptions. Moreover, we classify ROOT problems based on two ROOT-specific criteria: the requirements for changing or keeping deployed solutions and the number of deployed solutions. This classification helps researchers gain a better understanding of the characteristics and requirements of ROOT problems, which is crucial to systematic algorithm design and benchmarking. Additionally, we classify ROOT methods based on the approach they use for finding robust solutions and provide a comprehensive review of them. This survey also reviews ROOT benchmarks and performance indicators. Finally, we identify several future research directions

    Process algebra approach to parallel DBMS performance modelling

    Get PDF
    Abstract unavailable please refer to PD

    Monetary Rules for Small, Open, Emerging Economies

    Get PDF
    This paper develops a variant of the IMF's Global Economic Model (GEM) suitable to analyze macroeconomic dynamics in open economies, and uses it to assess the effectiveness of Taylor rules and Inflation-Forecast-Based (IFB) rules in stabilizing variability in output and inflation. Our findings suggest that a simple IFB rule that does not rely upon any direct estimates of the equilibrium real interest rate and places a relatively high weight on the inflation forecast may perform better in small open economies than conventional Taylor rules.
    • …
    corecore