23,107 research outputs found

    CAutoCSD-evolutionary search and optimisation enabled computer automated control system design

    Get PDF
    This paper attempts to set a unified scene for various linear time-invariant (LTI) control system design schemes, by transforming the existing concept of 'Computer-Aided Control System Design' (CACSD) to the novel 'Computer-Automated Control System Design' (CAutoCSD). The first step towards this goal is to accommodate, under practical constraints, various design objectives that are desirable in both time and frequency-domains. Such performance-prioritised unification is aimed to relieve practising engineers from having to select a particular control scheme and from sacrificing certain performance goals resulting from pre-committing to the adopted scheme. With the recent progress in evolutionary computing based extra-numeric, multi-criterion search and optimisation techniques, such unification of LTI control schemes becomes feasible, analytically and practically, and the resultant designs can be creative. The techniques developed are applied to, and illustrated by, three design problems. The unified approach automatically provides an integrator for zero-steady state error in velocity control of a DC motor, meets multiple objectives in designing an LTI controller for a non-minimum phase plant and offers a high-performing LTI controller network for a nonlinear chemical process

    Decision-based genetic algorithms for solving multi-period project scheduling with dynamically experienced workforce

    Get PDF
    The importance of the flexibility of resources increased rapidly with the turbulent changes in the industrial context, to meet the customers’ requirements. Among all resources, the most important and considered as the hardest to manage are human resources, in reasons of availability and/or conventions. In this article, we present an approach to solve project scheduling with multi-period human resources allocation taking into account two flexibility levers. The first is the annual hours and working time regulation, and the second is the actors’ multi-skills. The productivity of each operator was considered as dynamic, developing or degrading depending on the prior allocation decisions. The solving approach mainly uses decision-based genetic algorithms, in which, chromosomes don’t represent directly the problem solution; they simply present three decisions: tasks’ priorities for execution, actors’ priorities for carrying out these tasks, and finally the priority of working time strategy that can be considered during the specified working period. Also the principle of critical skill was taken into account. Based on these decisions and during a serial scheduling generating scheme, one can in a sequential manner introduce the project scheduling and the corresponding workforce allocations

    Practical Combinatorial Interaction Testing: Empirical Findings on Efficiency and Early Fault Detection

    Get PDF
    Combinatorial interaction testing (CIT) is important because it tests the interactions between the many features and parameters that make up the configuration space of software systems. Simulated Annealing (SA) and Greedy Algorithms have been widely used to find CIT test suites. From the literature, there is a widely-held belief that SA is slower, but produces more effective tests suites than Greedy and that SA cannot scale to higher strength coverage. We evaluated both algorithms on seven real-world subjects for the well-studied two-way up to the rarely-studied six-way interaction strengths. Our findings present evidence to challenge this current orthodoxy: real-world constraints allow SA to achieve higher strengths. Furthermore, there was no evidence that Greedy was less effective (in terms of time to fault revelation) compared to SA; the results for the greedy algorithm are actually slightly superior. However, the results are critically dependent on the approach adopted to constraint handling. Moreover, we have also evaluated a genetic algorithm for constrained CIT test suite generation. This is the first time strengths higher than 3 and constraint handling have been used to evaluate GA. Our results show that GA is competitive only for pairwise testing for subjects with a small number of constraints

    A Hybrid multi-agent architecture and heuristics generation for solving meeting scheduling problem

    Get PDF
    Agent-based computing has attracted much attention as a promising technique for application domains that are distributed, complex and heterogeneous. Current research on multi-agent systems (MAS) has become mature enough to be applied as a technology for solving problems in an increasingly wide range of complex applications. The main formal architectures used to describe the relationships between agents in MAS are centralised and distributed architectures. In computational complexity theory, researchers have classified the problems into the followings categories: (i) P problems, (ii) NP problems, (iii) NP-complete problems, and (iv) NP-hard problems. A method for computing the solution to NP-hard problems, using the algorithms and computational power available nowadays in reasonable time frame remains undiscovered. And unfortunately, many practical problems belong to this very class. On the other hand, it is essential that these problems are solved, and the only possibility of doing this is to use approximation techniques. Heuristic solution techniques are an alternative. A heuristic is a strategy that is powerful in general, but not absolutely guaranteed to provide the best (i.e. optimal) solutions or even find a solution. This demands adopting some optimisation techniques such as Evolutionary Algorithms (EA). This research has been undertaken to investigate the feasibility of running computationally intensive algorithms on multi-agent architectures while preserving the ability of small agents to run on small devices, including mobile devices. To achieve this, the present work proposes a new Hybrid Multi-Agent Architecture (HMAA) that generates new heuristics for solving NP-hard problems. This architecture is hybrid because it is "semi-distributed/semi-centralised" architecture where variables and constraints are distributed among small agents exactly as in distributed architectures, but when the small agents become stuck, a centralised control becomes active where the variables are transferred to a super agent, that has a central view of the whole system, and possesses much more computational power and intensive algorithms to generate new heuristics for the small agents, which find optimal solution for the specified problem. This research comes up with the followings: (1) Hybrid Multi-Agent Architecture (HMAA) that generates new heuristic for solving many NP-hard problems. (2) Two frameworks of HMAA have been implemented; search and optimisation frameworks. (3) New SMA meeting scheduling heuristic. (4) New SMA repair strategy for the scheduling process. (5) Small Agent (SMA) that is responsible for meeting scheduling has been developed. (6) “Local Search Programming” (LSP), a new concept for evolutionary approaches, has been introduced. (7) Two types of super-agent (LGP_SUA and LSP_SUA) have been implemented in the HMAA, and two SUAs (local and global optima) have been implemented for each type. (8) A prototype for HMAA has been implemented: this prototype employs the proposed meeting scheduling heuristic with the repair strategy on SMAs, and the four extensive algorithms on SUAs. The results reveal that this architecture is applicable to many different application domains because of its simplicity and efficiency. Its performance was better than many existing meeting scheduling architectures. HMAA can be modified and altered to other types of evolutionary approaches

    Pilot3 D2.1 - Trade-off report on multi criteria decision making techniques

    Get PDF
    This deliverable describes the decision making approach that will be followed in Pilot3. It presents a domain-driven analysis of the characteristics of Pilot3 objective function and optimisation framework. This has been done considering inputs from deliverable D1.1 - Technical Resources and Problem definition, from interaction with the Topic Manager, but most importantly from a dedicated Advisory Board workshop and follow-up consultation. The Advisory Board is formed by relevant stakeholders including airlines, flight operation experts, pilots, and other relevant ATM experts. A review of the different multi-criteria decision making techniques available in the literature is presented. Considering the domain-driven characteristics of Pilot3 and inputs on how the tool could be used by airlines and crew. Then, the most suitable methods for multi-criteria optimisation are selected for each of the phases of the optimisation framework

    The XMM-Newton Wide Angle Survey (XWAS): the X-ray spectrum of type-1 AGN

    Get PDF
    We discuss the broad band X-ray properties of one of the largest samples of X-ray selected type-1 AGN to date (487 objects in total), drawn from the XMM-Newton Wide Angle Survey. The objects cover 2-10 keV luminosities from ~10^{42}-10^{45} erg s^{-1} and are detected up to redshift ~4. We constrain the overall properties of the broad band continuum, soft excess and X-ray absorption, along with their dependence on the X-ray luminosity and redshift and we discuss the implications for models of AGN emission. We constrained the mean spectral index of the broad band X-ray continuum to =1.96+-0.02 with intrinsic dispersion sigma=0.27_{-0.02}^{+0.01}. The continuum becomes harder at faint fluxes and at higher redshifts and luminosities. The dependence of Gamma with flux is likely due to undetected absorption rather than to spectral variation. We found a strong dependence of the detection efficiency of objects on the spectral shape which can have a strong impact on the measured mean continuum shapes of sources at different redshifts and luminosities. We detected excess absorption in ~3% of our objects, with column densities ~a few x10^{22} cm^{-2}. The apparent mismatch between the optical classification and X-ray properties of these objects is a challenge for the standard AGN unification model. We found that the fraction of objects with detected soft excess is ~36%. Using a thermal model, we constrained the soft excess mean temperature and intrinsic dispersion to ~100 eV and sigma~34 eV. The origin of the soft excess as thermal emission from the accretion disk or Compton scattered disk emission is ruled out on the basis of the temperatures detected and the lack of correlation of the measured temperature with the X-ray luminosity (abridged).Comment: 13 pages, 24 figures, Accepted for publication in Astronomy and Astrophysic

    Modelling the Strategic Alignment of Software Requirements using Goal Graphs

    Get PDF
    This paper builds on existing Goal Oriented Requirements Engineering (GORE) research by presenting a methodology with a supporting tool for analysing and demonstrating the alignment between software requirements and business objectives. Current GORE methodologies can be used to relate business goals to software goals through goal abstraction in goal graphs. However, we argue that unless the extent of goal-goal contribution is quantified with verifiable metrics and confidence levels, goal graphs are not sufficient for demonstrating the strategic alignment of software requirements. We introduce our methodology using an example software project from Rolls-Royce. We conclude that our methodology can improve requirements by making the relationships to business problems explicit, thereby disambiguating a requirement's underlying purpose and value.Comment: v2 minor updates: 1) bitmap images replaced with vector, 2) reworded related work ref[6] for clarit

    Relating adults' lives and learning: participation and engagement in different settings

    Get PDF
    This report shows how an understanding of language, literacy andnumeracy as social practices can help practitioners to take account oflearners' lives. It demonstrates how people's histories, currentcircumstances and imagined futures can shape their learning andaffect their level of engagement. The study is based on the research ofthe Adult Learners' Lives project in community settings in Blackburn,Lancaster and Liverpool

    SPINN: Synergistic Progressive Inference of Neural Networks over Device and Cloud

    Full text link
    Despite the soaring use of convolutional neural networks (CNNs) in mobile applications, uniformly sustaining high-performance inference on mobile has been elusive due to the excessive computational demands of modern CNNs and the increasing diversity of deployed devices. A popular alternative comprises offloading CNN processing to powerful cloud-based servers. Nevertheless, by relying on the cloud to produce outputs, emerging mission-critical and high-mobility applications, such as drone obstacle avoidance or interactive applications, can suffer from the dynamic connectivity conditions and the uncertain availability of the cloud. In this paper, we propose SPINN, a distributed inference system that employs synergistic device-cloud computation together with a progressive inference method to deliver fast and robust CNN inference across diverse settings. The proposed system introduces a novel scheduler that co-optimises the early-exit policy and the CNN splitting at run time, in order to adapt to dynamic conditions and meet user-defined service-level requirements. Quantitative evaluation illustrates that SPINN outperforms its state-of-the-art collaborative inference counterparts by up to 2x in achieved throughput under varying network conditions, reduces the server cost by up to 6.8x and improves accuracy by 20.7% under latency constraints, while providing robust operation under uncertain connectivity conditions and significant energy savings compared to cloud-centric execution.Comment: Accepted at the 26th Annual International Conference on Mobile Computing and Networking (MobiCom), 202
    • 

    corecore