648 research outputs found

    Optimality Properties, Distributed Strategies, and Measurement-Based Evaluation of Coordinated Multicell OFDMA Transmission

    Full text link
    The throughput of multicell systems is inherently limited by interference and the available communication resources. Coordinated resource allocation is the key to efficient performance, but the demand on backhaul signaling and computational resources grows rapidly with number of cells, terminals, and subcarriers. To handle this, we propose a novel multicell framework with dynamic cooperation clusters where each terminal is jointly served by a small set of base stations. Each base station coordinates interference to neighboring terminals only, thus limiting backhaul signalling and making the framework scalable. This framework can describe anything from interference channels to ideal joint multicell transmission. The resource allocation (i.e., precoding and scheduling) is formulated as an optimization problem (P1) with performance described by arbitrary monotonic functions of the signal-to-interference-and-noise ratios (SINRs) and arbitrary linear power constraints. Although (P1) is non-convex and difficult to solve optimally, we are able to prove: 1) Optimality of single-stream beamforming; 2) Conditions for full power usage; and 3) A precoding parametrization based on a few parameters between zero and one. These optimality properties are used to propose low-complexity strategies: both a centralized scheme and a distributed version that only requires local channel knowledge and processing. We evaluate the performance on measured multicell channels and observe that the proposed strategies achieve close-to-optimal performance among centralized and distributed solutions, respectively. In addition, we show that multicell interference coordination can give substantial improvements in sum performance, but that joint transmission is very sensitive to synchronization errors and that some terminals can experience performance degradations.Comment: Published in IEEE Transactions on Signal Processing, 15 pages, 7 figures. This version corrects typos related to Eq. (4) and Eq. (28

    Adaptive structured parallelism

    Get PDF
    Algorithmic skeletons abstract commonly-used patterns of parallel computation, communication, and interaction. Parallel programs are expressed by interweaving parameterised skeletons analogously to the way in which structured sequential programs are developed, using well-defined constructs. Skeletons provide top-down design composition and control inheritance throughout the program structure. Based on the algorithmic skeleton concept, structured parallelism provides a high-level parallel programming technique which allows the conceptual description of parallel programs whilst fostering platform independence and algorithm abstraction. By decoupling the algorithm specification from machine-dependent structural considerations, structured parallelism allows programmers to code programs regardless of how the computation and communications will be executed in the system platform.Meanwhile, large non-dedicated multiprocessing systems have long posed a challenge to known distributed systems programming techniques as a result of the inherent heterogeneity and dynamism of their resources. Scant research has been devoted to the use of structural information provided by skeletons in adaptively improving program performance, based on resource utilisation. This thesis presents a methodology to improve skeletal parallel programming in heterogeneous distributed systems by introducing adaptivity through resource awareness. As we hypothesise that a skeletal program should be able to adapt to the dynamic resource conditions over time using its structural forecasting information, we have developed ASPara: Adaptive Structured Parallelism. ASPara is a generic methodology to incorporate structural information at compilation into a parallel program, which will help it to adapt at execution

    Adaptive Data Parallelism for Internet Clients on Heterogeneous Platforms

    Get PDF
    Il Web moderno ha da molto superato le pagine statiche, limitate alla formattazione HTML e poche immagini. Siamo entrati i un era di Rich Internet Applications come giochi, simulazioni fisiche, rendering di immagini, elaborazione di foto, etc eseguite localmente dai programmi client. Nonostante questo gli attuali linguaggi lato client hanno limitatissime capacità di utilizzare le capacità computazionali della piattaforma, tipicamente eterogenea, sottostante. Presentiamo un DSL (Domain Specific Language) chiamato ASDP (ActionScript Data Parallel) integrato in ActionScript, uno dei linguaggi più popolari per la programmazione lato client e un parente prossimo di JavaScript. ASDP è molto similare ad ActionScript e permette frequentemente di introdurre la programmazione parallela con minime modifiche al codice sorgente. Presentiamo anche un prototipo di un sistema in cui computazioni data parallel possono essere eseguite su CPU o GPU. Il sistema runtime si occuperà di selezionare in modo trasparente la miglior unità computazionale a seconda della computazione, dell'architettura e del carico attuale del sistema. Vengono inoltre valutate le performance del sistema su diversi benchmark, rappresentativi dei seguenti tipi di applicazioni: fisica, elaborazione di immagini, calcolo scientifico e crittografia. Today’s Internet is long past static web pages full of HTML-formatted text sprinkled with an occasional image or animation. We have entered an era of Rich Internet Applications executed locally on Internet clients such as web browsers: games, physics engines, image rendering, photo editing, etc. And yet today’s languages used to program Internet clients have limited ability to tap to the computational capabilities of the underlying, often heterogeneous, platforms. We present how a Domain Specific Language (DSL) can be integrated into ActionScript, one of the most popular scripting languages used to program Internet clients and a close cousin of JavaScript. Our DSL, called ASDP (ActionScript Data Parallel), closely resembles ActionScript and often only minimal changes to existing ActionScript programs are required to enable data parallelism. We also present a prototype of a system, where data parallel workloads can be executed on either CPU or a GPU, with the runtime system transparently selecting the best processing unit, depending on the type of workload as well as the architecture and current load of the execution platform. We evaluate performance of our system on a variety of benchmarks, representing different types of workloads: physics, image processing, scientific computing and cryptography

    Optimisation sous contraintes de problèmes distribués par auto-organisation coopérative

    Get PDF
    Quotidiennement, divers problèmes d'optimisation : minimiser un coût de production, optimiser le parcours d'un véhicule, etc sont à résoudre. Ces problèmes se caractérisent par un degré élevé de complexité dû à l'hétérogénéité et la diversité des acteurs en jeu, à la masse importante des données ainsi qu'à la dynamique des environnements dans lesquels ils sont plongés. Face à la complexité croissante de ces applications, les approches de résolution classiques ont montré leurs limites. Depuis quelques années, la communauté scientifique s'intéresse aux développements de nouvelles solutions basées sur la distribution du calcul et la décentralisation du contrôle plus adaptées à ce genre de problème. La théorie des AMAS (Adaptive Multi-Agents Systems) propose le développement de solutions utilisant des systèmes multi-agents auto-adaptatifs par auto-organisation coopérative. Cette théorie a montré son adéquation pour la résolution de problèmes complexes et dynamiques, mais son application reste à un niveau d'abstraction assez élevé. L'objectif de ce travail est de spécialiser cette théorie pour la résolution de ce genre de problèmes. Ainsi, son utilisation en sera facilitée. Pour cela, le modèle d'agents AMAS4Opt avec des comportements et des interactions coopératifs et locaux a été défini. La validation s'est effectuée sur deux problèmes clés d'optimisation : le contrôle manufacturier et la conception de produit complexe. De plus, afin de montrer la robustesse et l'adéquation des solutions développées, un ensemble de critères d'évaluation permettant de souligner les points forts et faibles des systèmes adaptatifs et de les comparer à des systèmes existants a été défini.We solve problems and make decisions all day long. Some problems and decisions are very challenging: What is the best itinerary to deliver orders given the weather, the traffic and the hour? How to improve product manufacturing performances? etc. Problems that are characterized by a high level of complexity due to the heterogeneity and diversity of the participating actors, to the increasing volume of manipulated data and to the dynamics of the applications environments. Classical solving approaches have shown their limits to cope with this growing complexity. For the last several years, the scientific community has been interested in the development of new solutions based on computation distribution and control decentralization. The AMAS (Adaptive Multi-Agent-Systems) theory proposes to build solutions based on self-adaptive multi-agent systems using cooperative self-organization. This theory has shown its adequacy to solve different complex and dynamic problems, but remains at a high abstraction level. This work proposes a specialization of this theory for complex optimization problem solving under constraints. Thus, the usage of this theory is made accessible to different non-AMAS experts' engineers. Thus, the AMAS4Opt agent model with cooperative, local and generic behaviours and interactions has been defined.This model is validated on two well-known optimization problems: scheduling in manufacturing control and complex product design. Finally, in order to show the robustness and adequacy of the developed solutions, a set of evaluation criteria is proposed to underline the advantages and limits of adaptive systems and to compare them with already existing systems
    corecore