1,947 research outputs found

    A Tutorial on Cross-layer Optimization Wireless Network System Using TOPSIS Method

    Get PDF
    Each other, leading to issues such as interference, limited bandwidth, and varying channel conditions. These challenges require specialized optimization techniques tailored to the wireless environment. In wireless communication networks is to maximize the overall system throughput while ensuring fairness among users and maintaining quality of service requirements. This objective can be achieved through resource allocation optimization, where the available network resources such as bandwidth, power, and time slots are allocated to users in an optimal manner. Optimization-based approaches in wireless resource allocation typically involve formulating the resource allocation problem as an optimization problem with certain constraints.. These techniques provide practical solutions with reduced computational complexity, although they may not guarantee optimality. In summary, optimization-based approaches have been instrumental in studying resource allocation problems in communication networks, including the wireless domain. While techniques from the Internet setting have influenced the understanding of congestion control and protocol design, specific challenges in wireless networks necessitate tailored optimization techniques that account for interference, limited bandwidth, and varying channel conditions. power allocation problem in wireless ad hoc networks Cross-layer optimization refers to the process of jointly optimizing the allocation of resources across different layers of wireless networks, the interactions between different layers become more complex due to the shared medium and time-varying channel conditions.  Nash equilibrium, where no user can unilaterally improve its own performance by changing its strategy. Game theory can capture the distributed nature of wireless networks and provide insights into the behavior of users in resource allocation scenarios Additionally, heuristics and approximation algorithms are often employed in wireless resource allocation due to the complexity of the optimization problems involved. In traditional cellular systems, each user is allocated a fixed time slot for transmission, regardless of their channel conditions. However, in opportunistic scheduling. Alternative parameters for “Data rate Ž kbps, Geographic coverage ,  Service requirements , cost ” Evaluation parameter for “Circuit-switched cell, CDPD, WLAN, Paging, Satellite.” “the first ranking training is obtained with the lowest quality of compensation.

    Solution strategies for nonlinear conservation laws

    Get PDF
    Nonlinear conservation laws form the basis for models for a wide range of physical phenomena. Finding an optimal strategy for solving these problems can be challenging, and a good strategy for one problem may fail spectacularly for others. As different problems have different challenging features, exploiting knowledge about the problem structure is a key factor in achieving an efficient solution strategy. Most strategies found in literature for solving nonlinear problems involve a linearization step, usually using Newton's method, which replaces the original nonlinear problem by an iteration process consisting of a series of linear problems. A large effort is then spent on finding a good strategy for solving these linear problems. This involves choosing suitable preconditioners and linear solvers. This approach is in many cases a good choice and a multitude of different methods have been developed. However, the linearization step to some degree involves a loss of information about the original problem. This is not necessarily critical, but in many cases the structure of the nonlinear problem can be exploited to a larger extent than what is possible when working solely on the linearized problem. This may involve knowledge about dominating physical processes and specifically on whether a process is near equilibrium. By using nonlinear preconditioning techniques developed in recent years, certain attractive features such as automatic localization of computations to parts of the problem domain with the highest degree of nonlinearities arise. In the present work, these methods are further refined to obtain a framework for nonlinear preconditioning that also takes into account equilibrium information. This framework is developed mainly in the context of porous media, but in a general manner, allowing for application to a wide range of problems. A scalability study shows that the method is scalable for challenging two-phase flow problems. It is also demonstrated for nonlinear elasticity problems. Some models arising from nonlinear conservation laws are best solved using completely different strategies than the approach outlined above. One such example can be found in the field of surface gravity waves. For special types of nonlinear waves, such as solitary waves and undular bores, the well-known Korteweg-de Vries (KdV) equation has been shown to be a suitable model. This equation has many interesting properties not typical of nonlinear equations which may be exploited in the solver, and strategies usually reserved to linear problems may be applied. In this work includes a comparative study of two discretization methods with highly different properties for this equation

    Solution of systems of nonlinear equations Final report

    Get PDF
    Algorithm and computer program of diagonal discrimination method for computing nonlinear and transcendental function

    Design and resource management of reconfigurable multiprocessors for data-parallel applications

    Get PDF
    FPGA (Field-Programmable Gate Array)-based custom reconfigurable computing machines have established themselves as low-cost and low-risk alternatives to ASIC (Application-Specific Integrated Circuit) implementations and general-purpose microprocessors in accelerating a wide range of computation-intensive applications. Most often they are Application Specific Programmable Circuiits (ASPCs), which are developer programmable instead of user programmable. The major disadvantages of ASPCs are minimal programmability, and significant time and energy overheads caused by required hardware reconfiguration when the problem size outnumbers the available reconfigurable resources; these problems are expected to become more serious with increases in the FPGA chip size. On the other hand, dominant high-performance computing systems, such as PC clusters and SMPs (Symmetric Multiprocessors), suffer from high communication latencies and/or scalability problems. This research introduces low-cost, user-programmable and reconfigurable MultiProcessor-on-a-Programmable-Chip (MPoPC) systems for high-performance, low-cost computing. It also proposes a relevant resource management framework that deals with performance, power consumption and energy issues. These semi-customized systems reduce significantly runtime device reconfiguration by employing userprogrammable processing elements that are reusable for different tasks in large, complex applications. For the sake of illustration, two different types of MPoPCs with hardware FPUs (floating-point units) are designed and implemented for credible performance evaluation and modeling: the coarse-grain MIMD (Multiple-Instruction, Multiple-Data) CG-MPoPC machine based on a processor IP (Intellectual Property) core and the mixed-mode (MIMD, SIMD or M-SIMD) variant-grain HERA (HEterogeneous Reconfigurable Architecture) machine. In addition to alleviating the above difficulties, MPoPCs can offer several performance and energy advantages to our data-parallel applications when compared to ASPCs; they are simpler and more scalable, and have less verification time and cost. Various common computation-intensive benchmark algorithms, such as matrix-matrix multiplication (MMM) and LU factorization, are studied and their parallel solutions are shown for the two MPoPCs. The performance is evaluated with large sparse real-world matrices primarily from power engineering. We expect even further performance gains on MPoPCs in the near future by employing ever improving FPGAs. The innovative nature of this work has the potential to guide research in this arising field of high-performance, low-cost reconfigurable computing. The largest advantage of reconfigurable logic lies in its large degree of hardware customization and reconfiguration which allows reusing the resources to match the computation and communication needs of applications. Therefore, a major effort in the presented design methodology for mixed-mode MPoPCs, like HERA, is devoted to effective resource management. A two-phase approach is applied. A mixed-mode weighted Task Flow Graph (w-TFG) is first constructed for any given application, where tasks are classified according to their most appropriate computing mode (e.g., SIMD or MIMD). At compile time, an architecture is customized and synthesized for the TFG using an Integer Linear Programming (ILP) formulation and a parameterized hardware component library. Various run-time scheduling schemes with different performanceenergy objectives are proposed. A system-level energy model for HERA, which is based on low-level implementation data and run-time statistics, is proposed to guide performance-energy trade-off decisions. A parallel power flow analysis technique based on Newton\u27s method is proposed and employed to verify the methodology

    Mathematical modeling of tumor therapy with oncolytic viruses: Effects of parametric heterogeneity on cell dynamics

    Get PDF
    One of the mechanisms that ensure cancer robustness is tumor heterogeneity, and its effects on tumor cells dynamics have to be taken into account when studying cancer progression. There is no unifying theoretical framework in mathematical modeling of carcinogenesis that would account for parametric heterogeneity. Here we formulate a modeling approach that naturally takes stock of inherent cancer cell heterogeneity and illustrate it with a model of interaction between a tumor and an oncolytic virus. We show that several phenomena that are absent in homogeneous models, such as cancer recurrence, tumor dormancy, an others, appear in heterogeneous setting. We also demonstrate that, within the applied modeling framework, to overcome the adverse effect of tumor cell heterogeneity on cancer progression, a heterogeneous population of an oncolytic virus must be used. Heterogeneity in parameters of the model, such as tumor cell susceptibility to virus infection and virus replication rate, can lead to complex, time-dependent behaviors of the tumor. Thus, irregular, quasi-chaotic behavior of the tumor-virus system can be caused not only by random perturbations but also by the heterogeneity of the tumor and the virus. The modeling approach described here reveals the importance of tumor cell and virus heterogeneity for the outcome of cancer therapy. It should be straightforward to apply these techniques to mathematical modeling of other types of anticancer therapy.Comment: 45 pages, 6 figures; submitted to Biology Direc

    On Quasi‐Newton methods in fast Fourier transform‐based micromechanics

    Get PDF
    This work is devoted to investigating the computational power of Quasi‐Newton methods in the context of fast Fourier transform (FFT)‐based computational micromechanics. We revisit FFT‐based Newton‐Krylov solvers as well as modern Quasi‐Newton approaches such as the recently introduced Anderson accelerated basic scheme. In this context, we propose two algorithms based on the Broyden‐Fletcher‐Goldfarb‐Shanno (BFGS) method, one of the most powerful Quasi‐Newton schemes. To be specific, we use the BFGS update formula to approximate the global Hessian or, alternatively, the local material tangent stiffness. Both for Newton and Quasi‐Newton methods, a globalization technique is necessary to ensure global convergence. Specific to the FFT‐based context, we promote a Dong‐type line search, avoiding function evaluations altogether. Furthermore, we investigate the influence of the forcing term, that is, the accuracy for solving the linear system, on the overall performance of inexact (Quasi‐)Newton methods. This work concludes with numerical experiments, comparing the convergence characteristics and runtime of the proposed techniques for complex microstructures with nonlinear material behavior and finite as well as infinite material contrast

    Fair scheduling of bag-of-tasks applications using distributed Lagrangian optimization

    Get PDF
    International audienceLarge scale distributed systems typically comprise hundreds to millions of entities (applications, users, companies, universities) that have only a partial view of resources (computers, communication links). How to fairly and efficiently share such resources between entities in a distributed way has thus become a critical question. Although not all applications are suitable for execution on large scale distributed computing platform, ideal are the Bag-of-Tasks (BoT) applications. Hence a large fraction of jobs in workloads imposed on Grids is made of sequential applications submitted in the form of BoTs. Up until now, mainly simple mechanisms have been used to ensure a fair sharing of resources among these applications. Although these mechanisms are proved to be efficient for CPU-bound applications, they are known to be ineffective in the presence of network-bound applications. A possible answer resorts to Lagrangian optimization and distributed gradient descent. Under certain conditions, the resource sharing problem can be formulated as a global optimization problem, which can be solved by a distributed self-stabilizing supply and demand algorithm. In the last decade, this technique has been applied to design various network protocols (variants of TCP, multi-path network protocols, wireless network protocols) and even distributed algorithms for smart grids. In this article, we explain how to use this technique for fairly scheduling concurrent BoT applications with arbitrary communication-to-computation ratio on a Grid. Yet, application heterogeneity raises severe convergence and stability issues that did not appear in the previous contexts and need to be addressed by non-trivial modifications. The effectiveness of our proposal is assessed through an extensive set of complex and realistic simulations

    Robust electronic circuit design using evolutionary and Taguchi methods

    Get PDF
    Bibliography: pages 80-81.In engineering, there is a wide range of applications where genetic optimizers are used. Two genetic optimizers used in this thesis namely, Population Based Incremental Learning ( PBIL ) and Cross generational selection Heterogeneous crossover Cataclysmic mutation ( CHC ), are tested on a series of circuit problems to fmd if robust electronic circuits can be built from evolutionary methods. The evolutionary algorithms were used to search the space of discrete component values from a range of manufactured preferred values to obtain robust electronic circuits. Parasitic effects were also modelled in the simulation to provide for a more realistic circuit

    A differential semblance algorithm for the inverse problem of reflection seismology

    Get PDF
    AbstractThis paper presents an analysis of stability and convergence for a special case of differential semblance optimization (DSO). This approach to model estimation for reflection seismology is a variant of the output least squares inversion of seismograms, enjoying analytical and numerical properties superior to those of more straightforward versions. We study a specialization of DSO appropriate to the inversion of convolutional-approximation planewave seismograms over layered constant-density acoustic media. We prove that the differential semblance variational principle is locally convex in suitable model classes for a range of data noise. Moreover, the structure of the convexity estimates suggest a family of quasi-Newton algorithms. We describe an implementation of one of these algorithms, and present some numerical results
    corecore