111 research outputs found
Interconnection networks for parallel and distributed computing
Parallel computers are generally either shared-memory machines or distributed- memory machines. There are currently technological limitations on shared-memory architectures and so parallel computers utilizing a large number of processors tend tube distributed-memory machines. We are concerned solely with distributed-memory multiprocessors. In such machines, the dominant factor inhibiting faster global computations is inter-processor communication. Communication is dependent upon the topology of the interconnection network, the routing mechanism, the flow control policy, and the method of switching. We are concerned with issues relating to the topology of the interconnection network. The choice of how we connect processors in a distributed-memory multiprocessor is a fundamental design decision. There are numerous, often conflicting, considerations to bear in mind. However, there does not exist an interconnection network that is optimal on all counts and trade-offs have to be made. A multitude of interconnection networks have been proposed with each of these networks having some good (topological) properties and some not so good. Existing noteworthy networks include trees, fat-trees, meshes, cube-connected cycles, butterflies, Möbius cubes, hypercubes, augmented cubes, k-ary n-cubes, twisted cubes, n-star graphs, (n, k)-star graphs, alternating group graphs, de Bruijn networks, and bubble-sort graphs, to name but a few. We will mainly focus on k-ary n-cubes and (n, k)-star graphs in this thesis. Meanwhile, we propose a new interconnection network called augmented k-ary n- cubes. The following results are given in the thesis.1. Let k ≥ 4 be even and let n ≥ 2. Consider a faulty k-ary n-cube Q(^k_n) in which the number of node faults f(_n) and the number of link faults f(_e) are such that f(_n) + f(_e) ≤ 2n - 2. We prove that given any two healthy nodes s and e of Q(^k_n), there is a path from s to e of length at least k(^n) - 2f(_n) - 1 (resp. k(^n) - 2f(_n) - 2) if the nodes s and e have different (resp. the same) parities (the parity of a node Q(^k_n) in is the sum modulo 2 of the elements in the n-tuple over 0, 1, ∙∙∙ , k - 1 representing the node). Our result is optimal in the sense that there are pairs of nodes and fault configurations for which these bounds cannot be improved, and it answers questions recently posed by Yang, Tan and Hsu, and by Fu. Furthermore, we extend known results, obtained by Kim and Park, for the case when n = 2.2. We give precise solutions to problems posed by Wang, An, Pan, Wang and Qu and by Hsieh, Lin and Huang. In particular, we show that Q(^k_n) is bi-panconnected and edge-bipancyclic, when k ≥ 3 and n ≥ 2, and we also show that when k is odd, Q(^k_n) is m-panconnected, for m = (^n(k - 1) + 2k - 6’ / ‘_2), and (k -1) pancyclic (these bounds are optimal). We introduce a path-shortening technique, called progressive shortening, and strengthen existing results, showing that when paths are formed using progressive shortening then these paths can be efficiently constructed and used to solve a problem relating to the distributed simulation of linear arrays and cycles in a parallel machine whose interconnection network is Q(^k_n) even in the presence of a faulty processor.3. We define an interconnection network AQ(^k_n) which we call the augmented k-ary n-cube by extending a k-ary n-cube in a manner analogous to the existing extension of an n-dimensional hypercube to an n-dimensional augmented cube. We prove that the augmented k-ary n-cube Q(^k_n) has a number of attractive properties (in the context of parallel computing). For example, we show that the augmented k-ary n-cube Q(^k_n) - is a Cayley graph (and so is vertex-symmetric); has connectivity 4n - 2, and is such that we can build a set of 4n - 2 mutually disjoint paths joining any two distinct vertices so that the path of maximal length has length at most max{{n- l)k- (n-2), k + 7}; has diameter [(^k) / (_3)] + [(^k - 1) /( _3)], when n = 2; and has diameter at most (^k) / (_4) (n+ 1), for n ≥ 3 and k even, and at most [(^k)/ (_4) (n + 1) + (^n) / (_4), for n ^, for n ≥ 3 and k odd.4. We present an algorithm which given a source node and a set of n - 1 target nodes in the (n, k)-star graph S(_n,k) where all nodes are distinct, builds a collection of n - 1 node-disjoint paths, one from each target node to the source. The collection of paths output from the algorithm is such that each path has length at most 6k - 7, and the algorithm has time complexity O(k(^3)n(^4))
Quasiconvex Programming
We define quasiconvex programming, a form of generalized linear programming
in which one seeks the point minimizing the pointwise maximum of a collection
of quasiconvex functions. We survey algorithms for solving quasiconvex programs
either numerically or via generalizations of the dual simplex method from
linear programming, and describe varied applications of this geometric
optimization technique in meshing, scientific computation, information
visualization, automated algorithm analysis, and robust statistics.Comment: 33 pages, 14 figure
Hypercube-Based Topologies With Incremental Link Redundancy.
Hypercube structures have received a great deal of attention due to the attractive properties inherent to their topology. Parallel algorithms targeted at this topology can be partitioned into many tasks, each of which running on one node processor. A high degree of performance is achievable by running every task individually and concurrently on each node processor available in the hypercube. Nevertheless, the performance can be greatly degraded if the node processors spend much time just communicating with one another. The goal in designing hypercubes is, therefore, to achieve a high ratio of computation time to communication time. The dissertation addresses primarily ways to enhance system performance by minimizing the communication time among processors. The need for improving the performance of hypercube networks is clearly explained. Three novel topologies related to hypercubes with improved performance are proposed and analyzed. Firstly, the Bridged Hypercube (BHC) is introduced. It is shown that this design is remarkably more efficient and cost-effective than the standard hypercube due to its low diameter. Basic routing algorithms such as one to one and broadcasting are developed for the BHC and proven optimal. Shortcomings of the BHC such as its asymmetry and limited application are clearly discussed. The Folded Hypercube (FHC), a symmetric network with low diameter and low degree of the node, is introduced. This new topology is shown to support highly efficient communications among the processors. For the FHC, optimal routing algorithms are developed and proven to be remarkably more efficient than those of the conventional hypercube. For both BHC and FHC, network parameters such as average distance, message traffic density, and communication delay are derived and comparatively analyzed. Lastly, to enhance the fault tolerance of the hypercube, a new design called Fault Tolerant Hypercube (FTH) is proposed. The FTH is shown to exhibit a graceful degradation in performance with the existence of faults. Probabilistic models based on Markov chain are employed to characterize the fault tolerance of the FTH. The results are verified by Monte Carlo simulation. The most attractive feature of all new topologies is the asymptotically zero overhead associated with them. The designs are simple and implementable. These designs can lead themselves to many parallel processing applications requiring high degree of performance
Efficient Interconnection Schemes for VLSI and Parallel Computation
This thesis is primarily concerned with two problems of interconnecting components in VLSI technologies. In the first case, the goal is to construct efficient interconnection networks for general-purpose parallel computers. The second problem is a more specialized problem in the design of VLSI chips, namely multilayer channel routing. In addition, a final part of this thesis provides lower bounds on the area required for VLSI implementations of finite-state machines. This thesis shows that networks based on Leiserson\u27s fat-tree architecture are nearly as good as any network built in a comparable amount of physical space. It shows that these universal networks can efficiently simulate competing networks by means of an appropriate correspondence between network components and efficient algorithms for routing messages on the universal network. In particular, a universal network of area A can simulate competing networks with O(lg^3A) slowdown (in bit-times), using a very simple randomized routing algorithm and simple network components. Alternatively, a packet routing scheme of Leighton, Maggs, and Rao can be used in conjunction with more sophisticated switching components to achieve O(lg^2 A) slowdown. Several other important aspects of universality are also discussed. It is shown that universal networks can be constructed in area linear in the number of processors, so that there is no need to restrict the density of processors in competing networks. Also results are presented for comparisons between networks of different size or with processors of different sizes (as determined by the amount of attached memory). Of particular interest is the fact that a universal network built from sufficiently small processors can simulate (with the slowdown already quoted) any competing network of comparable size regardless of the size of processors in the competing network. In addition, many of the results given do not require the usual assumption of unit wire delay. Finally, though most of the discussion is in the two-dimensional world, the results are shown to apply in three dimensions by way of a simple demonstration of general results on graph layout in three dimensions. The second main problem considered in this thesis is channel routing when many layers of interconnect are available, a scenario that is becoming more and more meaningful as chip fabrication technologies advance. This thesis describes a system MulCh for multilayer channel routing which extends the Chameleon system developed at U. C. Berkeley. Like Chameleon, MulCh divides a multilayer problem into essentially independent subproblems of at most three layers, but unlike Chameleon, MulCh considers the possibility of using partitions comprised of a single layer instead of only partitions of two or three layers. Experimental results show that MulCh often performs better than Chameleon in terms of channel width, total net length, and number of vias. In addition to a description of MulCh as implemented, this thesis provides improved algorithms for subtasks performed by MulCh, thereby indicating potential improvements in the speed and performance of multilayer channel routing. In particular, a linear time algorithm is given for determining the minimum width required for a single-layer channel routing problem, and an algorithm is given for maintaining the density of a collection of nets in logarithmic time per net insertion. The last part of this thesis shows that straightforward techniques for implementing finite-state machines are optimal in the worst case. Specifically, for any s and k, there is a deterministic finite-state machine with s states and k symbols such that any layout algorithm requires (ks lg s) area to lay out its realization. For nondeterministic machines, there is an analogous lower bound of (ks^2) area
Minkowski Sum Construction and other Applications of Arrangements of Geodesic Arcs on the Sphere
We present two exact implementations of efficient output-sensitive algorithms
that compute Minkowski sums of two convex polyhedra in 3D. We do not assume
general position. Namely, we handle degenerate input, and produce exact
results. We provide a tight bound on the exact maximum complexity of Minkowski
sums of polytopes in 3D in terms of the number of facets of the summand
polytopes. The algorithms employ variants of a data structure that represents
arrangements embedded on two-dimensional parametric surfaces in 3D, and they
make use of many operations applied to arrangements in these representations.
We have developed software components that support the arrangement
data-structure variants and the operations applied to them. These software
components are generic, as they can be instantiated with any number type.
However, our algorithms require only (exact) rational arithmetic. These
software components together with exact rational-arithmetic enable a robust,
efficient, and elegant implementation of the Minkowski-sum constructions and
the related applications. These software components are provided through a
package of the Computational Geometry Algorithm Library (CGAL) called
Arrangement_on_surface_2. We also present exact implementations of other
applications that exploit arrangements of arcs of great circles embedded on the
sphere. We use them as basic blocks in an exact implementation of an efficient
algorithm that partitions an assembly of polyhedra in 3D with two hands using
infinite translations. This application distinctly shows the importance of
exact computation, as imprecise computation might result with dismissal of
valid partitioning-motions.Comment: A Ph.D. thesis carried out at the Tel-Aviv university. 134 pages
long. The advisor was Prof. Dan Halperi
HERMESH : a geometrical domain composition method in computational mechanics
With this thesis we present the HERMESH method which has been classified by us as a a composition domain method. This term comes from the idea that HERMESH obtains a global solution of the problem from two independent meshes as a result of the mesh coupling. The global mesh maintains the same number of degrees of freedom as the sum of the independent meshes, which are coupled in the interfaces via new elements referred to by us as extension elements. For this reason we enunciate that the domain composition method is geometrical. The result of the global mesh is a non-conforming mesh in the interfaces between independent meshes due to these new connectivities formed with existing nodes and represented by the new extension elements.
The first requirements were that the method be implicit, be valid for any partial differential equation and not imply any additional effort or loss in efficiency in the parallel performance of the code in which the method has been implemented. In our opinion, these properties constitute the main contribution in mesh coupling for the computational mechanics framework.
From these requirements, we have been able to develop an automatic and topology-independent tool to compose independent meshes. The method can couple overlapping meshes with minimal intervention on the user's part. The overlapping can be partial or complete in the sense of overset meshes. The meshes can be disjoint with or without a gap between them. And we have demonstrated the flexibility of the method in the relative mesh size.
In this work we present a detailed description of HERMESH which has been implemented in a high-performance computing computational mechanics code within the framework of the finite element methods. This code is called Alya. The numerical properties will be proved with different benchmark-type problems and the manufactured solution technique.
Finally, the results in complex problems solved with HERMESH will be presented, clearly showing the versatility of the method.En este trabajo presentamos el metodo HERMESH al que hemos catalogado como un método de composición de dominios puesto que a partir de mallas independientes se obtiene una solución global del problema como la unión de los subproblemas que forman las mallas independientes. Como resultado, la malla global mantiene el mismo número de grados de libertad que la suma de los grados de libertad de las mallas independientes, las cuales se acoplan en las interfases internas a través de nuevos elementos a los que nos referimos como elementos de extensión. Por este motivo decimos que el método de composición de dominio es geométrico. El resultado de la malla global es una malla que
no es conforme en las interfases entre las distintas mallas debido a las nuevas conectividades generadas sobre los nodos existentes.
Los requerimientos de partida fueron que el método se implemente de forma implÃcita, sea válido para cualquier PDE y no implique ningún esfuerzo
adicional ni perdida de eficiencia para el funcionamiento paralelo del código de altas prestaciones en el que ha sido implementado. Creemos que estas propiedades son las principales aportaciones de esta tesis dentro del marco de acoplamiento de mallas en mecánica computacional.
A partir de estas premisas, hemos conseguido una herramienta automática e independiente de la topologÃa para componer mallas. Es capaz de acoplar sin necesidad de intervención del usuario, mallas con solapamiento parcial o total asà como mallas disjuntas con o sin "gap" entre ellas. También hemos visto que ofrece cierta flexibilidad en relación al tamaños relativos entre las mallas siendo un método válido como técnica de remallado local.
Presentamos una descripción detallada de la implementación de esta técnica, llevada a cabo en un código de altas prestaciones de mecánica computacional en el contexto de elementos finitos, Alya. Se demostrarán todas las propiedades numéricas que ofrece el métodos a través de distintos problemas tipo benchmark y el método de la solución manufacturada.
Finalmente se mostrarán los resultados en problemas complejos resueltos con el método HERMESH, que a su vez es una prueba de la gran flexibilidad que nos brinda
LIPIcs, Volume 258, SoCG 2023, Complete Volume
LIPIcs, Volume 258, SoCG 2023, Complete Volum
Scan-based immersed isogeometric analysis
Scan-based simulations contain innate topologically complex three-dimensional geometries, represented by large data sets in formats which are not directly suitable for analysis. Consequently, performing high-fidelity scan-based simulations at practical computational costs is still very challenging. The main objective of this dissertation is to develop an efficient and robust scan-based simulation strategy by acquiring a profound understanding of three prominent challenges in scan-based IGA, viz.: i) balancing the accuracy and computational effort associated with numerical integration; ii) the preservation of topology in the spline-based segmentation procedure; and iii) the control of accuracy using error estimation and adaptivity techniques.
In three-dimensional immersed isogeometric simulations, the computational effort associated with integration can be the critical component. A myriad of integration strategies has been proposed over the past years to ameliorate the difficulties associated with integration, but a general optimal integration framework that suits a broad class of engineering problems is not yet available. In this dissertation we provide a thorough investigation of the accuracy and computational effort of the octree integration technique. We quantify the contribution of the integration error using the theoretical basis provided by Strang’s first lemma. Based on this study we propose an error-estimate-based adaptive integration procedure for immersed IGA.
To exploit the advantageous properties of IGA in a scan-based setting, it is important to extract a smooth geometry. This can be established by convoluting the voxel data using B-splines, but this can induce problematic topological changes when features with a size similar to that of the voxels are encountered. This dissertation presents a topology-preserving segmentation procedure using truncated hierarchical (TH)B-splines. A moving-window-based topological anomaly detection algorithm is proposed to identify regions in which (TH)B-spline refinements must be performed. The criterion to identify topological anomalies is based on the Euler characteristic, giving it the capability to distinguish between topological and shape changes. A Fourier analysis is presented to explain the effectiveness of the developed procedure.
An additional computational challenge in the context of immersed IGA is the construction of optimal approximations using locally refined splines. For scan-based volumetric domains, hierarchical splines are particularly suitable, as they optimally leverage the advantages offered by the availability of a geometrically simple background mesh. Although truncated hierarchical B-splines have been successfully applied in the context of IGA, their application in the immersed setting is largely unexplored. In this dissertation we propose a computational strategy for the application of error estimation-based mesh adaptivity for stabilized immersed IGA.
The conducted analyses and developed computational techniques for scan-based immersed IGA are interrelated, and together constitute a significant improvement in the efficiency and robustness of the analysis paradigm. In combination with other state-of-the-art developments regarding immersed FEM/IGA (\emph{e.g.}, iterative solution techniques, parallel computing), the research in this thesis opens the doors to scan-based simulations with more sophisticated physical behavior, geometries of increased complexity, and larger scan-data sizes.Scan-based simulations contain innate topologically complex three-dimensional geometries, represented by large data sets in formats which are not directly suitable for analysis. Consequently, performing high-fidelity scan-based simulations at practical computational costs is still very challenging. The main objective of this dissertation is to develop an efficient and robust scan-based simulation strategy by acquiring a profound understanding of three prominent challenges in scan-based IGA, viz.: i) balancing the accuracy and computational effort associated with numerical integration; ii) the preservation of topology in the spline-based segmentation procedure; and iii) the control of accuracy using error estimation and adaptivity techniques.
In three-dimensional immersed isogeometric simulations, the computational effort associated with integration can be the critical component. A myriad of integration strategies has been proposed over the past years to ameliorate the difficulties associated with integration, but a general optimal integration framework that suits a broad class of engineering problems is not yet available. In this dissertation we provide a thorough investigation of the accuracy and computational effort of the octree integration technique. We quantify the contribution of the integration error using the theoretical basis provided by Strang’s first lemma. Based on this study we propose an error-estimate-based adaptive integration procedure for immersed IGA.
To exploit the advantageous properties of IGA in a scan-based setting, it is important to extract a smooth geometry. This can be established by convoluting the voxel data using B-splines, but this can induce problematic topological changes when features with a size similar to that of the voxels are encountered. This dissertation presents a topology-preserving segmentation procedure using truncated hierarchical (TH)B-splines. A moving-window-based topological anomaly detection algorithm is proposed to identify regions in which (TH)B-spline refinements must be performed. The criterion to identify topological anomalies is based on the Euler characteristic, giving it the capability to distinguish between topological and shape changes. A Fourier analysis is presented to explain the effectiveness of the developed procedure.
An additional computational challenge in the context of immersed IGA is the construction of optimal approximations using locally refined splines. For scan-based volumetric domains, hierarchical splines are particularly suitable, as they optimally leverage the advantages offered by the availability of a geometrically simple background mesh. Although truncated hierarchical B-splines have been successfully applied in the context of IGA, their application in the immersed setting is largely unexplored. In this dissertation we propose a computational strategy for the application of error estimation-based mesh adaptivity for stabilized immersed IGA.
The conducted analyses and developed computational techniques for scan-based immersed IGA are interrelated, and together constitute a significant improvement in the efficiency and robustness of the analysis paradigm. In combination with other state-of-the-art developments regarding immersed FEM/IGA (\emph{e.g.}, iterative solution techniques, parallel computing), the research in this thesis opens the doors to scan-based simulations with more sophisticated physical behavior, geometries of increased complexity, and larger scan-data sizes
LEGO : linear embedding via Green's operators
Reduction of lead time has long been an important target in product development. Owing to the advance of computer power product optimization has been moved from the production stage to the preceding design stage. In particular, the full electromagnetic behavior of the final product can now be predicted through numerical methods. However, for the tuning of device parameters in the optimization stage, commercial software packages often rely on brute-force parameter sweeps. Further, for each set of parameter values a full recomputation of the entire configuration is usually required. In case of stringent product specifications or large complex structures, the computational burden may become severe. Recently, "marching on in anything" has been introduced to accelerate parameter sweeps. Nevertheless, it remains necessary to further reduce the computational costs of electromagnetic device design. This is the main goal in this thesis. As an alternative to existing electromagnetic modeling methods, we propose a modular modeling technique called linear embedding via Green’s operators (LEGO). It is a so-called diakoptic method based on the Huygens principle, involving equivalent boundary current sources by which simply connected scattering domains of arbitrary shape may fully be characterized. Mathematically this may be achieved using either Love’s or Schelkunoff’s equivalence principles, LEP or SEP, respectively. LEGO may be considered as the electromagnetic generalization of decomposing an electric circuit into a system of multi-port subsystems. We have captured the pertaining equivalent current distributions in terms of a lucid Green’s operator formalism. For instance, our scattering operator expresses the equivalent sources that would produce the scattered field exterior to a scattering domain in terms of the equivalent sources that would produce the incident field inside that domain. The enclosed scattering objects may be of arbitrary shape and composition. The scattering domains together with their scattering operators constitute the LEGO building blocks. We have employed various alternative electromagnetic solution methods to construct the scattering operators. In its most elementary form, LEGO is a generalization of an embedding procedure introduced in inverse scattering to describe multiple scattering be tween adjacent blocks, by considering one of the blocks as the environment of the other and vice versa. To establish an interaction between current distributions on disjoint domain boundaries we define a source transfer operator. Through such transfer operators we obtain a closed loop that connects the scattering operators of both domains, which describes the total field including the multiple scattering. Subsequently, a combined scattering block is composed by merging the separate scattering operators via transfer operators, and removing common boundaries. We have validated the LEGO approach for both 2D and 3D configurations. In the field of electromagnetic bandgap (EBG) structures we have demonstrated that a cascade of embedding steps can be employed to form electromagnetically large complex composite blocks. LEGO is a modular method, in that previously combined blocks may be stored in a database for possible reuse in subsequent LEGO building step. Besides scattering operators that account for the exterior scattered field, we also use interior field operators by which the field may be reproduced within (sub)domains that have been combined at an earlier stage. Only the subdomains of interest are stored and updated to account for the presence of additional domains added in subsequent steps. We have also shown how the scattering operator can be utilized to compute the band diagram of EBG structures. Two alternative methods have been proposed to solve the pertaining eigenvalue problem. We have validated the results via a comparison with results from a plane-wave method for 2D EBG structures. In addition, we have demonstrated that our method also applies to unit cells containing scattering objects that are perfectly conducting or extend across the boundary of the unit cell. The optimization stage of a design process often involves tuning local medium properties. In LEGO we accommodated for this through a transfer of the equivalent sources on the boundary of a large scattering operator to the boundary of a relatively small designated domain in which local structure variations are to be tested. As a result, subsequent LEGO steps can be carried out with great efficiency. As demonstrators, we have locally tuned the transmission properties at the Y-junction of both a power splitter and a mode splitter in EBG waveguide technology. In these design examples the computational advantageous of the LEGO approach become clearly manifest, as computation times reduce from hours to minutes. This efficient optimization stage of the LEGO method may also be integrated with existing software packages as an additional design tool. In addition to the acceleration of the computations, the reusability of the composite building constitute an important advantage. The Green’s operators are expressed in terms of equivalent boundary currents. These operators have been obtained using integral equations. In the numerical implementation of the LEGO method we have discretized the operators via the method of moments with a flat-facetted mesh using local test and expansion functions for the fields and currents, respectively. In the 2D case we have investigated the influence of using piecewise constant and piecewise linear functions. For the 3D implementation, we have applied the Rao-Wilton-Glisson (RWG) functions in combination with rotated RWG functions. After discretization, operators and operator compositions are matrices and matrix multiplications, respectively. Since the matrix multiplications in a LEGO step dominate the computational costs, we aim at a maximum accuracy of the field for a minimum mesh density. For LEGO with SEP, we have determined the unknown currents through inverse field propagators, whereas with LEP, the currents are directly obtained from the tangential field components via inverse Gram matrices. After a careful assessment of the computational costs of the LEGO method, it turns out that owing to the removal of common boundaries and the reusability of scattering domains, the most efficient application of LEGO involves a closely-packed configuration of identical blocks. In terms of the number of array elements, N, the complexity of a sequence of LEGO steps for 2D and 3D applications increases as O(N1.5) and O(N2), respectively. We have discussed possible improvements that can be expected from "marching on in anything" or multi-level fast-multipole algorithms. From an evaluation of the resulting scattered field, it turns out that LEGO with SEP is more accurate than with LEP. However, the spurious interior resonance effect common to SEP in the construction of composite building blocks can not simply be avoided through a combined field integral equation. By contrast, LEGO based on LEP is robust. Further, we have demonstrated that additional errors due to the choice of domain shape or building sequence, or the accumulation of errors due to long LEGO sequences are negligible. Further, we have investigated integral equations for the scattering from 2D and 3D perfectly conducting and dielectric objects. The discretized integral operators directly apply to the LEGO method. For scattering objects that are not canonical, these integral equations are used in the construction of the elementary LEGO blocks. Since we aim at a maximum accuracy of the field for a minimum mesh density, the regular test and expansion integral parts are primarily determined through adaptive quadrature rules, while analytic expressions are used for the singular integral parts. It turns out that the convergence of the scattered field is a direct measure for the accuracy of the scattered field computed with LEGO based on SEP or LEP. As an alternative to the PMCHW and the M¨uller integral equations, we have proposed an new integral equation formulation, which leads to cubic convergence in the 2D case, irrespective of the mesh density and object shape. In case of scattering object with a regular boundary domain scaling may be used to improve the convergence rate of the scattered field
- …