191 research outputs found
Manipulating light at distance by a metasurface using momentum transformation
A momentum conservation approach is introduced to manipulate light at
distance using metasurfaces. Given a specified field existing on one side of
the metasurface and specified desired field transmitted from the opposite side,
a general momentum boundary condition is established, which determines the
amplitude, phase and polarization transformation to be induced by the
metasurface. This approach, named momentum transformation, enables a systematic
way to synthesize metasurfaces with complete control over the reflected and
transmitted fields. Several synthesis illustrative examples are provided: a
vortex hypergeometric-Gaussian beam and a "delayed-start" accelerated beam for
Fresnel region manipulation, and a pencil beam radiator and a holographic
repeater for Frauenhofer region manipulation
AUTOMATING DATA-LAYOUT DECISIONS IN DOMAIN-SPECIFIC LANGUAGES
A long-standing challenge in High-Performance Computing (HPC) is the simultaneous achievement of programmer productivity and hardware computational efficiency. The challenge has been exacerbated by the onset of multi- and many-core CPUs and accelerators. Only a few expert programmers have been able to hand-code domain-specific data transformations and vectorization schemes needed to extract the best possible performance on such architectures. In this research, we examined the possibility of automating these methods by developing a Domain-Specific Language (DSL) framework. Our DSL approach extends C++14 by embedding into it a high-level data-parallel array language, and by using a domain-specific compiler to compile to hybrid-parallel code. We also implemented an array index-space transformation algebra within this high-level array language to manipulate array data-layouts and data-distributions. The compiler introduces a novel method for SIMD auto-vectorization based on array data-layouts. Our new auto-vectorization technique is shown to outperform the default auto-vectorization strategy by up to 40% for stencil computations. The compiler also automates distributed data movement with overlapping of local compute with remote data movement using polyhedral integer set analysis. Along with these main innovations, we developed a new technique using C++ template metaprogramming for developing embedded DSLs using C++. We also proposed a domain-specific compiler intermediate representation that simplifies data flow analysis of abstract DSL constructs. We evaluated our framework by constructing a DSL for the HPC grand-challenge domain of lattice quantum chromodynamics. Our DSL yielded performance gains of up to twice the flop rate over existing production C code for selected kernels. This gain in performance was obtained while using less than one-tenth the lines of code. The performance of this DSL was also competitive with the best hand-optimized and hand-vectorized code, and is an order of magnitude better than existing production DSLs.Doctor of Philosoph
Transformations of High-Level Synthesis Codes for High-Performance Computing
Specialized hardware architectures promise a major step in performance and
energy efficiency over the traditional load/store devices currently employed in
large scale computing systems. The adoption of high-level synthesis (HLS) from
languages such as C/C++ and OpenCL has greatly increased programmer
productivity when designing for such platforms. While this has enabled a wider
audience to target specialized hardware, the optimization principles known from
traditional software design are no longer sufficient to implement
high-performance codes. Fast and efficient codes for reconfigurable platforms
are thus still challenging to design. To alleviate this, we present a set of
optimizing transformations for HLS, targeting scalable and efficient
architectures for high-performance computing (HPC) applications. Our work
provides a toolbox for developers, where we systematically identify classes of
transformations, the characteristics of their effect on the HLS code and the
resulting hardware (e.g., increases data reuse or resource consumption), and
the objectives that each transformation can target (e.g., resolve interface
contention, or increase parallelism). We show how these can be used to
efficiently exploit pipelining, on-chip distributed fast memory, and on-chip
streaming dataflow, allowing for massively parallel architectures. To quantify
the effect of our transformations, we use them to optimize a set of
throughput-oriented FPGA kernels, demonstrating that our enhancements are
sufficient to scale up parallelism within the hardware constraints. With the
transformations covered, we hope to establish a common framework for
performance engineers, compiler developers, and hardware developers, to tap
into the performance potential offered by specialized hardware architectures
using HLS
Data-Driven Process Optimization of Additive Manufacturing Systems
The goal of the present dissertation is to develop and apply novel and systematic data-driven optimization approaches that can efficiently optimize Additive Manufacturing (AM) systems with respect to targeted properties of final parts. The proposed approaches are capable of achieving sets of process parameters that result in the satisfactory level of part quality in an accelerated manner. First, an Accelerated Process Optimization (APO) methodology is developed to optimize an individual scalar property of parts. The APO leverages data from similar—but non-identical—prior studies to accelerate sequential experimentation for optimizing the AM system in the current study. Using Bayesian updating, the APO characterizes and updates the difference between prior and current experimental studies. The APO accounts for the differences in experimental conditions and utilizes prior data to facilitate the optimization procedure in the current study. The efficiency and robustness of the APO is tested against an extensive simulation studies and a real-world case study for optimizing relative density of stainless steel parts fabricated by a Selective Laser Melting (SLM) system. Then, we extend the idea behind the APO in order to handle multi-objective process optimization problems in which some of the characteristics of the AMabricated parts are uncorrelated. The proposed Multi-objective Process Optimization (m-APO) breaks down the master multi-objective optimization problem into a series of convex combinations of single-objective sub-problems. The m-APO maps and scales experimental data from previous sub-problems to guide remaining sub-problems that improve the solutions while reducing the number of experiments required. The robustness and efficiency of the m-APO is verified by conducting a series of challenging simulation studies and a real-world case study to minimize geometric inaccuracy of parts fabricated by a Fused Filament Fabrication () system. At the end, we apply the proposed m-APO to maximize the mechanical properties of AMabricated parts that show conflicting behavior in the optimal window, namely relative density and elongation-toailure. Numerical studies show that the m-APO can achieve the best trade-off among conflicting mechanical properties while significantly reducing the number of experimental runs compared with existing methods
Fitting Analysis using Differential Evolution Optimization (FADO): Spectral population synthesis through genetic optimization under self-consistency boundary conditions
The goal of population spectral synthesis (PSS) is to decipher from the
spectrum of a galaxy the mass, age and metallicity of its constituent stellar
populations. This technique has been established as a fundamental tool in
extragalactic research. It has been extensively applied to large spectroscopic
data sets, notably the SDSS, leading to important insights into the galaxy
assembly history. However, despite significant improvements over the past
decade, all current PSS codes suffer from two major deficiencies that inhibit
us from gaining sharp insights into the star-formation history (SFH) of
galaxies and potentially introduce substantial biases in studies of their
physical properties (e.g., stellar mass, mass-weighted stellar age and specific
star formation rate). These are i) the neglect of nebular emission in spectral
fits, consequently, ii) the lack of a mechanism that ensures consistency
between the best-fitting SFH and the observed nebular emission characteristics
of a star-forming (SF) galaxy. In this article, we present FADO (Fitting
Analysis using Differential evolution Optimization): a conceptually novel,
publicly available PSS tool with the distinctive capability of permitting
identification of the SFH that reproduces the observed nebular characteristics
of a SF galaxy. This so-far unique self-consistency concept allows us to
significantly alleviate degeneracies in current spectral synthesis. The
innovative character of FADO is further augmented by its mathematical
foundation: FADO is the first PSS code employing genetic differential evolution
optimization. This, in conjunction with other unique elements in its
mathematical concept (e.g., optimization of the spectral library using
artificial intelligence, convergence test, quasi-parallelization) results in
key improvements with respect to computational efficiency and uniqueness of the
best-fitting SFHs.Comment: 25 pages, 12 figures, A&A accepte
Multiobjective strategies for New Product Development in the pharmaceutical industry
New Product Development (NPD) constitutes a challenging problem in the pharmaceutical industry, due to the characteristics of the development pipeline. Formally, the NPD problem can be stated as follows: select a set of R&D projects from a pool of candidate projects in order to satisfy several criteria (economic profitability, time to market) while coping with the uncertain nature of the projects. More precisely, the recurrent key issues are to determine the projects to develop once target molecules have been identified, their order and the level of resources to assign. In this context, the proposed approach combines discrete event stochastic simulation (Monte Carlo approach) with multiobjective genetic algorithms (NSGAII type, Non-Sorted Genetic Algorithm II) to optimize the highly combinatorial portfolio management problem. In that context, Genetic Algorithms (GAs) are particularly attractive for treating this kind of problem, due to their ability to directly lead to the so-called Pareto front and to account for the combinatorial aspect. This work is illustrated with a study case involving nine interdependent new product candidates targeting three diseases. An analysis is performed for this test bench on the different pairs of criteria both for the bi- and tricriteria optimization: large portfolios cause resource queues and delays time to launch and are eliminated by the bi- and tricriteria optimization strategy. The optimization strategy is thus interesting to detect the sequence candidates. Time is an important criterion to consider simultaneously with NPV and risk criteria. The order in which drugs are released in the pipeline is of great importance as with scheduling problems
Multiobjective strategies for New Product Development in the pharmaceutical industry
New Product Development (NPD) constitutes a challenging problem in the pharmaceutical industry, due to the characteristics of the development pipeline. Formally, the NPD problem can be stated as follows: select a set of R&D projects from a pool of candidate projects in order to satisfy several criteria (economic profitability, time to market) while coping with the uncertain nature of the projects. More precisely, the recurrent key issues are to determine the projects to develop once target molecules have been identified, their order and the level of resources to assign. In this context, the proposed approach combines discrete event stochastic simulation (Monte Carlo approach) with multiobjective genetic algorithms (NSGAII type, Non-Sorted Genetic Algorithm II) to optimize the highly combinatorial portfolio management problem. In that context, Genetic Algorithms (GAs) are particularly attractive for treating this kind of problem, due to their ability to directly lead to the so-called Pareto front and to account for the combinatorial aspect. This work is illustrated with a study case involving nine interdependent new product candidates targeting three diseases. An analysis is performed for this test bench on the different pairs of criteria both for the bi- and tricriteria optimization: large portfolios cause resource queues and delays time to launch and are eliminated by the bi- and tricriteria optimization strategy. The optimization strategy is thus interesting to detect the sequence candidates. Time is an important criterion to consider simultaneously with NPV and risk criteria. The order in which drugs are released in the pipeline is of great importance as with scheduling problems
- …