160,489 research outputs found
Boosting the Efficiency of Parametric Detection with Hierarchical Neural Networks
Gravitational wave astronomy is a vibrant field that leverages both classic
and modern data processing techniques for the understanding of the universe.
Various approaches have been proposed for improving the efficiency of the
detection scheme, with hierarchical matched filtering being an important
strategy. Meanwhile, deep learning methods have recently demonstrated both
consistency with matched filtering methods and remarkable statistical
performance. In this work, we propose Hierarchical Detection Network (HDN), a
novel approach to efficient detection that combines ideas from hierarchical
matching and deep learning. The network is trained using a novel loss function,
which encodes simultaneously the goals of statistical accuracy and efficiency.
We discuss the source of complexity reduction of the proposed model, and
describe a general recipe for initialization with each layer specializing in
different regions. We demonstrate the performance of HDN with experiments using
open LIGO data and synthetic injections, and observe with two-layer models a
efficiency gain compared with matched filtering at an equal error rate
of . Furthermore, we show how training a three-layer HDN initialized
using two-layer model can further boost both accuracy and efficiency,
highlighting the power of multiple simple layers in efficient detection
Bayesian sparsification for deep neural networks with Bayesian model reduction
Deep learning's immense capabilities are often constrained by the complexity
of its models, leading to an increasing demand for effective sparsification
techniques. Bayesian sparsification for deep learning emerges as a crucial
approach, facilitating the design of models that are both computationally
efficient and competitive in terms of performance across various deep learning
applications. The state-of-the-art -- in Bayesian sparsification of deep neural
networks -- combines structural shrinkage priors on model weights with an
approximate inference scheme based on stochastic variational inference.
However, model inversion of the full generative model is exceptionally
computationally demanding, especially when compared to standard deep learning
of point estimates. In this context, we advocate for the use of Bayesian model
reduction (BMR) as a more efficient alternative for pruning of model weights.
As a generalization of the Savage-Dickey ratio, BMR allows a post-hoc
elimination of redundant model weights based on the posterior estimates under a
straightforward (non-hierarchical) generative model. Our comparative study
highlights the advantages of the BMR method relative to established approaches
based on hierarchical horseshoe priors over model weights. We illustrate the
potential of BMR across various deep learning architectures, from classical
networks like LeNet to modern frameworks such as Vision Transformers and
MLP-Mixers
Simulation-time reduction techniques for a retrofit planning tool
The design of retrofitted energy efficient buildings is
a promising option towards achieving a cost-effective
improvement of the overall building sector’s energy
performance. With the aim of discovering the best design
for a retrofitting project in an automatic manner,
a decision making (or optimization) process is usually
adopted, utilizing accurate building simulation
models towards evaluating the candidate retrofitting
scenarios. A major factor which affects the overall
computational time of such a process is the simulation
execution time. Since high complexity and prohibitive
simulation execution time are predominantly
due to the full-scale, detailed simulation, in this work,
the following simulation-time reduction methodologies
are evaluated with respect to accuracy and computational
effort in a test building: Hierarchical clustering;
Koopman modes; and Meta-models. The simplified
model that would be the outcome of these
approaches, can be utilized by any optimization approach
to discover the best retrofitting option
Tensor Computation: A New Framework for High-Dimensional Problems in EDA
Many critical EDA problems suffer from the curse of dimensionality, i.e. the
very fast-scaling computational burden produced by large number of parameters
and/or unknown variables. This phenomenon may be caused by multiple spatial or
temporal factors (e.g. 3-D field solvers discretizations and multi-rate circuit
simulation), nonlinearity of devices and circuits, large number of design or
optimization parameters (e.g. full-chip routing/placement and circuit sizing),
or extensive process variations (e.g. variability/reliability analysis and
design for manufacturability). The computational challenges generated by such
high dimensional problems are generally hard to handle efficiently with
traditional EDA core algorithms that are based on matrix and vector
computation. This paper presents "tensor computation" as an alternative general
framework for the development of efficient EDA algorithms and tools. A tensor
is a high-dimensional generalization of a matrix and a vector, and is a natural
choice for both storing and solving efficiently high-dimensional EDA problems.
This paper gives a basic tutorial on tensors, demonstrates some recent examples
of EDA applications (e.g., nonlinear circuit modeling and high-dimensional
uncertainty quantification), and suggests further open EDA problems where the
use of tensor computation could be of advantage.Comment: 14 figures. Accepted by IEEE Trans. CAD of Integrated Circuits and
System
Truncated Variational EM for Semi-Supervised Neural Simpletrons
Inference and learning for probabilistic generative networks is often very
challenging and typically prevents scalability to as large networks as used for
deep discriminative approaches. To obtain efficiently trainable, large-scale
and well performing generative networks for semi-supervised learning, we here
combine two recent developments: a neural network reformulation of hierarchical
Poisson mixtures (Neural Simpletrons), and a novel truncated variational EM
approach (TV-EM). TV-EM provides theoretical guarantees for learning in
generative networks, and its application to Neural Simpletrons results in
particularly compact, yet approximately optimal, modifications of learning
equations. If applied to standard benchmarks, we empirically find, that
learning converges in fewer EM iterations, that the complexity per EM iteration
is reduced, and that final likelihood values are higher on average. For the
task of classification on data sets with few labels, learning improvements
result in consistently lower error rates if compared to applications without
truncation. Experiments on the MNIST data set herein allow for comparison to
standard and state-of-the-art models in the semi-supervised setting. Further
experiments on the NIST SD19 data set show the scalability of the approach when
a manifold of additional unlabeled data is available
An error-controlled methodology for approximate hierarchical symbolic analysis
Limitations of existing approaches for symbolic analysis of large analog circuits are discussed. To address their solution, a new methodology for hierarchical symbolic analysis is introduced. The combination of a hierarchical modeling technique and approximation strategies, comprising circuit reduction, graph-based symbolic solution of circuit equations and matrix-based error control, provides optimum results in terms of speech and quality of results.European Commission ESPRIT 21812Comisión Interministerial de Ciencia y Tecnología TIC97-058
- …