42,863 research outputs found
Calibration of Computational Models with Categorical Parameters and Correlated Outputs via Bayesian Smoothing Spline ANOVA
It has become commonplace to use complex computer models to predict outcomes
in regions where data does not exist. Typically these models need to be
calibrated and validated using some experimental data, which often consists of
multiple correlated outcomes. In addition, some of the model parameters may be
categorical in nature, such as a pointer variable to alternate models (or
submodels) for some of the physics of the system. Here we present a general
approach for calibration in such situations where an emulator of the
computationally demanding models and a discrepancy term from the model to
reality are represented within a Bayesian Smoothing Spline (BSS) ANOVA
framework. The BSS-ANOVA framework has several advantages over the traditional
Gaussian Process, including ease of handling categorical inputs and correlated
outputs, and improved computational efficiency. Finally this framework is then
applied to the problem that motivated its design; a calibration of a
computational fluid dynamics model of a bubbling fluidized which is used as an
absorber in a CO2 capture system
Piecewise Latent Variables for Neural Variational Text Processing
Advances in neural variational inference have facilitated the learning of
powerful directed graphical models with continuous latent variables, such as
variational autoencoders. The hope is that such models will learn to represent
rich, multi-modal latent factors in real-world data, such as natural language
text. However, current models often assume simplistic priors on the latent
variables - such as the uni-modal Gaussian distribution - which are incapable
of representing complex latent factors efficiently. To overcome this
restriction, we propose the simple, but highly flexible, piecewise constant
distribution. This distribution has the capacity to represent an exponential
number of modes of a latent target distribution, while remaining mathematically
tractable. Our results demonstrate that incorporating this new latent
distribution into different models yields substantial improvements in natural
language processing tasks such as document modeling and natural language
generation for dialogue.Comment: 19 pages, 2 figures, 8 tables; EMNLP 201
Semantic Image Retrieval via Active Grounding of Visual Situations
We describe a novel architecture for semantic image retrieval---in
particular, retrieval of instances of visual situations. Visual situations are
concepts such as "a boxing match," "walking the dog," "a crowd waiting for a
bus," or "a game of ping-pong," whose instantiations in images are linked more
by their common spatial and semantic structure than by low-level visual
similarity. Given a query situation description, our architecture---called
Situate---learns models capturing the visual features of expected objects as
well the expected spatial configuration of relationships among objects. Given a
new image, Situate uses these models in an attempt to ground (i.e., to create a
bounding box locating) each expected component of the situation in the image
via an active search procedure. Situate uses the resulting grounding to compute
a score indicating the degree to which the new image is judged to contain an
instance of the situation. Such scores can be used to rank images in a
collection as part of a retrieval system. In the preliminary study described
here, we demonstrate the promise of this system by comparing Situate's
performance with that of two baseline methods, as well as with a related
semantic image-retrieval system based on "scene graphs.
- …