4,223 research outputs found
Reduced order modeling of fluid flows: Machine learning, Kolmogorov barrier, closure modeling, and partitioning
In this paper, we put forth a long short-term memory (LSTM) nudging framework
for the enhancement of reduced order models (ROMs) of fluid flows utilizing
noisy measurements. We build on the fact that in a realistic application, there
are uncertainties in initial conditions, boundary conditions, model parameters,
and/or field measurements. Moreover, conventional nonlinear ROMs based on
Galerkin projection (GROMs) suffer from imperfection and solution instabilities
due to the modal truncation, especially for advection-dominated flows with slow
decay in the Kolmogorov width. In the presented LSTM-Nudge approach, we fuse
forecasts from a combination of imperfect GROM and uncertain state estimates,
with sparse Eulerian sensor measurements to provide more reliable predictions
in a dynamical data assimilation framework. We illustrate the idea with the
viscous Burgers problem, as a benchmark test bed with quadratic nonlinearity
and Laplacian dissipation. We investigate the effects of measurements noise and
state estimate uncertainty on the performance of the LSTM-Nudge behavior. We
also demonstrate that it can sufficiently handle different levels of temporal
and spatial measurement sparsity. This first step in our assessment of the
proposed model shows that the LSTM nudging could represent a viable realtime
predictive tool in emerging digital twin systems
Sparsity in Dynamics of Spontaneous Subtle Emotions: Analysis \& Application
Spontaneous subtle emotions are expressed through micro-expressions, which
are tiny, sudden and short-lived dynamics of facial muscles; thus poses a great
challenge for visual recognition. The abrupt but significant dynamics for the
recognition task are temporally sparse while the rest, irrelevant dynamics, are
temporally redundant. In this work, we analyze and enforce sparsity constrains
to learn significant temporal and spectral structures while eliminate
irrelevant facial dynamics of micro-expressions, which would ease the challenge
in the visual recognition of spontaneous subtle emotions. The hypothesis is
confirmed through experimental results of automatic spontaneous subtle emotion
recognition with several sparsity levels on CASME II and SMIC, the only two
publicly available spontaneous subtle emotion databases. The overall
performances of the automatic subtle emotion recognition are boosted when only
significant dynamics are preserved from the original sequences.Comment: IEEE Transaction of Affective Computing (2016
Cortex, countercurrent context, and dimensional integration of lifetime memory
The correlation between relative neocortex size and longevity in mammals encourages a search for a cortical function specifically related to the life-span. A candidate in the domain of permanent and cumulative memory storage is proposed and explored in relation to basic aspects of cortical organization. The pattern of cortico-cortical connectivity between functionally specialized areas and the laminar organization of that connectivity converges on a globally coherent representational space in which contextual embedding of information emerges as an obligatory feature of cortical function. This brings a powerful mode of inductive knowledge within reach of mammalian adaptations, a mode which combines item specificity with classificatory generality. Its neural implementation is proposed to depend on an obligatory interaction between the oppositely directed feedforward and feedback currents of cortical activity, in countercurrent fashion. Direct interaction of the two streams along their cortex-wide local interface supports a scheme of "contextual capture" for information storage responsible for the lifelong cumulative growth of a uniquely cortical form of memory termed "personal history." This approach to cortical function helps elucidate key features of cortical organization as well as cognitive aspects of mammalian life history strategies
A Compressed Sampling and Dictionary Learning Framework for WDM-Based Distributed Fiber Sensing
We propose a compressed sampling and dictionary learning framework for
fiber-optic sensing using wavelength-tunable lasers. A redundant dictionary is
generated from a model for the reflected sensor signal. Imperfect prior
knowledge is considered in terms of uncertain local and global parameters. To
estimate a sparse representation and the dictionary parameters, we present an
alternating minimization algorithm that is equipped with a pre-processing
routine to handle dictionary coherence. The support of the obtained sparse
signal indicates the reflection delays, which can be used to measure
impairments along the sensing fiber. The performance is evaluated by
simulations and experimental data for a fiber sensor system with common core
architecture.Comment: Accepted for publication in Journal of the Optical Society of America
A [ \copyright\ 2017 Optical Society of America.]. One print or electronic
copy may be made for personal use only. Systematic reproduction and
distribution, duplication of any material in this paper for a fee or for
commercial purposes, or modifications of the content of this paper are
prohibite
An efficient surrogate model for emulation and physics extraction of large eddy simulations
In the quest for advanced propulsion and power-generation systems,
high-fidelity simulations are too computationally expensive to survey the
desired design space, and a new design methodology is needed that combines
engineering physics, computer simulations and statistical modeling. In this
paper, we propose a new surrogate model that provides efficient prediction and
uncertainty quantification of turbulent flows in swirl injectors with varying
geometries, devices commonly used in many engineering applications. The novelty
of the proposed method lies in the incorporation of known physical properties
of the fluid flow as {simplifying assumptions} for the statistical model. In
view of the massive simulation data at hand, which is on the order of hundreds
of gigabytes, these assumptions allow for accurate flow predictions in around
an hour of computation time. To contrast, existing flow emulators which forgo
such simplications may require more computation time for training and
prediction than is needed for conducting the simulation itself. Moreover, by
accounting for coupling mechanisms between flow variables, the proposed model
can jointly reduce prediction uncertainty and extract useful flow physics,
which can then be used to guide further investigations.Comment: Submitted to JASA A&C
Substructure and Boundary Modeling for Continuous Action Recognition
This paper introduces a probabilistic graphical model for continuous action
recognition with two novel components: substructure transition model and
discriminative boundary model. The first component encodes the sparse and
global temporal transition prior between action primitives in state-space model
to handle the large spatial-temporal variations within an action class. The
second component enforces the action duration constraint in a discriminative
way to locate the transition boundaries between actions more accurately. The
two components are integrated into a unified graphical structure to enable
effective training and inference. Our comprehensive experimental results on
both public and in-house datasets show that, with the capability to incorporate
additional information that had not been explicitly or efficiently modeled by
previous methods, our proposed algorithm achieved significantly improved
performance for continuous action recognition.Comment: Detailed version of the CVPR 2012 paper. 15 pages, 6 figure
Deep-learning assisted reduced order model for high-dimensional flow prediction from sparse data
The reconstruction and prediction of full-state flows from sparse data are of
great scientific and engineering significance yet remain challenging,
especially in applications where data are sparse and/or subjected to noise. To
this end, this study proposes a deep-learning assisted non-intrusive reduced
order model (named DCDMD) for high-dimensional flow prediction from sparse
data. Based on the compressed sensing (CS)-Dynamic Mode Decomposition (DMD),
the DCDMD model is distinguished by two novelties. Firstly, a sparse matrix is
defined to overcome the strict random distribution condition of sensor
locations in CS, thus allowing flexible sensor deployments and requiring very
few sensors. Secondly, a deep-learning-based proxy is invoked to acquire
coherent flow modes from the sparse data of high-dimensional flows, thereby
addressing the issue of defining sparsity and the stringent incoherence
condition in the conventional CSDMD. The two advantageous features, combined
with the fact that the model retains flow physics in the online stage, lead to
significant enhancements in accuracy and efficiency, as well as superior
insensitivity to data noises (i.e., robustness), in both reconstruction and
prediction of full-state flows. These are demonstrated by three benchmark
examples, i.e., cylinder wake, weekly-mean sea surface temperature and
isotropic turbulence in a periodic square area.Comment: 36 Pages, 23 Figures, 5 Table
A Survey of Location Prediction on Twitter
Locations, e.g., countries, states, cities, and point-of-interests, are
central to news, emergency events, and people's daily lives. Automatic
identification of locations associated with or mentioned in documents has been
explored for decades. As one of the most popular online social network
platforms, Twitter has attracted a large number of users who send millions of
tweets on daily basis. Due to the world-wide coverage of its users and
real-time freshness of tweets, location prediction on Twitter has gained
significant attention in recent years. Research efforts are spent on dealing
with new challenges and opportunities brought by the noisy, short, and
context-rich nature of tweets. In this survey, we aim at offering an overall
picture of location prediction on Twitter. Specifically, we concentrate on the
prediction of user home locations, tweet locations, and mentioned locations. We
first define the three tasks and review the evaluation metrics. By summarizing
Twitter network, tweet content, and tweet context as potential inputs, we then
structurally highlight how the problems depend on these inputs. Each dependency
is illustrated by a comprehensive review of the corresponding strategies
adopted in state-of-the-art approaches. In addition, we also briefly review two
related problems, i.e., semantic location prediction and point-of-interest
recommendation. Finally, we list future research directions.Comment: Accepted to TKDE. 30 pages, 1 figur
Nonlinear Hebbian learning as a unifying principle in receptive field formation
The development of sensory receptive fields has been modeled in the past by a
variety of models including normative models such as sparse coding or
independent component analysis and bottom-up models such as spike-timing
dependent plasticity or the Bienenstock-Cooper-Munro model of synaptic
plasticity. Here we show that the above variety of approaches can all be
unified into a single common principle, namely Nonlinear Hebbian Learning. When
Nonlinear Hebbian Learning is applied to natural images, receptive field shapes
were strongly constrained by the input statistics and preprocessing, but
exhibited only modest variation across different choices of nonlinearities in
neuron models or synaptic plasticity rules. Neither overcompleteness nor sparse
network activity are necessary for the development of localized receptive
fields. The analysis of alternative sensory modalities such as auditory models
or V2 development lead to the same conclusions. In all examples, receptive
fields can be predicted a priori by reformulating an abstract model as
nonlinear Hebbian learning. Thus nonlinear Hebbian learning and natural
statistics can account for many aspects of receptive field formation across
models and sensory modalities
Recommended from our members
Computational cytometer based on magnetically modulated coherent imaging and deep learning.
Detecting rare cells within blood has numerous applications in disease diagnostics. Existing rare cell detection techniques are typically hindered by their high cost and low throughput. Here, we present a computational cytometer based on magnetically modulated lensless speckle imaging, which introduces oscillatory motion to the magnetic-bead-conjugated rare cells of interest through a periodic magnetic force and uses lensless time-resolved holographic speckle imaging to rapidly detect the target cells in three dimensions (3D). In addition to using cell-specific antibodies to magnetically label target cells, detection specificity is further enhanced through a deep-learning-based classifier that is based on a densely connected pseudo-3D convolutional neural network (P3D CNN), which automatically detects rare cells of interest based on their spatio-temporal features under a controlled magnetic force. To demonstrate the performance of this technique, we built a high-throughput, compact and cost-effective prototype for detecting MCF7 cancer cells spiked in whole blood samples. Through serial dilution experiments, we quantified the limit of detection (LoD) as 10 cells per millilitre of whole blood, which could be further improved through multiplexing parallel imaging channels within the same instrument. This compact, cost-effective and high-throughput computational cytometer can potentially be used for rare cell detection and quantification in bodily fluids for a variety of biomedical applications
- …