938 research outputs found
Variational Deep Semantic Hashing for Text Documents
As the amount of textual data has been rapidly increasing over the past
decade, efficient similarity search methods have become a crucial component of
large-scale information retrieval systems. A popular strategy is to represent
original data samples by compact binary codes through hashing. A spectrum of
machine learning methods have been utilized, but they often lack expressiveness
and flexibility in modeling to learn effective representations. The recent
advances of deep learning in a wide range of applications has demonstrated its
capability to learn robust and powerful feature representations for complex
data. Especially, deep generative models naturally combine the expressiveness
of probabilistic generative models with the high capacity of deep neural
networks, which is very suitable for text modeling. However, little work has
leveraged the recent progress in deep learning for text hashing.
In this paper, we propose a series of novel deep document generative models
for text hashing. The first proposed model is unsupervised while the second one
is supervised by utilizing document labels/tags for hashing. The third model
further considers document-specific factors that affect the generation of
words. The probabilistic generative formulation of the proposed models provides
a principled framework for model extension, uncertainty estimation, simulation,
and interpretability. Based on variational inference and reparameterization,
the proposed models can be interpreted as encoder-decoder deep neural networks
and thus they are capable of learning complex nonlinear distributed
representations of the original documents. We conduct a comprehensive set of
experiments on four public testbeds. The experimental results have demonstrated
the effectiveness of the proposed supervised learning models for text hashing.Comment: 11 pages, 4 figure
The Relativistic Hopfield network: rigorous results
The relativistic Hopfield model constitutes a generalization of the standard
Hopfield model that is derived by the formal analogy between the
statistical-mechanic framework embedding neural networks and the Lagrangian
mechanics describing a fictitious single-particle motion in the space of the
tuneable parameters of the network itself. In this analogy the cost-function of
the Hopfield model plays as the standard kinetic-energy term and its related
Mattis overlap (naturally bounded by one) plays as the velocity. The
Hamiltonian of the relativisitc model, once Taylor-expanded, results in a
P-spin series with alternate signs: the attractive contributions enhance the
information-storage capabilities of the network, while the repulsive
contributions allow for an easier unlearning of spurious states, conferring
overall more robustness to the system as a whole. Here we do not deepen the
information processing skills of this generalized Hopfield network, rather we
focus on its statistical mechanical foundation. In particular, relying on
Guerra's interpolation techniques, we prove the existence of the infinite
volume limit for the model free-energy and we give its explicit expression in
terms of the Mattis overlaps. By extremizing the free energy over the latter we
get the generalized self-consistent equations for these overlaps, as well as a
picture of criticality that is further corroborated by a fluctuation analysis.
These findings are in full agreement with the available previous results.Comment: 11 pages, 1 figur
TransNets: Learning to Transform for Recommendation
Recently, deep learning methods have been shown to improve the performance of
recommender systems over traditional methods, especially when review text is
available. For example, a recent model, DeepCoNN, uses neural nets to learn one
latent representation for the text of all reviews written by a target user, and
a second latent representation for the text of all reviews for a target item,
and then combines these latent representations to obtain state-of-the-art
performance on recommendation tasks. We show that (unsurprisingly) much of the
predictive value of review text comes from reviews of the target user for the
target item. We then introduce a way in which this information can be used in
recommendation, even when the target user's review for the target item is not
available. Our model, called TransNets, extends the DeepCoNN model by
introducing an additional latent layer representing the target user-target item
pair. We then regularize this layer, at training time, to be similar to another
latent representation of the target user's review of the target item. We show
that TransNets and extensions of it improve substantially over the previous
state-of-the-art.Comment: Accepted for publication in the 11th ACM Conference on Recommender
Systems (RecSys 2017
Learning the Tangent Space of Dynamical Instabilities from Data
For a large class of dynamical systems, the optimally time-dependent (OTD)
modes, a set of deformable orthonormal tangent vectors that track directions of
instabilities along any trajectory, are known to depend "pointwise" on the
state of the system on the attractor, and not on the history of the trajectory.
We leverage the power of neural networks to learn this "pointwise" mapping from
phase space to OTD space directly from data. The result of the learning process
is a cartography of directions associated with strongest instabilities in phase
space. Implications for data-driven prediction and control of dynamical
instabilities are discussed
The Challenge of Machine Learning in Space Weather Nowcasting and Forecasting
The numerous recent breakthroughs in machine learning (ML) make imperative to
carefully ponder how the scientific community can benefit from a technology
that, although not necessarily new, is today living its golden age. This Grand
Challenge review paper is focused on the present and future role of machine
learning in space weather. The purpose is twofold. On one hand, we will discuss
previous works that use ML for space weather forecasting, focusing in
particular on the few areas that have seen most activity: the forecasting of
geomagnetic indices, of relativistic electrons at geosynchronous orbits, of
solar flares occurrence, of coronal mass ejection propagation time, and of
solar wind speed. On the other hand, this paper serves as a gentle introduction
to the field of machine learning tailored to the space weather community and as
a pointer to a number of open challenges that we believe the community should
undertake in the next decade. The recurring themes throughout the review are
the need to shift our forecasting paradigm to a probabilistic approach focused
on the reliable assessment of uncertainties, and the combination of
physics-based and machine learning approaches, known as gray-box.Comment: under revie
Polypyrimidine tract binding protein functions as a negative regulator of feline calicivirus translation.
Positive strand RNA viruses rely heavily on host cell RNA binding proteins for various aspects of their life cycle. Such proteins interact with sequences usually present at the 5' or 3' extremities of the viral RNA genome, to regulate viral translation and/or replication. We have previously reported that the well characterized host RNA binding protein polypyrimidine tract binding protein (PTB) interacts with the 5'end of the feline calicivirus (FCV) genomic and subgenomic RNAs, playing a role in the FCV life cycle.We have demonstrated that PTB interacts with at least two binding sites within the 5'end of the FCV genome. In vitro translation indicated that PTB may function as a negative regulator of FCV translation and this was subsequently confirmed as the translation of the viral subgenomic RNA in PTB siRNA treated cells was stimulated under conditions in which RNA replication could not occur. We also observed that PTB redistributes from the nucleus to the cytoplasm during FCV infection, partially localizing to viral replication complexes, suggesting that PTB binding may be involved in the switch from translation to replication. Reverse genetics studies demonstrated that synonymous mutations in the PTB binding sites result in a cell-type specific defect in FCV replication.Our data indicates that PTB may function to negatively regulate FCV translation initiation. To reconcile this with efficient virus replication in cells, we propose a putative model for the function of PTB in the FCV life cycle. It is possible that during the early stages of infection, viral RNA is translated in the absence of PTB, however, as the levels of viral proteins increase, the nuclear-cytoplasmic shuttling of PTB is altered, increasing the cytoplasmic levels of PTB, inhibiting viral translation. Whether PTB acts directly to repress translation initiation or via the recruitment of other factors remains to be determined but this may contribute to the stimulation of viral RNA replication via clearance of ribosomes from viral RNA
Geometric deep learning
The goal of these course notes is to describe the main mathematical ideas behind geometric deep learning and to provide implementation details for several applications in shape analysis and synthesis, computer vision and computer graphics. The text in the course materials is primarily based on previously published work. With these notes we gather and provide a clear picture of the key concepts and techniques that fall under the umbrella of geometric deep learning, and illustrate the applications they enable. We also aim to provide practical implementation details for the methods presented in these works, as well as suggest further readings and extensions of these ideas
- …