115 research outputs found
Kernel Exponential Family Estimation via Doubly Dual Embedding
We investigate penalized maximum log-likelihood estimation for exponential
family distributions whose natural parameter resides in a reproducing kernel
Hilbert space. Key to our approach is a novel technique, doubly dual embedding,
that avoids computation of the partition function. This technique also allows
the development of a flexible sampling strategy that amortizes the cost of
Monte-Carlo sampling in the inference stage. The resulting estimator can be
easily generalized to kernel conditional exponential families. We establish a
connection between kernel exponential family estimation and MMD-GANs, revealing
a new perspective for understanding GANs. Compared to the score matching based
estimators, the proposed method improves both memory and time efficiency while
enjoying stronger statistical properties, such as fully capturing smoothness in
its statistical convergence rate while the score matching estimator appears to
saturate. Finally, we show that the proposed estimator empirically outperforms
state-of-the-artComment: 22 pages, 20 figures; AISTATS 201
Variational Reasoning for Question Answering with Knowledge Graph
Knowledge graph (KG) is known to be helpful for the task of question
answering (QA), since it provides well-structured relational information
between entities, and allows one to further infer indirect facts. However, it
is challenging to build QA systems which can learn to reason over knowledge
graphs based on question-answer pairs alone. First, when people ask questions,
their expressions are noisy (for example, typos in texts, or variations in
pronunciations), which is non-trivial for the QA system to match those
mentioned entities to the knowledge graph. Second, many questions require
multi-hop logic reasoning over the knowledge graph to retrieve the answers. To
address these challenges, we propose a novel and unified deep learning
architecture, and an end-to-end variational learning algorithm which can handle
noise in questions, and learn multi-hop reasoning simultaneously. Our method
achieves state-of-the-art performance on a recent benchmark dataset in the
literature. We also derive a series of new benchmark datasets, including
questions for multi-hop reasoning, questions paraphrased by neural translation
model, and questions in human voice. Our method yields very promising results
on all these challenging datasets
- …
