20 research outputs found

    A Deep Architecture for Semantic Matching with Multiple Positional Sentence Representations

    Full text link
    Matching natural language sentences is central for many applications such as information retrieval and question answering. Existing deep models rely on a single sentence representation or multiple granularity representations for matching. However, such methods cannot well capture the contextualized local information in the matching process. To tackle this problem, we present a new deep architecture to match two sentences with multiple positional sentence representations. Specifically, each positional sentence representation is a sentence representation at this position, generated by a bidirectional long short term memory (Bi-LSTM). The matching score is finally produced by aggregating interactions between these different positional sentence representations, through kk-Max pooling and a multi-layer perceptron. Our model has several advantages: (1) By using Bi-LSTM, rich context of the whole sentence is leveraged to capture the contextualized local information in each positional sentence representation; (2) By matching with multiple positional sentence representations, it is flexible to aggregate different important contextualized local information in a sentence to support the matching; (3) Experiments on different tasks such as question answering and sentence completion demonstrate the superiority of our model.Comment: Accepted by AAAI-201

    A Quantum Many-body Wave Function Inspired Language Modeling Approach

    Full text link
    The recently proposed quantum language model (QLM) aimed at a principled approach to modeling term dependency by applying the quantum probability theory. The latest development for a more effective QLM has adopted word embeddings as a kind of global dependency information and integrated the quantum-inspired idea in a neural network architecture. While these quantum-inspired LMs are theoretically more general and also practically effective, they have two major limitations. First, they have not taken into account the interaction among words with multiple meanings, which is common and important in understanding natural language text. Second, the integration of the quantum-inspired LM with the neural network was mainly for effective training of parameters, yet lacking a theoretical foundation accounting for such integration. To address these two issues, in this paper, we propose a Quantum Many-body Wave Function (QMWF) inspired language modeling approach. The QMWF inspired LM can adopt the tensor product to model the aforesaid interaction among words. It also enables us to reveal the inherent necessity of using Convolutional Neural Network (CNN) in QMWF language modeling. Furthermore, our approach delivers a simple algorithm to represent and match text/sentence pairs. Systematic evaluation shows the effectiveness of the proposed QMWF-LM algorithm, in comparison with the state of the art quantum-inspired LMs and a couple of CNN-based methods, on three typical Question Answering (QA) datasets.Comment: 10 pages,4 figures,CIK

    Review-guided Helpful Answer Identification in E-commerce

    Full text link
    Product-specific community question answering platforms can greatly help address the concerns of potential customers. However, the user-provided answers on such platforms often vary a lot in their qualities. Helpfulness votes from the community can indicate the overall quality of the answer, but they are often missing. Accurately predicting the helpfulness of an answer to a given question and thus identifying helpful answers is becoming a demanding need. Since the helpfulness of an answer depends on multiple perspectives instead of only topical relevance investigated in typical QA tasks, common answer selection algorithms are insufficient for tackling this task. In this paper, we propose the Review-guided Answer Helpfulness Prediction (RAHP) model that not only considers the interactions between QA pairs but also investigates the opinion coherence between the answer and crowds' opinions reflected in the reviews, which is another important factor to identify helpful answers. Moreover, we tackle the task of determining opinion coherence as a language inference problem and explore the utilization of pre-training strategy to transfer the textual inference knowledge obtained from a specifically designed trained network. Extensive experiments conducted on real-world data across seven product categories show that our proposed model achieves superior performance on the prediction task.Comment: Accepted by WWW202

    Learning Contextualized Document Representations for Healthcare Answer Retrieval

    Full text link
    We present Contextual Discourse Vectors (CDV), a distributed document representation for efficient answer retrieval from long healthcare documents. Our approach is based on structured query tuples of entities and aspects from free text and medical taxonomies. Our model leverages a dual encoder architecture with hierarchical LSTM layers and multi-task training to encode the position of clinical entities and aspects alongside the document discourse. We use our continuous representations to resolve queries with short latency using approximate nearest neighbor search on sentence level. We apply the CDV model for retrieving coherent answer passages from nine English public health resources from the Web, addressing both patients and medical professionals. Because there is no end-to-end training data available for all application scenarios, we train our model with self-supervised data from Wikipedia. We show that our generalized model significantly outperforms several state-of-the-art baselines for healthcare passage ranking and is able to adapt to heterogeneous domains without additional fine-tuning.Comment: The Web Conference 2020 (WWW '20

    Text Matching as Image Recognition

    No full text
    Matching two texts is a fundamental problem in many natural language processing tasks. An effective way is to extract meaningful matching patterns from words, phrases, and sentences to produce the matching score. Inspired by the success of convolutional neural network in image recognition, where neurons can capture many complicated patterns based on the extracted elementary visual patterns such as oriented edges and corners, we propose to model text matching as the problem of image recognition. Firstly, a matching matrix whose entries represent the similarities between words is constructed and viewed as an image. Then a convolutional neural network is utilized to capture rich matching patterns in a layer-by-layer way. We show that by resembling the compositional hierarchies of patterns in image recognition, our model can successfully identify salient signals such as n-gram and n-term matchings. Experimental results demonstrate its superiority against the baselines

    aMV-LSTM: an attention-based model with multiple positional text matching

    Get PDF
    National audienceDeep models are getting a wide interest in recent NLP and IR state-of-the-art. Among the proposed models, position-based models and attention-based models take into account the word position in the text, in the former, and the importance of a word among other words in the latter. The positional information are some of the important features that help text representation learning. However, the importance of a given word among others in a given text, which is an important aspect in text matching, is not considered in positional features. In this paper, we propose a model that combines position-based representation learning approach with the attention-based weighting process. The latter learns an importance coefficient for each word of the input text. We propose an extension of a position-based model MV-LSTM with an attention layer, allowing a parameterizable architecture. We believe that when the model is aware of both word position and importance, the learned representations will get more relevant features for the matching process. Our model, namely aMV-LSTM, learns the attention based coefficients to weight words of the different input sentences, before computing their position-based representations. Experimental results, in question/answer matching and question pairs identification tasks, show that the proposed model outperforms the MV-LSTM baseline and several state-of-the-art models

    High Efficiency Dye-sensitized Solar Cells Constructed with Composites of TiO2 and the Hot-bubbling Synthesized Ultra-Small SnO2 Nanocrystals

    No full text
    International audienceAn efficient photo-anode for the dye-sensitized solar cells (DSSCs) should have features of high loading of dye molecules, favorable band alignments and good efficiency in electron transport. Herein, the 3.4 nm-sized SnO2 nanocrystals (NCs) of high crystallinity, synthesized via the hot-bubbling method, were incorporated with the commercial TiO2 (P25) particles to fabricate the photo-anodes. The optimal percentage of the doped SnO2 NCs was found at ~7.5% (SnO2/TiO2, w/w), and the fabricated DSSC delivers a power conversion efficiency up to 6.7%, which is 1.52 times of the P25 based DSSCs. The ultra-small SnO2 NCs offer three benefits, (1) the incorporation of SnO2 NCs enlarges surface areas of the photo-anode films, and higher dye-loading amounts were achieved; (2) the high charge mobility provided by SnO2 was confirmed to accelerate the electron transport, and the photo-electron recombination was suppressed by the highly-crystallized NCs; (3) the conduction band minimum (CBM) of the SnO2 NCs was uplifted due to the quantum size effects, and this was found to alleviate the decrement in the open-circuit voltage. This work highlights great contributions of the SnO2 NCs to the improvement of the photovoltaic performances in the DSSCs

    Solid Material Formula of Small Fragrant Chicken Soup Stock and Improvement of Boiling Process

    No full text
    [Objectives] To study the solid material formula and boiling process of Small Fragrant Chicken Soup Stock. [Methods] Small fragrant chicken and edible fungi were used as the main raw materials to prepare instant chicken and Morchella esculenta soup stock, and the influencing factors such as solid material ratio, ratio of liquid to material, time and temperature were studied, and the optimal formula and process of the soup stock were screened out through sensory and instrument evaluation experiments. [Results] The optimal formula and process of Small Fragrant Chicken Soup Stock were as follows: small fragrant chicken 40%, M. esculenta 6%, Lyophyllum decastes 3% and Lentinus edodes 1.5%; the ratio of liquid to material was 2.51, the boiling time was 2 h, and the boiling temperature was 120 ℃.Under these conditions, the sensory score of Small Fragrant Chicken Soup Stock was the best. [Conclusions] This study provided a theoretical basis for the industrial production of edible mushroom prepared dishes
    corecore