516 research outputs found

    Deep Recurrent Generative Decoder for Abstractive Text Summarization

    Full text link
    We propose a new framework for abstractive text summarization based on a sequence-to-sequence oriented encoder-decoder model equipped with a deep recurrent generative decoder (DRGN). Latent structure information implied in the target summaries is learned based on a recurrent latent random model for improving the summarization quality. Neural variational inference is employed to address the intractable posterior inference for the recurrent latent variables. Abstractive summaries are generated based on both the generative latent variables and the discriminative deterministic states. Extensive experiments on some benchmark datasets in different languages show that DRGN achieves improvements over the state-of-the-art methods.Comment: 10 pages, EMNLP 201

    Supervised topic models with word order structure for document classification and retrieval learning

    Get PDF
    One limitation of most existing probabilistic latent topic models for document classification is that the topic model itself does not consider useful side-information, namely, class labels of documents. Topic models, which in turn consider the side-information, popularly known as supervised topic models, do not consider the word order structure in documents. One of the motivations behind considering the word order structure is to capture the semantic fabric of the document. We investigate a low-dimensional latent topic model for document classification. Class label information and word order structure are integrated into a supervised topic model enabling a more effective interaction among such information for solving document classification. We derive a collapsed Gibbs sampler for our model. Likewise, supervised topic models with word order structure have not been explored in document retrieval learning. We propose a novel supervised topic model for document retrieval learning which can be regarded as a pointwise model for tackling the learning-to-rank task. Available relevance assessments and word order structure are integrated into the topic model itself. We conduct extensive experiments on several publicly available benchmark datasets, and show that our model improves upon the state-of-the-art models

    Robust Intrinsic Ferromagnetism and Half Semiconductivity in Stable Two-Dimensional Single-Layer Chromium Trihalides

    Full text link
    Two-dimensional (2D) intrinsic ferromagnetic (FM) semiconductors are crucial to develop low-dimensional spintronic devices. Using density functional theory, we show that single-layer chromium trihalides (SLCTs) (CrX3_3,X=F, Cl, Br and I) constitute a series of stable 2D intrinsic FM semiconductors. A free-standing SLCT can be easily exfoliated from the bulk crystal, due to a low cleavage energy and a high in-plane stiffness. Electronic structure calculations using the HSE06 functional indicate that both bulk and single-layer CrX3_3 are half semiconductors with indirect gaps and their valence bands and conduction bands are fully spin-polarized in the same spin direction. The energy gaps and absorption edges of CrBr3_3 and CrI3_3 are found to be in the visible frequency range, which implies possible opt-electronic applications. Furthermore, SLCTs are found to possess a large magnetic moment of 3Ī¼B\mu_B per formula unit and a sizable magnetic anisotropy energy. The magnetic exchange constants of SLCTs are then extracted using the Heisenberg spin Hamiltonian and the microscopic origins of the various exchange interactions are analyzed. A competition between a near 90āˆ˜^\circ FM superexchange and a direct antiferromagnetic (AFM) exchange results in a FM nearest-neighbour exchange interaction. The next and third nearest-neighbour exchange interactions are found to be FM and AFM respectively and this can be understood by the angle-dependent extended Cr-X-X-Cr superexchange interaction. Moreover, the Curie temperatures of SLCTs are also predicted using Monte Carlo simulations and the values can further increase by applying a biaxial tensile strain. The unique combination of robust intrinsic ferromagnetism, half semiconductivity and large magnetic anisotropy energies renders the SLCTs as promising candidates for next-generation semiconductor spintronic applications.Comment: 12 pages, 14 figures. published in J. Mater. Chem.

    Web Query Reformulation via Joint Modeling of Latent Topic Dependency and Term Context

    Get PDF
    An important way to improve usersā€™ satisfaction in Web search is to assist them by issuing more effective queries. One such approach is query reformulation, which generates new queries according to the current query issued by users. A common procedure for conducting reformulation is to generate some candidate queries first, then a scoring method is employed to assess these candidates. Currently, most of the existing methods are context based. They rely heavily on the context relation of terms in the history queries and cannot detect and maintain the semantic consistency of queries. In this article, we propose a graphical model to score queries. The proposed model exploits a latent topic space, which is automatically derived from the query log, to detect semantic dependency of terms in a query and dependency among topics. Meanwhile, the graphical model also captures the term context in the history query by skip-bigram and n-gram language models. In addition, our model can be easily extended to consider usersā€™ history search interests when we conduct query reformulation for different users. In the task of candidate query generation, we investigate a social tagging data resourceā€”Delicious bookmarkā€”to generate addition and substitution patterns that are employed as supplements to the patterns generated from query log data

    Abstractive Multi-Document Summarization via Phrase Selection and Merging

    Full text link
    We propose an abstraction-based multi-document summarization framework that can construct new sentences by exploring more fine-grained syntactic units than sentences, namely, noun/verb phrases. Different from existing abstraction-based approaches, our method first constructs a pool of concepts and facts represented by phrases from the input documents. Then new sentences are generated by selecting and merging informative phrases to maximize the salience of phrases and meanwhile satisfy the sentence construction constraints. We employ integer linear optimization for conducting phrase selection and merging simultaneously in order to achieve the global optimal solution for a summary. Experimental results on the benchmark data set TAC 2011 show that our framework outperforms the state-of-the-art models under automated pyramid evaluation metric, and achieves reasonably well results on manual linguistic quality evaluation.Comment: 11 pages, 1 figure, accepted as a full paper at ACL 201

    Bureaucracy and red tape ureaucracy and red tape

    Get PDF
    Thesis (B.Sc)--University of Hong Kong, 2004.published_or_final_versio
    • ā€¦
    corecore