975 research outputs found

    Table-to-text Generation by Structure-aware Seq2seq Learning

    Full text link
    Table-to-text generation aims to generate a description for a factual table which can be viewed as a set of field-value records. To encode both the content and the structure of a table, we propose a novel structure-aware seq2seq architecture which consists of field-gating encoder and description generator with dual attention. In the encoding phase, we update the cell memory of the LSTM unit by a field gate and its corresponding field value in order to incorporate field information into table representation. In the decoding phase, dual attention mechanism which contains word level attention and field level attention is proposed to model the semantic relevance between the generated description and the table. We conduct experiments on the \texttt{WIKIBIO} dataset which contains over 700k biographies and corresponding infoboxes from Wikipedia. The attention visualizations and case studies show that our model is capable of generating coherent and informative descriptions based on the comprehensive understanding of both the content and the structure of a table. Automatic evaluations also show our model outperforms the baselines by a great margin. Code for this work is available on https://github.com/tyliupku/wiki2bio.Comment: Accepted by AAAI201

    Harnessing the Plug-and-Play Controller by Prompting

    Full text link
    Controllable text generation is a growing field within natural language generation (NLG) that focuses on producing text that meets specific constraints in real-world applications. Previous approaches, such as plug-and-play controllers (PPCs), aimed to steer the properties of generated text in a flexible manner. However, these methods often compromised the integrity of the language model's decoding process, resulting in less smooth text generation. Alternatively, other techniques utilized multiple attribute prompts to align the generated text with desired attributes, but this approach required prompt design for each attribute and was dependent on the size of the language model. This paper introduces a novel method for flexible attribute control in text generation using pre-trained language models (PLMs). The proposed approach aims to enhance the fluency of generated text by guiding the generation process with PPCs. The key idea is to dynamically adjust the distribution of generated text by modifying prompts, effectively constraining the output space of the language model and influencing the desired attribute. To enable smooth cooperation between the PLM and the PPC, our work innovatively proposes a new model fine-tuning method: Reinforcement Learning with Dynamic Adjust Feedback (RLDAF).This fine-tuning process adapts a small subset of the language model's parameters based on the generating actions taken during the PPC control process. The resulting harmonious collaboration between the PLM and PPC leads to improved smoothness in text generation during inference. Extensive experiments were conducted on the SST2 dataset, and the proposed method outperformed previous approaches in various evaluation metrics, including text fluency and attribute consistency.Comment: The Third Version of the Generation, Evaluation & Metrics (GEM) Workshop in EMNLP 202

    Text Attribute Control via Closed-Loop Disentanglement

    Full text link
    Changing an attribute of a text without changing the content usually requires to first disentangle the text into irrelevant attributes and content representations. After that, in the inference phase, the representation of one attribute is tuned to a different value, expecting that the corresponding attribute of the text can also be changed accordingly. The usual way of disentanglement is to add some constraints on the latent space of an encoder-decoder architecture, including adversarial-based constraints and mutual-information-based constraints. However, the previous semi-supervised processes of attribute change are usually not enough to guarantee the success of attribute change and content preservation. In this paper, we propose a novel approach to achieve a robust control of attributes while enhancing content preservation. In this approach, we use a semi-supervised contrastive learning method to encourage the disentanglement of attributes in latent spaces. Differently from previous works, we re-disentangle the reconstructed sentence and compare the re-disentangled latent space with the original latent space, which makes a closed-loop disentanglement process. This also helps content preservation. In addition, the contrastive learning method is also able to replace the role of minimizing mutual information and adversarial training in the disentanglement process, which alleviates the computation cost. We conducted experiments on three text datasets, including the Yelp Service review dataset, the Amazon Product review dataset, and the GoEmotions dataset. The experimental results show the effectiveness of our model.Comment: accepted by TACL 202

    Multi-type Disentanglement without Adversarial Training

    Full text link
    Controlling the style of natural language by disentangling the latent space is an important step towards interpretable machine learning. After the latent space is disentangled, the style of a sentence can be transformed by tuning the style representation without affecting other features of the sentence. Previous works usually use adversarial training to guarantee that disentangled vectors do not affect each other. However, adversarial methods are difficult to train. Especially when there are multiple features (e.g., sentiment, or tense, which we call style types in this paper), each feature requires a separate discriminator for extracting a disentangled style vector corresponding to that feature. In this paper, we propose a unified distribution-controlling method, which provides each specific style value (the value of style types, e.g., positive sentiment, or past tense) with a unique representation. This method contributes a solid theoretical basis to avoid adversarial training in multi-type disentanglement. We also propose multiple loss functions to achieve a style-content disentanglement as well as a disentanglement among multiple style types. In addition, we observe that if two different style types always have some specific style values that occur together in the dataset, they will affect each other when transferring the style values. We call this phenomenon training bias, and we propose a loss function to alleviate such training bias while disentangling multiple types. We conduct experiments on two datasets (Yelp service reviews and Amazon product reviews) to evaluate the style-disentangling effect and the unsupervised style transfer performance on two style types: sentiment and tense. The experimental results show the effectiveness of our model

    Updated insights into 3D architecture electrodes for micropower sources

    Get PDF
    Microbatteries (MBs) and microsupercapacitors (MSCs) are primary on-chip micropower sources that drive autonomous and stand-alone microelectronic devices for implementation of the Internet of Things (IoT). However, the performance of conventional MBs and MSCs is restricted by their 2D thin-film electrode design, and these devices struggle to satisfy the increasing IoT energy demands for high energy density, high power density, and long lifespan. The energy densities of MBs and MSCs can be improved significantly through adoption of a 2D thick-film electrode design; however, their power densities and lifespans deteriorate with increased electrode thickness. In contrast, 3D architecture electrodes offer remarkable opportunities to simultaneously improve MB and MSC energy density, power density, and lifespan. To date, various 3D architecture electrodes have been designed, fabricated, and investigated for MBs and MSCs. This review provides an update on the principal superiorities of 3D architecture electrodes over 2D thick-film electrodes in the context of improved MB and MSC energy density, power density, and lifespan. In addition, the most recent and representative progress in 3D architecture electrode development for MBs and MSCs is highlighted. Finally, present challenges are discussed and key perspectives for future research in this field are outlined

    DETERMINANTS OF AGRICULTURE-RELATED LOAN DEFAULT: EVIDENCE FROM CHINA

    Get PDF
    This paper investigates agriculture-related loan default in 2002–2009 through a large data set from a leading Chinese state-owned bank. Using logit regression, we find the default rate on agriculture-related loans is significantly higher than that on non–agriculture-related loans. We find that base interest rates, loan maturity, the type of collateral, firm size, ownership structure, and managerial quality rating have a significant impact on agriculture-related loan default, but this also depends on how agriculture-related loans are defined. The results provide insight into the real impact of monetary policy on agriculture-related lending.This paper investigates agriculture-related loan default in 2002–2009 through a large data set from a leading Chinese state-owned bank. Using logit regression, we find the default rate on agriculture-related loans is significantly higher than that on non–agriculture-related loans. We find that base interest rates, loan maturity, the type of collateral, firm size, ownership structure, and managerial quality rating have a significant impact on agriculture-related loan default, but this also depends on how agriculture-related loans are defined. The results provide insight into the real impact of monetary policy on agriculture-related lending

    Order-Planning Neural Text Generation From Structured Data

    Full text link
    Generating texts from structured data (e.g., a table) is important for various natural language processing tasks such as question answering and dialog systems. In recent studies, researchers use neural language models and encoder-decoder frameworks for table-to-text generation. However, these neural network-based approaches do not model the order of contents during text generation. When a human writes a summary based on a given table, he or she would probably consider the content order before wording. In a biography, for example, the nationality of a person is typically mentioned before occupation in a biography. In this paper, we propose an order-planning text generation model to capture the relationship between different fields and use such relationship to make the generated text more fluent and smooth. We conducted experiments on the WikiBio dataset and achieve significantly higher performance than previous methods in terms of BLEU, ROUGE, and NIST scores

    Self-assembly of noble metal nanoparticles into sub-100 nm colloidosomes with collective optical and catalytic properties.

    Get PDF
    Self-assembly at the nanoscale represents a powerful tool for creating materials with new structures and intriguing collective properties. Here, we report a novel strategy to synthesize nanoscale colloidosomes of noble metals by assembling primary metal nanoparticles at the interface of emulsion droplets formed by their capping agent. This strategy produces noble metal colloidosomes of unprecedentedly small sizes (<100 nm) in high yield and uniformity, which is highly desirable for practical applications. In addition, it enables the high tunability of the composition, producing a diversity of monometallic and bimetallic alloy colloidosomes. The colloidosomes exhibit interesting collective properties that are different from those of individual colloidal nanoparticles. Specifically, we demonstrate Au colloidosomes with well-controlled interparticle plasmon coupling and Au-Pd alloy colloidosomes with superior electrocatalytic performance, both thanks to the special structural features that arise from the assembly. We believe this strategy provides a general platform for producing a rich class of miniature colloidosomes that may have fascinating collective properties for a broad range of applications
    • …
    corecore