61 research outputs found

    Number Sequence Prediction Problems for Evaluating Computational Powers of Neural Networks

    Full text link
    Inspired by number series tests to measure human intelligence, we suggest number sequence prediction tasks to assess neural network models' computational powers for solving algorithmic problems. We define the complexity and difficulty of a number sequence prediction task with the structure of the smallest automaton that can generate the sequence. We suggest two types of number sequence prediction problems: the number-level and the digit-level problems. The number-level problems format sequences as 2-dimensional grids of digits and the digit-level problems provide a single digit input per a time step. The complexity of a number-level sequence prediction can be defined with the depth of an equivalent combinatorial logic, and the complexity of a digit-level sequence prediction can be defined with an equivalent state automaton for the generation rule. Experiments with number-level sequences suggest that CNN models are capable of learning the compound operations of sequence generation rules, but the depths of the compound operations are limited. For the digit-level problems, simple GRU and LSTM models can solve some problems with the complexity of finite state automata. Memory augmented models such as Stack-RNN, Attention, and Neural Turing Machines can solve the reverse-order task which has the complexity of simple pushdown automaton. However, all of above cannot solve general Fibonacci, Arithmetic or Geometric sequence generation problems that represent the complexity of queue automata or Turing machines. The results show that our number sequence prediction problems effectively evaluate machine learning models' computational capabilities.Comment: Accepted to 2019 AAAI Conference on Artificial Intelligenc

    Improving Neural Question Generation using Answer Separation

    Full text link
    Neural question generation (NQG) is the task of generating a question from a given passage with deep neural networks. Previous NQG models suffer from a problem that a significant proportion of the generated questions include words in the question target, resulting in the generation of unintended questions. In this paper, we propose answer-separated seq2seq, which better utilizes the information from both the passage and the target answer. By replacing the target answer in the original passage with a special token, our model learns to identify which interrogative word should be used. We also propose a new module termed keyword-net, which helps the model better capture the key information in the target answer and generate an appropriate question. Experimental results demonstrate that our answer separation method significantly reduces the number of improper questions which include answers. Consequently, our model significantly outperforms previous state-of-the-art NQG models.Comment: The paper is accepted to AAAI 201

    Neural Sequence-to-grid Module for Learning Symbolic Rules

    Full text link
    Logical reasoning tasks over symbols, such as learning arithmetic operations and computer program evaluations, have become challenges to deep learning. In particular, even state-of-the-art neural networks fail to achieve \textit{out-of-distribution} (OOD) generalization of symbolic reasoning tasks, whereas humans can easily extend learned symbolic rules. To resolve this difficulty, we propose a neural sequence-to-grid (seq2grid) module, an input preprocessor that automatically segments and aligns an input sequence into a grid. As our module outputs a grid via a novel differentiable mapping, any neural network structure taking a grid input, such as ResNet or TextCNN, can be jointly trained with our module in an end-to-end fashion. Extensive experiments show that neural networks having our module as an input preprocessor achieve OOD generalization on various arithmetic and algorithmic problems including number sequence prediction problems, algebraic word problems, and computer program evaluation problems while other state-of-the-art sequence transduction models cannot. Moreover, we verify that our module enhances TextCNN to solve the bAbI QA tasks without external memory.Comment: 9 pages, 9 figures, AAAI 202

    Weakly Supervised Semantic Parsing with Execution-based Spurious Program Filtering

    Full text link
    The problem of spurious programs is a longstanding challenge when training a semantic parser from weak supervision. To eliminate such programs that have wrong semantics but correct denotation, existing methods focus on exploiting similarities between examples based on domain-specific knowledge. In this paper, we propose a domain-agnostic filtering mechanism based on program execution results. Specifically, for each program obtained through the search process, we first construct a representation that captures the program's semantics as execution results under various inputs. Then, we run a majority vote on these representations to identify and filter out programs with significantly different semantics from the other programs. In particular, our method is orthogonal to the program search process so that it can easily augment any of the existing weakly supervised semantic parsing frameworks. Empirical evaluations on the Natural Language Visual Reasoning and WikiTableQuestions demonstrate that applying our method to the existing semantic parsers induces significantly improved performances.Comment: EMNLP 202

    Propagate-Selector: Detecting Supporting Sentences for Question Answering via Graph Neural Networks

    Full text link
    In this study, we propose a novel graph neural network called propagate-selector (PS), which propagates information over sentences to understand information that cannot be inferred when considering sentences in isolation. First, we design a graph structure in which each node represents an individual sentence, and some pairs of nodes are selectively connected based on the text structure. Then, we develop an iterative attentive aggregation and a skip-combine method in which a node interacts with its neighborhood nodes to accumulate the necessary information. To evaluate the performance of the proposed approaches, we conduct experiments with the standard HotpotQA dataset. The empirical results demonstrate the superiority of our proposed approach, which obtains the best performances, compared to the widely used answer-selection models that do not consider the intersentential relationship.Comment: 8 pages, Accepted as a conference paper at LREC 202

    Anisotropic Thermal Conductivity of Nickel-Based Superalloy CM247LC Fabricated via Selective Laser Melting

    Get PDF
    Efforts to enhance thermal efficiency of turbines by increasing the turbine inlet temperature have been further accelerated by the introduction of 3D printing to turbine components as complex cooling geometry can be implemented using this technique. However, as opposed to the properties of materials fabricated by conventional methods, the properties of materials manufactured by 3D printing are not isotropic. In this study, we analyzed the anisotropic thermal conductivity of nickel-based superalloy CM247LC manufactured by selective laser melting (SLM). We found that as the density decreases, so does the thermal conductivity. In addition, the anisotropy in thermal conductivity is more pronounced at lower densities. It was confirmed that the samples manufactured with low energy density have the same electron thermal conductivity with respect to the orientation, but the lattice thermal conductivity was about 16.5% higher in the in-plane direction than in the cross-plane direction. This difference in anisotropic lattice thermal conductivity is proportional to the difference in square root of elastic modulus. We found that ellipsoidal pores contributed to a direction-dependent elastic modulus, resulting in anisotropy in thermal conductivity. The results of this study should be beneficial not only for designing next-generation gas turbines, but also for any system produced by 3D printing
    • โ€ฆ
    corecore