61 research outputs found
Number Sequence Prediction Problems for Evaluating Computational Powers of Neural Networks
Inspired by number series tests to measure human intelligence, we suggest
number sequence prediction tasks to assess neural network models' computational
powers for solving algorithmic problems. We define the complexity and
difficulty of a number sequence prediction task with the structure of the
smallest automaton that can generate the sequence. We suggest two types of
number sequence prediction problems: the number-level and the digit-level
problems. The number-level problems format sequences as 2-dimensional grids of
digits and the digit-level problems provide a single digit input per a time
step. The complexity of a number-level sequence prediction can be defined with
the depth of an equivalent combinatorial logic, and the complexity of a
digit-level sequence prediction can be defined with an equivalent state
automaton for the generation rule. Experiments with number-level sequences
suggest that CNN models are capable of learning the compound operations of
sequence generation rules, but the depths of the compound operations are
limited. For the digit-level problems, simple GRU and LSTM models can solve
some problems with the complexity of finite state automata. Memory augmented
models such as Stack-RNN, Attention, and Neural Turing Machines can solve the
reverse-order task which has the complexity of simple pushdown automaton.
However, all of above cannot solve general Fibonacci, Arithmetic or Geometric
sequence generation problems that represent the complexity of queue automata or
Turing machines. The results show that our number sequence prediction problems
effectively evaluate machine learning models' computational capabilities.Comment: Accepted to 2019 AAAI Conference on Artificial Intelligenc
Improving Neural Question Generation using Answer Separation
Neural question generation (NQG) is the task of generating a question from a
given passage with deep neural networks. Previous NQG models suffer from a
problem that a significant proportion of the generated questions include words
in the question target, resulting in the generation of unintended questions. In
this paper, we propose answer-separated seq2seq, which better utilizes the
information from both the passage and the target answer. By replacing the
target answer in the original passage with a special token, our model learns to
identify which interrogative word should be used. We also propose a new module
termed keyword-net, which helps the model better capture the key information in
the target answer and generate an appropriate question. Experimental results
demonstrate that our answer separation method significantly reduces the number
of improper questions which include answers. Consequently, our model
significantly outperforms previous state-of-the-art NQG models.Comment: The paper is accepted to AAAI 201
Neural Sequence-to-grid Module for Learning Symbolic Rules
Logical reasoning tasks over symbols, such as learning arithmetic operations
and computer program evaluations, have become challenges to deep learning. In
particular, even state-of-the-art neural networks fail to achieve
\textit{out-of-distribution} (OOD) generalization of symbolic reasoning tasks,
whereas humans can easily extend learned symbolic rules. To resolve this
difficulty, we propose a neural sequence-to-grid (seq2grid) module, an input
preprocessor that automatically segments and aligns an input sequence into a
grid. As our module outputs a grid via a novel differentiable mapping, any
neural network structure taking a grid input, such as ResNet or TextCNN, can be
jointly trained with our module in an end-to-end fashion. Extensive experiments
show that neural networks having our module as an input preprocessor achieve
OOD generalization on various arithmetic and algorithmic problems including
number sequence prediction problems, algebraic word problems, and computer
program evaluation problems while other state-of-the-art sequence transduction
models cannot. Moreover, we verify that our module enhances TextCNN to solve
the bAbI QA tasks without external memory.Comment: 9 pages, 9 figures, AAAI 202
Weakly Supervised Semantic Parsing with Execution-based Spurious Program Filtering
The problem of spurious programs is a longstanding challenge when training a
semantic parser from weak supervision. To eliminate such programs that have
wrong semantics but correct denotation, existing methods focus on exploiting
similarities between examples based on domain-specific knowledge. In this
paper, we propose a domain-agnostic filtering mechanism based on program
execution results. Specifically, for each program obtained through the search
process, we first construct a representation that captures the program's
semantics as execution results under various inputs. Then, we run a majority
vote on these representations to identify and filter out programs with
significantly different semantics from the other programs. In particular, our
method is orthogonal to the program search process so that it can easily
augment any of the existing weakly supervised semantic parsing frameworks.
Empirical evaluations on the Natural Language Visual Reasoning and
WikiTableQuestions demonstrate that applying our method to the existing
semantic parsers induces significantly improved performances.Comment: EMNLP 202
Propagate-Selector: Detecting Supporting Sentences for Question Answering via Graph Neural Networks
In this study, we propose a novel graph neural network called
propagate-selector (PS), which propagates information over sentences to
understand information that cannot be inferred when considering sentences in
isolation. First, we design a graph structure in which each node represents an
individual sentence, and some pairs of nodes are selectively connected based on
the text structure. Then, we develop an iterative attentive aggregation and a
skip-combine method in which a node interacts with its neighborhood nodes to
accumulate the necessary information. To evaluate the performance of the
proposed approaches, we conduct experiments with the standard HotpotQA dataset.
The empirical results demonstrate the superiority of our proposed approach,
which obtains the best performances, compared to the widely used
answer-selection models that do not consider the intersentential relationship.Comment: 8 pages, Accepted as a conference paper at LREC 202
Anisotropic Thermal Conductivity of Nickel-Based Superalloy CM247LC Fabricated via Selective Laser Melting
Efforts to enhance thermal efficiency of turbines by increasing the turbine inlet temperature have been further accelerated by the introduction of 3D printing to turbine components as complex cooling geometry can be implemented using this technique. However, as opposed to the properties of materials fabricated by conventional methods, the properties of materials manufactured by 3D printing are not isotropic. In this study, we analyzed the anisotropic thermal conductivity of nickel-based superalloy CM247LC manufactured by selective laser melting (SLM). We found that as the density decreases, so does the thermal conductivity. In addition, the anisotropy in thermal conductivity is more pronounced at lower densities. It was confirmed that the samples manufactured with low energy density have the same electron thermal conductivity with respect to the orientation, but the lattice thermal conductivity was about 16.5% higher in the in-plane direction than in the cross-plane direction. This difference in anisotropic lattice thermal conductivity is proportional to the difference in square root of elastic modulus. We found that ellipsoidal pores contributed to a direction-dependent elastic modulus, resulting in anisotropy in thermal conductivity. The results of this study should be beneficial not only for designing next-generation gas turbines, but also for any system produced by 3D printing
- โฆ