7,672 research outputs found
End-to-End Entity Detection with Proposer and Regressor
Named entity recognition is a traditional task in natural language
processing. In particular, nested entity recognition receives extensive
attention for the widespread existence of the nesting scenario. The latest
research migrates the well-established paradigm of set prediction in object
detection to cope with entity nesting. However, the manual creation of query
vectors, which fail to adapt to the rich semantic information in the context,
limits these approaches. An end-to-end entity detection approach with proposer
and regressor is presented in this paper to tackle the issues. First, the
proposer utilizes the feature pyramid network to generate high-quality entity
proposals. Then, the regressor refines the proposals for generating the final
prediction. The model adopts encoder-only architecture and thus obtains the
advantages of the richness of query semantics, high precision of entity
localization, and easiness of model training. Moreover, we introduce the novel
spatially modulated attention and progressive refinement for further
improvement. Extensive experiments demonstrate that our model achieves advanced
performance in flat and nested NER, achieving a new state-of-the-art F1 score
of 80.74 on the GENIA dataset and 72.38 on the WeiboNER dataset
Deep Learning Techniques for Music Generation -- A Survey
This paper is a survey and an analysis of different ways of using deep
learning (deep artificial neural networks) to generate musical content. We
propose a methodology based on five dimensions for our analysis:
Objective - What musical content is to be generated? Examples are: melody,
polyphony, accompaniment or counterpoint. - For what destination and for what
use? To be performed by a human(s) (in the case of a musical score), or by a
machine (in the case of an audio file).
Representation - What are the concepts to be manipulated? Examples are:
waveform, spectrogram, note, chord, meter and beat. - What format is to be
used? Examples are: MIDI, piano roll or text. - How will the representation be
encoded? Examples are: scalar, one-hot or many-hot.
Architecture - What type(s) of deep neural network is (are) to be used?
Examples are: feedforward network, recurrent network, autoencoder or generative
adversarial networks.
Challenge - What are the limitations and open challenges? Examples are:
variability, interactivity and creativity.
Strategy - How do we model and control the process of generation? Examples
are: single-step feedforward, iterative feedforward, sampling or input
manipulation.
For each dimension, we conduct a comparative analysis of various models and
techniques and we propose some tentative multidimensional typology. This
typology is bottom-up, based on the analysis of many existing deep-learning
based systems for music generation selected from the relevant literature. These
systems are described and are used to exemplify the various choices of
objective, representation, architecture, challenge and strategy. The last
section includes some discussion and some prospects.Comment: 209 pages. This paper is a simplified version of the book: J.-P.
Briot, G. Hadjeres and F.-D. Pachet, Deep Learning Techniques for Music
Generation, Computational Synthesis and Creative Systems, Springer, 201
SMART: A Situation Model for Algebra Story Problems via Attributed Grammar
Solving algebra story problems remains a challenging task in artificial
intelligence, which requires a detailed understanding of real-world situations
and a strong mathematical reasoning capability. Previous neural solvers of math
word problems directly translate problem texts into equations, lacking an
explicit interpretation of the situations, and often fail to handle more
sophisticated situations. To address such limits of neural solvers, we
introduce the concept of a \emph{situation model}, which originates from
psychology studies to represent the mental states of humans in problem-solving,
and propose \emph{SMART}, which adopts attributed grammar as the representation
of situation models for algebra story problems. Specifically, we first train an
information extraction module to extract nodes, attributes, and relations from
problem texts and then generate a parse graph based on a pre-defined attributed
grammar. An iterative learning strategy is also proposed to improve the
performance of SMART further. To rigorously study this task, we carefully
curate a new dataset named \emph{ASP6.6k}. Experimental results on ASP6.6k show
that the proposed model outperforms all previous neural solvers by a large
margin while preserving much better interpretability. To test these models'
generalization capability, we also design an out-of-distribution (OOD)
evaluation, in which problems are more complex than those in the training set.
Our model exceeds state-of-the-art models by 17\% in the OOD evaluation,
demonstrating its superior generalization ability
- …