120,738 research outputs found
Event-related potentials elicited by spoken relative clauses
Sentence-length event-related potential (ERP) waveforms were obtained from 23 scalp sites as 24 subjects listened to normally spoken sentences of various syntactic structures. The critical materials consisted of 36 sentences each containing one of 2 types of relative clauses that differ in processing difficulty, namely Subject Object (SO) and Subject Subject (SS) relative clauses. Sentence-length ERPs showed several differences in the slow scalp potentials elicited by SO and SS sentences that were similar in their temporal dynamics to those elicited by the same stimuli in a word-by-word reading experiment, although the effects in the two modalities have non identical distributions. Just as for written sentences, there was a large, fronto-central negativity beginning at the linguistically defined "gap" in the SO sentences; this effect was largest for listeners with above-median comprehension rates, and is hypothesized to index changes in on-line processing demands during comprehension
Complex Networks Measures for Differentiation between Normal and Shuffled Croatian Texts
This paper studies the properties of the Croatian texts via complex networks.
We present network properties of normal and shuffled Croatian texts for
different shuffling principles: on the sentence level and on the text level. In
both experiments we preserved the vocabulary size, word and sentence frequency
distributions. Additionally, in the first shuffling approach we preserved the
sentence structure of the text and the number of words per sentence. Obtained
results showed that degree rank distributions exhibit no substantial deviation
in shuffled networks, and strength rank distributions are preserved due to the
same word frequencies. Therefore, standard approach to study the structure of
linguistic co-occurrence networks showed no clear difference among the
topologies of normal and shuffled texts. Finally, we showed that the in- and
out- selectivity values from shuffled texts are constantly below selectivity
values calculated from normal texts. Our results corroborate that the node
selectivity measure can capture structural differences between original and
shuffled Croatian texts
Self-Adaptive Hierarchical Sentence Model
The ability to accurately model a sentence at varying stages (e.g.,
word-phrase-sentence) plays a central role in natural language processing. As
an effort towards this goal we propose a self-adaptive hierarchical sentence
model (AdaSent). AdaSent effectively forms a hierarchy of representations from
words to phrases and then to sentences through recursive gated local
composition of adjacent segments. We design a competitive mechanism (through
gating networks) to allow the representations of the same sentence to be
engaged in a particular learning task (e.g., classification), therefore
effectively mitigating the gradient vanishing problem persistent in other
recursive models. Both qualitative and quantitative analysis shows that AdaSent
can automatically form and select the representations suitable for the task at
hand during training, yielding superior classification performance over
competitor models on 5 benchmark data sets.Comment: 8 pages, 7 figures, accepted as a full paper at IJCAI 201
Frequency Value Grammar and Information Theory
I previously laid the groundwork for Frequency Value Grammar (FVG) in papers I submitted in the proceedings of the 4th International Conference on Cognitive Science (2003), Sydney Australia, and Corpus Linguistics Conference (2003), Lancaster, UK. FVG is a formal syntax theoretically based in large part on Information Theory principles. FVG relies on dynamic physical principles external to the corpus which shape and mould the corpus whereas generative grammar and other formal syntactic theories are based exclusively on patterns (fractals) found occurring within the well-formed portion of the corpus. However, FVG should not be confused with Probability Syntax, (PS), as described by Manning (2003). PS is a corpus based approach that will yield the probability distribution of possible syntax constructions over a fixed corpus. PS makes no distinction between well and ill formed sentence constructions and assumes everything found in the corpus is well formed. In contrast, FVG’s primary objective is to distinguish between well and ill formed sentence constructions and, in so doing, relies on corpus based parameters which determine sentence competency. In PS, a syntax of high probability will not necessarily yield a well formed sentence. However, in FVG, a syntax or sentence construction of high ‘frequency value’ will yield a well-formed sentence, at least, 95% of the time satisfying most empirical standards. Moreover, in FVG, a sentence construction of ‘high frequency value’ could very well be represented by an underlying syntactic construction of low probability as determined by PS. The characteristic ‘frequency values’ calculated in FVG are not measures of probability but rather are fundamentally determined values derived from exogenous principles which impact and determine corpus based parameters serving as an index of sentence competency. The theoretical framework of FVG has broad applications beyond that of formal syntax and NLP. In this paper, I will demonstrate how FVG can be used as a model for improving the upper bound calculation of entropy of written English. Generally speaking, when a function word precedes an open class word, the backward n-gram analysis will be homomorphic with the information source and will result in frequency values more representative of co-occurrences in the information source
Synergies between processing and memory in children's reading span.
Previous research has established the relevance of working memory for cognitive development. Yet the factors responsible for shaping performance in the complex span tasks used to assess working memory capacity are not fully understood. We report a study of reading span in 7- to 11-year old children that addresses several contemporary theoretical issues. We demonstrate that both the timing and the accuracy of recall are affected by the presence or absence of a semantic connection between the processing requirement and the memoranda. Evidence that there can be synergies between processing and memory argues against the view that complex span simply measures the competition between these activities. We also demonstrate a consistent relationship between the rate of completing processing operations (sentence reading) and recall accuracy. At the same time, the shape and strength of this function varies with the task configuration. Taken together, these results demonstrate the potential for reconstructive influences to shape working memory performance among children
Temporal Deformable Convolutional Encoder-Decoder Networks for Video Captioning
It is well believed that video captioning is a fundamental but challenging
task in both computer vision and artificial intelligence fields. The prevalent
approach is to map an input video to a variable-length output sentence in a
sequence to sequence manner via Recurrent Neural Network (RNN). Nevertheless,
the training of RNN still suffers to some degree from vanishing/exploding
gradient problem, making the optimization difficult. Moreover, the inherently
recurrent dependency in RNN prevents parallelization within a sequence during
training and therefore limits the computations. In this paper, we present a
novel design --- Temporal Deformable Convolutional Encoder-Decoder Networks
(dubbed as TDConvED) that fully employ convolutions in both encoder and decoder
networks for video captioning. Technically, we exploit convolutional block
structures that compute intermediate states of a fixed number of inputs and
stack several blocks to capture long-term relationships. The structure in
encoder is further equipped with temporal deformable convolution to enable
free-form deformation of temporal sampling. Our model also capitalizes on
temporal attention mechanism for sentence generation. Extensive experiments are
conducted on both MSVD and MSR-VTT video captioning datasets, and superior
results are reported when comparing to conventional RNN-based encoder-decoder
techniques. More remarkably, TDConvED increases CIDEr-D performance from 58.8%
to 67.2% on MSVD.Comment: AAAI 201
Keystroke dynamics as signal for shallow syntactic parsing
Keystroke dynamics have been extensively used in psycholinguistic and writing
research to gain insights into cognitive processing. But do keystroke logs
contain actual signal that can be used to learn better natural language
processing models?
We postulate that keystroke dynamics contain information about syntactic
structure that can inform shallow syntactic parsing. To test this hypothesis,
we explore labels derived from keystroke logs as auxiliary task in a multi-task
bidirectional Long Short-Term Memory (bi-LSTM). Our results show promising
results on two shallow syntactic parsing tasks, chunking and CCG supertagging.
Our model is simple, has the advantage that data can come from distinct
sources, and produces models that are significantly better than models trained
on the text annotations alone.Comment: In COLING 201
Recommended from our members
Simulating the Noun-Verb Asymmetry in the Productivity of Children’s Speech
Several authors propose that children may acquire syntactic categories on the basis of co-occurrence statistics of words in the input. This paper assesses the relative merits of two such accounts by assessing the type and amount of productive language that results from computing co-occurrence statistics over conjoint and independent preceding and following contexts. This is achieved through the implementation of these methods in MOSAIC, a computational model of syntax acquisition that produces utterances that can be directly compared to child speech, and has a developmental component (i.e. produces increasingly long utterances). It is shown that the computation of co-occurrence statistics over conjoint contexts or frames results in a pattern of productive speech that more closely resembles that displayed by language learning children. The simulation of the developmental patterning of children’s productive speech furthermore suggests two refinements to this basic mechanism: inclusion of utterance boundaries, and the weighting of frames for their lexical content
- …