26,063 research outputs found
The Optimisation of Stochastic Grammars to Enable Cost-Effective Probabilistic Structural Testing
The effectiveness of probabilistic structural testing depends on the characteristics of the probability distribution from which test inputs are sampled at random. Metaheuristic search has been shown to be a practical method of optimis- ing the characteristics of such distributions. However, the applicability of the existing search-based algorithm is lim- ited by the requirement that the software’s inputs must be a fixed number of numeric values. In this paper we relax this limitation by means of a new representation for the probability distribution. The repre- sentation is based on stochastic context-free grammars but incorporates two novel extensions: conditional production weights and the aggregation of terminal symbols represent- ing numeric values. We demonstrate that an algorithm which combines the new representation with hill-climbing search is able to effi- ciently derive probability distributions suitable for testing software with structurally-complex input domains
Adapting Sequence to Sequence models for Text Normalization in Social Media
Social media offer an abundant source of valuable raw data, however informal
writing can quickly become a bottleneck for many natural language processing
(NLP) tasks. Off-the-shelf tools are usually trained on formal text and cannot
explicitly handle noise found in short online posts. Moreover, the variety of
frequently occurring linguistic variations presents several challenges, even
for humans who might not be able to comprehend the meaning of such posts,
especially when they contain slang and abbreviations. Text Normalization aims
to transform online user-generated text to a canonical form. Current text
normalization systems rely on string or phonetic similarity and classification
models that work on a local fashion. We argue that processing contextual
information is crucial for this task and introduce a social media text
normalization hybrid word-character attention-based encoder-decoder model that
can serve as a pre-processing step for NLP applications to adapt to noisy text
in social media. Our character-based component is trained on synthetic
adversarial examples that are designed to capture errors commonly found in
online user-generated text. Experiments show that our model surpasses neural
architectures designed for text normalization and achieves comparable
performance with state-of-the-art related work.Comment: Accepted at the 13th International AAAI Conference on Web and Social
Media (ICWSM 2019
Feature-based time-series analysis
This work presents an introduction to feature-based time-series analysis. The
time series as a data type is first described, along with an overview of the
interdisciplinary time-series analysis literature. I then summarize the range
of feature-based representations for time series that have been developed to
aid interpretable insights into time-series structure. Particular emphasis is
given to emerging research that facilitates wide comparison of feature-based
representations that allow us to understand the properties of a time-series
dataset that make it suited to a particular feature-based representation or
analysis algorithm. The future of time-series analysis is likely to embrace
approaches that exploit machine learning methods to partially automate human
learning to aid understanding of the complex dynamical patterns in the time
series we measure from the world.Comment: 28 pages, 9 figure
General anesthesia reduces complexity and temporal asymmetry of the informational structures derived from neural recordings in Drosophila
We apply techniques from the field of computational mechanics to evaluate the
statistical complexity of neural recording data from fruit flies. First, we
connect statistical complexity to the flies' level of conscious arousal, which
is manipulated by general anesthesia (isoflurane). We show that the complexity
of even single channel time series data decreases under anesthesia. The
observed difference in complexity between the two states of conscious arousal
increases as higher orders of temporal correlations are taken into account. We
then go on to show that, in addition to reducing complexity, anesthesia also
modulates the informational structure between the forward- and reverse-time
neural signals. Specifically, using three distinct notions of temporal
asymmetry we show that anesthesia reduces temporal asymmetry on
information-theoretic and information-geometric grounds. In contrast to prior
work, our results show that: (1) Complexity differences can emerge at very
short timescales and across broad regions of the fly brain, thus heralding the
macroscopic state of anesthesia in a previously unforeseen manner, and (2) that
general anesthesia also modulates the temporal asymmetry of neural signals.
Together, our results demonstrate that anesthetized brains become both less
structured and more reversible.Comment: 14 pages, 6 figures. Comments welcome; Added time-reversal analysis,
updated discussion, new figures (Fig. 5 & Fig. 6) and Tables (Tab. 1
- …