127,993 research outputs found
Confidence Sets in Time-Series Filtering
The problem of filtering of finite-alphabet stationary ergodic time series is
considered. A method for constructing a confidence set for the (unknown) signal
is proposed, such that the resulting set has the following properties: First,
it includes the unknown signal with probability , where is a
parameter supplied to the filter. Second, the size of the confidence sets grows
exponentially with the rate that is asymptotically equal to the conditional
entropy of the signal given the data. Moreover, it is shown that this rate is
optimal.Comment: some of the results were reported at ISIT2011, St. Petersburg,
Russia, pp. 2436-243
On planetary mass determination in the case of super-Earths orbiting active stars. The case of the CoRoT-7 system
This investigation uses the excellent HARPS radial velocity measurements of
CoRoT-7 to re-determine the planet masses and to explore techniques able to
determine mass and elements of planets discovered around active stars when the
relative variation of the radial velocity due to the star activity cannot be
considered as just noise and can exceed the variation due to the planets. The
main technique used here is a self-consistent version of the high-pass filter
used by Queloz et al. (2009) in the first mass determination of CoRoT-7b and
CoRoT-7c. The results are compared to those given by two alternative
techniques: (1) The approach proposed by Hatzes et al. (2010) using only those
nights in which 2 or 3 observations were done; (2) A pure Fourier analysis. In
all cases, the eccentricities are taken equal to zero as indicated by the study
of the tidal evolution of the system; the periods are also kept fixed at the
values given by Queloz et al. Only the observations done in the time interval
BJD 2,454,847 - 873 are used because they include many nights with multiple
observations; otherwise it is not possible to separate the effects of the
rotation fourth harmonic (5.91d = Prot/4) from the alias of the orbital period
of CoRoT-7b (0.853585 d). The results of the various approaches are combined to
give for the planet masses the values 8.0 \pm 1.2 MEarth for CoRoT-7b and 13.6
\pm 1.4 MEarth for CoRoT 7c. An estimation of the variation of the radial
velocity of the star due to its activity is also given.The results obtained
with 3 different approaches agree to give masses larger than those in previous
determinations. From the existing internal structure models they indicate that
CoRoT-7b is a much denser super-Earth. The bulk density is 11 \pm 3.5 g.cm-3 .
CoRoT-7b may be rocky with a large iron core.Comment: 12 pages, 11 figure
Online Robot Introspection via Wrench-based Action Grammars
Robotic failure is all too common in unstructured robot tasks. Despite
well-designed controllers, robots often fail due to unexpected events. How do
robots measure unexpected events? Many do not. Most robots are driven by the
sense-plan act paradigm, however more recently robots are undergoing a
sense-plan-act-verify paradigm. In this work, we present a principled
methodology to bootstrap online robot introspection for contact tasks. In
effect, we are trying to enable the robot to answer the question: what did I
do? Is my behavior as expected or not? To this end, we analyze noisy wrench
data and postulate that the latter inherently contains patterns that can be
effectively represented by a vocabulary. The vocabulary is generated by
segmenting and encoding the data. When the wrench information represents a
sequence of sub-tasks, we can think of the vocabulary forming a sentence (set
of words with grammar rules) for a given sub-task; allowing the latter to be
uniquely represented. The grammar, which can also include unexpected events,
was classified in offline and online scenarios as well as for simulated and
real robot experiments. Multiclass Support Vector Machines (SVMs) were used
offline, while online probabilistic SVMs were are used to give temporal
confidence to the introspection result. The contribution of our work is the
presentation of a generalizable online semantic scheme that enables a robot to
understand its high-level state whether nominal or abnormal. It is shown to
work in offline and online scenarios for a particularly challenging contact
task: snap assemblies. We perform the snap assembly in one-arm simulated and
real one-arm experiments and a simulated two-arm experiment. This verification
mechanism can be used by high-level planners or reasoning systems to enable
intelligent failure recovery or determine the next most optima manipulation
skill to be used.Comment: arXiv admin note: substantial text overlap with arXiv:1609.0494
Linking objective and subjective modeling in engineering design through arc-elastic dominance
Engineering design in mechanics is a complex activity taking into account both objective modeling processes derived from physical analysis and designers’ subjective reasoning. This paper introduces arc-elastic dominance as a suitable concept for ranking design solutions according to a combination of objective and subjective models. Objective models lead to the aggregation of information derived from physics, economics or eco-environmental analysis into a performance indicator. Subjective models result in a confidence indicator for the solutions’ feasibility. Arc-elastic dominant design solutions achieve an optimal compromise between gain in performance and degradation in confidence. Due to the definition of arc-elasticity, this compromise value is expressive and easy for designers to interpret despite the difference in the nature of the objective and subjective models. From the investigation of arc-elasticity mathematical properties, a filtering algorithm of Pareto-efficient solutions is proposed and illustrated through a design knowledge modeling framework. This framework notably takes into account Harrington’s desirability functions and Derringer’s aggregation method. It is carried out through the re-design of a geothermal air conditioning system
Improved Noisy Student Training for Automatic Speech Recognition
Recently, a semi-supervised learning method known as "noisy student training"
has been shown to improve image classification performance of deep networks
significantly. Noisy student training is an iterative self-training method that
leverages augmentation to improve network performance. In this work, we adapt
and improve noisy student training for automatic speech recognition, employing
(adaptive) SpecAugment as the augmentation method. We find effective methods to
filter, balance and augment the data generated in between self-training
iterations. By doing so, we are able to obtain word error rates (WERs)
4.2%/8.6% on the clean/noisy LibriSpeech test sets by only using the clean 100h
subset of LibriSpeech as the supervised set and the rest (860h) as the
unlabeled set. Furthermore, we are able to achieve WERs 1.7%/3.4% on the
clean/noisy LibriSpeech test sets by using the unlab-60k subset of LibriLight
as the unlabeled set for LibriSpeech 960h. We are thus able to improve upon the
previous state-of-the-art clean/noisy test WERs achieved on LibriSpeech 100h
(4.74%/12.20%) and LibriSpeech (1.9%/4.1%).Comment: 5 pages, 5 figures, 4 tables; v2: minor revisions, reference adde
- …