871 research outputs found
Silent MST approximation for tiny memory
In network distributed computing, minimum spanning tree (MST) is one of the
key problems, and silent self-stabilization one of the most demanding
fault-tolerance properties. For this problem and this model, a polynomial-time
algorithm with memory is known for the state model. This is
memory optimal for weights in the classic range (where
is the size of the network). In this paper, we go below this
memory, using approximation and parametrized complexity.
More specifically, our contributions are two-fold. We introduce a second
parameter~, which is the space needed to encode a weight, and we design a
silent polynomial-time self-stabilizing algorithm, with space . In turn, this allows us to get an approximation algorithm for the problem,
with a trade-off between the approximation ratio of the solution and the space
used. For polynomial weights, this trade-off goes smoothly from memory for an -approximation, to memory for exact solutions,
with for example memory for a 2-approximation
Graph-Based Shape Analysis Beyond Context-Freeness
We develop a shape analysis for reasoning about relational properties of data
structures. Both the concrete and the abstract domain are represented by
hypergraphs. The analysis is parameterized by user-supplied indexed graph
grammars to guide concretization and abstraction. This novel extension of
context-free graph grammars is powerful enough to model complex data structures
such as balanced binary trees with parent pointers, while preserving most
desirable properties of context-free graph grammars. One strength of our
analysis is that no artifacts apart from grammars are required from the user;
it thus offers a high degree of automation. We implemented our analysis and
successfully applied it to various programs manipulating AVL trees,
(doubly-linked) lists, and combinations of both
Predicting the operability of damaged compressors using machine learning
Abstract
The application of machine learning to aerospace problems faces a particular challenge. For successful learning a large amount of good quality training data is required, typically tens of thousands of cases. However, due to the time and cost of experimental aerospace testing, this data is scarce. This paper shows that successful learning is possible with two novel techniques: The first technique is rapid testing. Over the last five years the Whittle Laboratory has developed a capability where rebuild and test times of a compressor stage now take 15 minutes instead of weeks. The second technique is to base machine learning on physical parameters, derived from engineering wisdom developed in industry over many decades.
The method is applied to the important industry problem of predicting the effect of blade damage on compressor operability. The current approach has high uncertainty, it is based on human judgement and correlation of a handful of experimental test cases. It is shown using 100 training cases and 25 test cases that the new method is able to predict the operability of damaged compressor stages with an accuracy of 2% in a 95% confidence interval; far better than is possible by even the most experienced compressor designers. Use of the method is also shown to generate new physical understanding, previously unknown by any of the experts involved in this work. Using this method in the future offers an exciting opportunity to generate understanding of previously intractable problems in aerospace.Aerospace Technology Institute
Rolls-Royce plc
Ten Years of Pedestrian Detection, What Have We Learned?
Paper-by-paper results make it easy to miss the forest for the trees.We
analyse the remarkable progress of the last decade by discussing the main ideas
explored in the 40+ detectors currently present in the Caltech pedestrian
detection benchmark. We observe that there exist three families of approaches,
all currently reaching similar detection quality. Based on our analysis, we
study the complementarity of the most promising ideas by combining multiple
published strategies. This new decision forest detector achieves the current
best known performance on the challenging Caltech-USA dataset.Comment: To appear in ECCV 2014 CVRSUAD workshop proceeding
Distributed Edge Connectivity in Sublinear Time
We present the first sublinear-time algorithm for a distributed
message-passing network sto compute its edge connectivity exactly in
the CONGEST model, as long as there are no parallel edges. Our algorithm takes
time to compute and a
cut of cardinality with high probability, where and are the
number of nodes and the diameter of the network, respectively, and
hides polylogarithmic factors. This running time is sublinear in (i.e.
) whenever is. Previous sublinear-time
distributed algorithms can solve this problem either (i) exactly only when
[Thurimella PODC'95; Pritchard, Thurimella, ACM
Trans. Algorithms'11; Nanongkai, Su, DISC'14] or (ii) approximately [Ghaffari,
Kuhn, DISC'13; Nanongkai, Su, DISC'14].
To achieve this we develop and combine several new techniques. First, we
design the first distributed algorithm that can compute a -edge connectivity
certificate for any in time .
Second, we show that by combining the recent distributed expander decomposition
technique of [Chang, Pettie, Zhang, SODA'19] with techniques from the
sequential deterministic edge connectivity algorithm of [Kawarabayashi, Thorup,
STOC'15], we can decompose the network into a sublinear number of clusters with
small average diameter and without any mincut separating a cluster (except the
`trivial' ones). Finally, by extending the tree packing technique from [Karger
STOC'96], we can find the minimum cut in time proportional to the number of
components. As a byproduct of this technique, we obtain an -time
algorithm for computing exact minimum cut for weighted graphs.Comment: Accepted at 51st ACM Symposium on Theory of Computing (STOC 2019
Estimation of the hydraulic parameters of unsaturated samples by electrical resistivity tomography
In situ and laboratory experiments have shown that electrical resistivity tomography (ERT) is an effective tool to image transient phenomena in soils. However, its application in quantifying soil hydraulic parameters has been limited. In this study, experiments of water inflow in unsaturated soil samples were conducted in an oedometer equipped to perform three-dimensional electrical measurements. Reconstructions of the electrical conductivity at different times confirmed the usefulness of ERT for monitoring the evolution of water content. The tomographic reconstructions were subsequently used in conjunction with a finite-element simulation to infer the water retention curve and the unsaturated hydraulic conductivity. The parameters estimated with ERT agree satisfactorily with those determined using established techniques, hence the proposed approach shows good potential for relatively fast characterisations. Similar experiments could be carried out on site to study the hydraulic behaviour of the entire soil deposi
EL-GAN: Embedding Loss Driven Generative Adversarial Networks for Lane Detection
Convolutional neural networks have been successfully applied to semantic
segmentation problems. However, there are many problems that are inherently not
pixel-wise classification problems but are nevertheless frequently formulated
as semantic segmentation. This ill-posed formulation consequently necessitates
hand-crafted scenario-specific and computationally expensive post-processing
methods to convert the per pixel probability maps to final desired outputs.
Generative adversarial networks (GANs) can be used to make the semantic
segmentation network output to be more realistic or better
structure-preserving, decreasing the dependency on potentially complex
post-processing. In this work, we propose EL-GAN: a GAN framework to mitigate
the discussed problem using an embedding loss. With EL-GAN, we discriminate
based on learned embeddings of both the labels and the prediction at the same
time. This results in more stable training due to having better discriminative
information, benefiting from seeing both `fake' and `real' predictions at the
same time. This substantially stabilizes the adversarial training process. We
use the TuSimple lane marking challenge to demonstrate that with our proposed
framework it is viable to overcome the inherent anomalies of posing it as a
semantic segmentation problem. Not only is the output considerably more similar
to the labels when compared to conventional methods, the subsequent
post-processing is also simpler and crosses the competitive 96% accuracy
threshold.Comment: 14 pages, 7 figure
Finite Automata for the Sub- and Superword Closure of CFLs: Descriptional and Computational Complexity
We answer two open questions by (Gruber, Holzer, Kutrib, 2009) on the
state-complexity of representing sub- or superword closures of context-free
grammars (CFGs): (1) We prove a (tight) upper bound of on
the size of nondeterministic finite automata (NFAs) representing the subword
closure of a CFG of size . (2) We present a family of CFGs for which the
minimal deterministic finite automata representing their subword closure
matches the upper-bound of following from (1).
Furthermore, we prove that the inequivalence problem for NFAs representing sub-
or superword-closed languages is only NP-complete as opposed to PSPACE-complete
for general NFAs. Finally, we extend our results into an approximation method
to attack inequivalence problems for CFGs
Single-Cell, Genome-wide Sequencing Identifies Clonal Somatic Copy-Number Variation in the Human Brain
SUMMARY De novo copy-number variants (CNVs) can cause neuropsychiatric disease, but the degree to which they occur somatically, and during development, is unknown. Single-cell whole-genome sequencing (WGS) in >200 single cells, including >160 neurons from three normal and two pathological human brains, sensitively identified germline trisomy of chromosome 18 but found most (≥95%) neurons in normal brain tissue to be euploid. Analysis of a patient with hemimegalencephaly (HMG) due to a somatic CNV of chromosome 1q found unexpected tetrasomy 1q in ~20% of neurons, suggesting that CNVs in a minority of cells can cause widespread brain dysfunction. Single-cell analysis identified large (>1 Mb) clonal CNVs in lymphoblasts and in single neurons from normal human brain tissue, suggesting that some CNVs occur during neurogenesis. Many neurons contained one or more large candidate private CNVs, including one at chromosome 15q13.2-13.3, a site of duplication in neuropsychiatric conditions. Large private and clonal somatic CNVs occur in normal and diseased human brains
Model Adaptation with Synthetic and Real Data for Semantic Dense Foggy Scene Understanding
This work addresses the problem of semantic scene understanding under dense
fog. Although considerable progress has been made in semantic scene
understanding, it is mainly related to clear-weather scenes. Extending
recognition methods to adverse weather conditions such as fog is crucial for
outdoor applications. In this paper, we propose a novel method, named
Curriculum Model Adaptation (CMAda), which gradually adapts a semantic
segmentation model from light synthetic fog to dense real fog in multiple
steps, using both synthetic and real foggy data. In addition, we present three
other main stand-alone contributions: 1) a novel method to add synthetic fog to
real, clear-weather scenes using semantic input; 2) a new fog density
estimator; 3) the Foggy Zurich dataset comprising real foggy images,
with pixel-level semantic annotations for images with dense fog. Our
experiments show that 1) our fog simulation slightly outperforms a
state-of-the-art competing simulation with respect to the task of semantic
foggy scene understanding (SFSU); 2) CMAda improves the performance of
state-of-the-art models for SFSU significantly by leveraging unlabeled real
foggy data. The datasets and code are publicly available.Comment: final version, ECCV 201
- …