3,393 research outputs found
Collision Helps - Algebraic Collision Recovery for Wireless Erasure Networks
Current medium access control mechanisms are based on collision avoidance and
collided packets are discarded. The recent work on ZigZag decoding departs from
this approach by recovering the original packets from multiple collisions. In
this paper, we present an algebraic representation of collisions which allows
us to view each collision as a linear combination of the original packets. The
transmitted, colliding packets may themselves be a coded version of the
original packets.
We propose a new acknowledgment (ACK) mechanism for collisions based on the
idea that if a set of packets collide, the receiver can afford to ACK exactly
one of them and still decode all the packets eventually. We analytically
compare delay and throughput performance of such collision recovery schemes
with other collision avoidance approaches in the context of a single hop
wireless erasure network. In the multiple receiver case, the broadcast
constraint calls for combining collision recovery methods with network coding
across packets at the sender. From the delay perspective, our scheme, without
any coordination, outperforms not only a ALOHA-type random access mechanisms,
but also centralized scheduling. For the case of streaming arrivals, we propose
a priority-based ACK mechanism and show that its stability region coincides
with the cut-set bound of the packet erasure network
Coded Cooperative Data Exchange for a Secret Key
We consider a coded cooperative data exchange problem with the goal of
generating a secret key. Specifically, we investigate the number of public
transmissions required for a set of clients to agree on a secret key with
probability one, subject to the constraint that it remains private from an
eavesdropper.
Although the problems are closely related, we prove that secret key
generation with fewest number of linear transmissions is NP-hard, while it is
known that the analogous problem in traditional cooperative data exchange can
be solved in polynomial time. In doing this, we completely characterize the
best possible performance of linear coding schemes, and also prove that linear
codes can be strictly suboptimal. Finally, we extend the single-key results to
characterize the minimum number of public transmissions required to generate a
desired integer number of statistically independent secret keys.Comment: Full version of a paper that appeared at ISIT 2014. 19 pages, 2
figure
Adaptive speaker diarization of broadcast news based on factor analysis
The introduction of factor analysis techniques in a speaker diarization system enhances its performance by facilitating the use of speaker specific information, by improving the suppression of nuisance factors such as phonetic content, and by facilitating various forms of adaptation. This paper describes a state-of-the-art iVector-based diarization system which employs factor analysis and adaptation on all levels. The diarization modules relevant for this work are: the speaker segmentation which searches for speaker boundaries and the speaker clustering which aims at grouping speech segments of the same speaker. The speaker segmentation relies on speaker factors which are extracted on a frame-by-frame basis using eigenvoices. We incorporate soft voice activity detection in this extraction process as the speaker change detection should be based on speaker information only and we want it to disregard the non-speech frames by applying speech posteriors. Potential speaker boundaries are inserted at positions where rapid changes in speaker factors are witnessed. By employing Mahalanobis distances, the effect of the phonetic content can be further reduced, which results in more accurate speaker boundaries. This iVector-based segmentation significantly outperforms more common segmentation methods based on the Bayesian Information Criterion (BIC) or speech activity marks. The speaker clustering employs two-step Agglomerative Hierarchical Clustering (AHC): after initial BIC clustering, the second cluster stage is realized by either an iVector Probabilistic Linear Discriminant Analysis (PLDA) system or Cosine Distance Scoring (CDS) of extracted speaker factors. The segmentation system is made adaptive on a file-by-file basis by iterating the diarization process using eigenvoice matrices adapted (unsupervised) on the output of the previous iteration. Assuming that for most use cases material similar to the recording in question is readily available, unsupervised domain adaptation of the speaker clustering is possible as well. We obtain this by expanding the eigenvoice matrix used during speaker factor extraction for the CDS clustering stage with a small set of new eigenvoices that, in combination with the initial generic eigenvoices, models the recurring speakers and acoustic conditions more accurately. Experiments on the COST278 multilingual broadcast news database show the generation of significantly more accurate speaker boundaries by using adaptive speaker segmentation which also results in more accurate clustering. The obtained speaker error rate (SER) can be further reduced by another 13% relative to 7.4% via domain adaptation of the CDS clustering. (C) 2017 Elsevier Ltd. All rights reserved
BRIDGE: Byzantine-resilient Decentralized Gradient Descent
Decentralized optimization techniques are increasingly being used to learn
machine learning models from data distributed over multiple locations without
gathering the data at any one location. Unfortunately, methods that are
designed for faultless networks typically fail in the presence of node
failures. In particular, Byzantine failures---corresponding to the scenario in
which faulty/compromised nodes are allowed to arbitrarily deviate from an
agreed-upon protocol---are the hardest to safeguard against in decentralized
settings. This paper introduces a Byzantine-resilient decentralized gradient
descent (BRIDGE) method for decentralized learning that, when compared to
existing works, is more efficient and scalable in higher-dimensional settings
and that is deployable in networks having topologies that go beyond the star
topology. The main contributions of this work include theoretical analysis of
BRIDGE for strongly convex learning objectives and numerical experiments
demonstrating the efficacy of BRIDGE for both convex and nonconvex learning
tasks.Comment: 18 pages, 1 figure, 1 table; preprint of a conference pape
- …