25,705 research outputs found

    Probing the Information Encoded in X-vectors

    Full text link
    Deep neural network based speaker embeddings, such as x-vectors, have been shown to perform well in text-independent speaker recognition/verification tasks. In this paper, we use simple classifiers to investigate the contents encoded by x-vector embeddings. We probe these embeddings for information related to the speaker, channel, transcription (sentence, words, phones), and meta information about the utterance (duration and augmentation type), and compare these with the information encoded by i-vectors across a varying number of dimensions. We also study the effect of data augmentation during extractor training on the information captured by x-vectors. Experiments on the RedDots data set show that x-vectors capture spoken content and channel-related information, while performing well on speaker verification tasks.Comment: Accepted at IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU) 201

    Tight Cell Probe Bounds for Succinct Boolean Matrix-Vector Multiplication

    Full text link
    The conjectured hardness of Boolean matrix-vector multiplication has been used with great success to prove conditional lower bounds for numerous important data structure problems, see Henzinger et al. [STOC'15]. In recent work, Larsen and Williams [SODA'17] attacked the problem from the upper bound side and gave a surprising cell probe data structure (that is, we only charge for memory accesses, while computation is free). Their cell probe data structure answers queries in O~(n7/4)\tilde{O}(n^{7/4}) time and is succinct in the sense that it stores the input matrix in read-only memory, plus an additional O~(n7/4)\tilde{O}(n^{7/4}) bits on the side. In this paper, we essentially settle the cell probe complexity of succinct Boolean matrix-vector multiplication. We present a new cell probe data structure with query time O~(n3/2)\tilde{O}(n^{3/2}) storing just O~(n3/2)\tilde{O}(n^{3/2}) bits on the side. We then complement our data structure with a lower bound showing that any data structure storing rr bits on the side, with n<r<n2n < r < n^2 must have query time tt satisfying tr=Ω~(n3)t r = \tilde{\Omega}(n^3). For rnr \leq n, any data structure must have t=Ω~(n2)t = \tilde{\Omega}(n^2). Since lower bounds in the cell probe model also apply to classic word-RAM data structures, the lower bounds naturally carry over. We also prove similar lower bounds for matrix-vector multiplication over F2\mathbb{F}_2

    Compressing Sparse Sequences under Local Decodability Constraints

    Full text link
    We consider a variable-length source coding problem subject to local decodability constraints. In particular, we investigate the blocklength scaling behavior attainable by encodings of rr-sparse binary sequences, under the constraint that any source bit can be correctly decoded upon probing at most dd codeword bits. We consider both adaptive and non-adaptive access models, and derive upper and lower bounds that often coincide up to constant factors. Notably, such a characterization for the fixed-blocklength analog of our problem remains unknown, despite considerable research over the last three decades. Connections to communication complexity are also briefly discussed.Comment: 8 pages, 1 figure. First five pages to appear in 2015 International Symposium on Information Theory. This version contains supplementary materia

    How to Counteract Systematic Errors in Quantum State Transfer

    Full text link
    In the absence of errors, the dynamics of a spin chain, with a suitably engineered local Hamiltonian, allow the perfect, coherent transfer of a quantum state over large distances. Here, we propose encoding and decoding procedures to recover perfectly from low rates of systematic errors. The encoding and decoding regions, located at opposite ends of the chain, are small compared to the length of the chain, growing linearly with the size of the error. We also describe how these errors can be identified, again by only acting on the encoding and decoding regions.Comment: 16 pages, 1 figur
    corecore