31,753 research outputs found

    Reduced Dimensional Optimal Vector Linear Index Codes for Index Coding Problems with Symmetric Neighboring and Consecutive Side-information

    Full text link
    A single unicast index coding problem (SUICP) with symmetric neighboring and consecutive side-information (SNCS) has KK messages and KK receivers, the kkth receiver RkR_k wanting the kkth message xkx_k and having the side-information Kk={xkU,,xk2,xk1}{xk+1,xk+2,,xk+D}\mathcal{K}_k=\{x_{k-U},\dots,x_{k-2},x_{k-1}\}\cup\{x_{k+1}, x_{k+2},\dots,x_{k+D}\}. The single unicast index coding problem with symmetric neighboring and consecutive side-information, SUICP(SNCS), is motivated by topological interference management problems in wireless communication networks. Maleki, Cadambe and Jafar obtained the symmetric capacity of this SUICP(SNCS) and proposed optimal length codes by using Vandermonde matrices. In our earlier work, we gave optimal length (U+1)(U+1)-dimensional vector linear index codes for SUICP(SNCS) satisfying some conditions on K,DK,D and UU \cite{VaR1}. In this paper, for SUICP(SNCS) with arbitrary K,DK,D and UU, we construct optimal length U+1gcd(K,DU,U+1)\frac{U+1}{\text{gcd}(K,D-U,U+1)}-dimensional vector linear index codes. We prove that the constructed vector linear index code is of minimal dimension if gcd(KD+U,U+1)\text{gcd}(K-D+U,U+1) is equal to gcd(K,DU,U+1)\text{gcd}(K,D-U,U+1). The proposed construction gives optimal length scalar linear index codes for the SUICP(SNCS) if (U+1)(U+1) divides both KK and DUD-U. The proposed construction is independent of field size and works over every field. We give a low-complexity decoding for the SUICP(SNCS). By using the proposed decoding method, every receiver is able to decode its wanted message symbol by simply adding some index code symbols (broadcast symbols).Comment: 13 pages, 1 figure and 5 table

    Optimal Error Correcting Delivery Scheme for Coded Caching with Symmetric Batch Prefetching

    Full text link
    Coded caching is used to reduce network congestion during peak hours. A single server is connected to a set of users through a bottleneck link, which generally is assumed to be error-free. During non-peak hours, all the users have full access to the files and they fill their local cache with portions of the files available. During delivery phase, each user requests a file and the server delivers coded transmissions to meet the demands taking into consideration their cache contents. In this paper we assume that the shared link is error prone. A new delivery scheme is required to meet the demands of each user even after receiving finite number of transmissions in error. We characterize the minimum average rate and minimum peak rate for this problem. We find closed form expressions of these rates for a particular caching scheme namely \textit{symmetric batch prefetching}. We also propose an optimal error correcting delivery scheme for coded caching problem with symmetric batch prefetching.Comment: 9 pages and 4 figure

    On Approximating the Sum-Rate for Multiple-Unicasts

    Full text link
    We study upper bounds on the sum-rate of multiple-unicasts. We approximate the Generalized Network Sharing Bound (GNS cut) of the multiple-unicasts network coding problem with kk independent sources. Our approximation algorithm runs in polynomial time and yields an upper bound on the joint source entropy rate, which is within an O(log2k)O(\log^2 k) factor from the GNS cut. It further yields a vector-linear network code that achieves joint source entropy rate within an O(log2k)O(\log^2 k) factor from the GNS cut, but \emph{not} with independent sources: the code induces a correlation pattern among the sources. Our second contribution is establishing a separation result for vector-linear network codes: for any given field F\mathbb{F} there exist networks for which the optimum sum-rate supported by vector-linear codes over F\mathbb{F} for independent sources can be multiplicatively separated by a factor of k1δk^{1-\delta}, for any constant δ>0{\delta>0}, from the optimum joint entropy rate supported by a code that allows correlation between sources. Finally, we establish a similar separation result for the asymmetric optimum vector-linear sum-rates achieved over two distinct fields Fp\mathbb{F}_{p} and Fq\mathbb{F}_{q} for independent sources, revealing that the choice of field can heavily impact the performance of a linear network code.Comment: 10 pages; Shorter version appeared at ISIT (International Symposium on Information Theory) 2015; some typos correcte
    corecore