2,855 research outputs found

    Diversity-Multiplexing Tradeoff of Asynchronous Cooperative Diversity in Wireless Networks

    Full text link
    Synchronization of relay nodes is an important and critical issue in exploiting cooperative diversity in wireless networks. In this paper, two asynchronous cooperative diversity schemes are proposed, namely, distributed delay diversity and asynchronous space-time coded cooperative diversity schemes. In terms of the overall diversity-multiplexing (DM) tradeoff function, we show that the proposed independent coding based distributed delay diversity and asynchronous space-time coded cooperative diversity schemes achieve the same performance as the synchronous space-time coded approach which requires an accurate symbol-level timing synchronization to ensure signals arriving at the destination from different relay nodes are perfectly synchronized. This demonstrates diversity order is maintained even at the presence of asynchronism between relay node. Moreover, when all relay nodes succeed in decoding the source information, the asynchronous space-time coded approach is capable of achieving better DM-tradeoff than synchronous schemes and performs equivalently to transmitting information through a parallel fading channel as far as the DM-tradeoff is concerned. Our results suggest the benefits of fully exploiting the space-time degrees of freedom in multiple antenna systems by employing asynchronous space-time codes even in a frequency flat fading channel. In addition, it is shown asynchronous space-time coded systems are able to achieve higher mutual information than synchronous space-time coded systems for any finite signal-to-noise-ratio (SNR) when properly selected baseband waveforms are employed

    Optimality of entropic uncertainty relations

    Get PDF
    The entropic uncertainty relation proven by Maassen and Uffink for arbitrary pairs of two observables is known to be non-optimal. Here, we call an uncertainty relation optimal, if the lower bound can be attained for any value of either of the corresponding uncertainties. In this work we establish optimal uncertainty relations by characterising the optimal lower bound in scenarios similar to the Maassen-Uffink type. We disprove a conjecture by Englert et al. and generalise various previous results. However, we are still far from a complete understanding and, based on numerical investigation and analytical results in small dimension, we present a number of conjectures.Comment: 24 pages, 10 figure

    A note on weighted distributed match-making

    Get PDF

    Information theoretic treatment of tripartite systems and quantum channels

    Full text link
    A Holevo measure is used to discuss how much information about a given POVM on system aa is present in another system bb, and how this influences the presence or absence of information about a different POVM on aa in a third system cc. The main goal is to extend information theorems for mutually unbiased bases or general bases to arbitrary POVMs, and especially to generalize "all-or-nothing" theorems about information located in tripartite systems to the case of \emph{partial information}, in the form of quantitative inequalities. Some of the inequalities can be viewed as entropic uncertainty relations that apply in the presence of quantum side information, as in recent work by Berta et al. [Nature Physics 6, 659 (2010)]. All of the results also apply to quantum channels: e.g., if \EC accurately transmits certain POVMs, the complementary channel \FC will necessarily be noisy for certain other POVMs. While the inequalities are valid for mixed states of tripartite systems, restricting to pure states leads to the basis-invariance of the difference between the information about aa contained in bb and cc.Comment: 21 pages. An earlier version of this paper attempted to prove our main uncertainty relation, Theorem 5, using the achievability of the Holevo quantity in a coding task, an approach that ultimately failed because it did not account for locking of classical correlations, e.g. see [DiVincenzo et al. PRL. 92, 067902 (2004)]. In the latest version, we use a very different approach to prove Theorem

    Interference Mitigation in Large Random Wireless Networks

    Full text link
    A central problem in the operation of large wireless networks is how to deal with interference -- the unwanted signals being sent by transmitters that a receiver is not interested in. This thesis looks at ways of combating such interference. In Chapters 1 and 2, we outline the necessary information and communication theory background, including the concept of capacity. We also include an overview of a new set of schemes for dealing with interference known as interference alignment, paying special attention to a channel-state-based strategy called ergodic interference alignment. In Chapter 3, we consider the operation of large regular and random networks by treating interference as background noise. We consider the local performance of a single node, and the global performance of a very large network. In Chapter 4, we use ergodic interference alignment to derive the asymptotic sum-capacity of large random dense networks. These networks are derived from a physical model of node placement where signal strength decays over the distance between transmitters and receivers. (See also arXiv:1002.0235 and arXiv:0907.5165.) In Chapter 5, we look at methods of reducing the long time delays incurred by ergodic interference alignment. We analyse the tradeoff between reducing delay and lowering the communication rate. (See also arXiv:1004.0208.) In Chapter 6, we outline a problem that is equivalent to the problem of pooled group testing for defective items. We then present some new work that uses information theoretic techniques to attack group testing. We introduce for the first time the concept of the group testing channel, which allows for modelling of a wide range of statistical error models for testing. We derive new results on the number of tests required to accurately detect defective items, including when using sequential `adaptive' tests.Comment: PhD thesis, University of Bristol, 201

    Optimal Parochialism: The Dynamics of Trust and Exclusion in Networks

    Get PDF
    Networks such as ethnic credit associations, close-knit residential neighborhoods, 'old boy' networks, and ethnically linked businesses play an important role in economic life but have been little studied by economists. These networks are often supported by cultural distinctions between insiders and outsiders and engage in exclusionary practices which we call parochialism. We provide an economic analysis of parochial networks in which the losses incurred by not trading with outsiders are offset by an enhanced ability to enforce informal contracts by fostering trust among insiders. We first model one-shot social interactions among self-regarding agents, demonstrating that trust (i.e., cooperating without using information about one's trading partner) is a best response in a mixed-strategy Nash equilibrium if the quality of information about one's partner is sufficiently high. We show that since larger networks have lower quality information about specific individuals and greater trading opportunities, there may be an optimal (payoff-maximizing) network size. We then model the growth and decline of networks, as well as their equilibrium size and number. We show that in the absence of parochialism, networks may not exist, and the appropriate level of parochialism may implement an optimal network size. Finally, we explore the welfare implications and reasons for the evolutionary success of exclusion on parochial and other grounds.

    Time Versus Cost Tradeoffs for Deterministic Rendezvous in Networks

    Full text link
    Two mobile agents, starting from different nodes of a network at possibly different times, have to meet at the same node. This problem is known as rendezvous\mathit{rendezvous}. Agents move in synchronous rounds. Each agent has a distinct integer label from the set {1,,L}\{1,\dots,L\}. Two main efficiency measures of rendezvous are its time\mathit{time} (the number of rounds until the meeting) and its cost\mathit{cost} (the total number of edge traversals). We investigate tradeoffs between these two measures. A natural benchmark for both time and cost of rendezvous in a network is the number of edge traversals needed for visiting all nodes of the network, called the exploration time. Hence we express the time and cost of rendezvous as functions of an upper bound EE on the time of exploration (where EE and a corresponding exploration procedure are known to both agents) and of the size LL of the label space. We present two natural rendezvous algorithms. Algorithm Cheap\mathtt{Cheap} has cost O(E)O(E) (and, in fact, a version of this algorithm for the model where the agents start simultaneously has cost exactly EE) and time O(EL)O(EL). Algorithm Fast\mathtt{Fast} has both time and cost O(ElogL)O(E\log L). Our main contributions are lower bounds showing that, perhaps surprisingly, these two algorithms capture the tradeoffs between time and cost of rendezvous almost tightly. We show that any deterministic rendezvous algorithm of cost asymptotically EE (i.e., of cost E+o(E)E+o(E)) must have time Ω(EL)\Omega(EL). On the other hand, we show that any deterministic rendezvous algorithm with time complexity O(ElogL)O(E\log L) must have cost Ω(ElogL)\Omega (E\log L)

    Better Sooner Rather Than Later

    Full text link
    This article unifies and generalizes fundamental results related to nn-process asynchronous crash-prone distributed computing. More precisely, it proves that for every 0kn0\leq k \leq n, assuming that process failures occur only before the number of participating processes bypasses a predefined threshold that equals nkn-k (a participating process is a process that has executed at least one statement of its code), an asynchronous algorithm exists that solves consensus for nn processes in the presence of ff crash failures if and only if fkf \leq k. In a very simple and interesting way, the "extreme" case k=0k=0 boils down to the celebrated FLP impossibility result (1985, 1987). Moreover, the second extreme case, namely k=nk=n, captures the celebrated mutual exclusion result by E.W. Dijkstra (1965) that states that mutual exclusion can be solved for nn processes in an asynchronous read/write shared memory system where any number of processes may crash (but only) before starting to participate in the algorithm (that is, participation is not required, but once a process starts participating it may not fail). More generally, the possibility/impossibility stated above demonstrates that more failures can be tolerated when they occur earlier in the computation (hence the title).Comment: 10 page

    Physical limits on cooperative protein-DNA binding and the kinetics of combinatorial transcription regulation

    Get PDF
    Much of the complexity observed in gene regulation originates from cooperative protein-DNA binding. While studies of the target search of proteins for their specific binding sites on the DNA have revealed design principles for the quantitative characteristics of protein-DNA interactions, no such principles are known for the cooperative interactions between DNA-binding proteins. We consider a simple theoretical model for two interacting transcription factor (TF) species, searching for and binding to two adjacent target sites hidden in the genomic background. We study the kinetic competition of a dimer search pathway and a monomer search pathway, as well as the steady-state regulation function mediated by the two TFs over a broad range of TF-TF interaction strengths. Using a transcriptional AND-logic as exemplary functional context, we identify the functionally desirable regime for the interaction. We find that both weak and very strong TF-TF interactions are favorable, albeit with different characteristics. However, there is also an unfavorable regime of intermediate interactions where the genetic response is prohibitively slow.Comment: manuscript and supplementary material combined into a single document; to be published in Biophysical Journa

    Testing Bounded Rationality Against Full Rationality in Job Changing Behavior

    Get PDF
    In this paper I question the hypothesis of full rationality in the context of job changing behaviour, via simple econometric explorations on microdata drawn from WHIP (Worker Histories Italian Panel). Workers’ performance is compared at the end of a three-year time window that starts when choices are expressed, under the accepted notion that the main driving forces of job change are future real wages and expected job quality. Bounded rationality suggests that individuals will search for new options capable to attain “satisfactory” targets (aspirations levels, standards, reference points, norms), based on conditions prevailing in their own local environments. The empirical strategy consists of appropriately defining such environments (cells) and observing the ex-post individual performance vis-à-vis their targets, in terms of degree of dispersion, clustering and mobility within and between cells. Under full rationality the following are to be expected: - large dispersion around the targets; - clustering in the vicinity of the theoretical efficiency frontier; - high inter-cell mobility; None of the above expectations are confirmed in this exploration. My conclusion is that workers appear to behave according to principles of rationality different from those of “full rationality” assumed in the vast majority of contemporary empirical (and theoretical) studies. The idea of “bounded rationality” à la Simon seems to provide a better fit to our observations.
    corecore