12,793 research outputs found

    Latency Bounds of Packet-Based Fronthaul for Cloud-RAN with Functionality Split

    Get PDF
    The emerging Cloud-RAN architecture within the fifth generation (5G) of wireless networks plays a vital role in enabling higher flexibility and granularity. On the other hand, Cloud-RAN architecture introduces an additional link between the central, cloudified unit and the distributed radio unit, namely fronthaul (FH). Therefore, the foreseen reliability and latency for 5G services should also be provisioned over the FH link. In this paper, focusing on Ethernet as FH, we present a reliable packet-based FH communication and demonstrate the upper and lower bounds of latency that can be offered. These bounds yield insights into the trade-off between reliability and latency, and enable the architecture design through choice of splitting point, focusing on high layer split between PDCP and RLC and low layer split between MAC and PHY, under different FH bandwidth and traffic properties. Presented model is then analyzed both numerically and through simulation, with two classes of 5G services that are ultra reliable low latency (URLL) and enhanced mobile broadband (eMBB).Comment: 6 pages, 7 figures, 3 tables, conference paper (ICC19

    High speed all optical networks

    Get PDF
    An inherent problem of conventional point-to-point wide area network (WAN) architectures is that they cannot translate optical transmission bandwidth into comparable user available throughput due to the limiting electronic processing speed of the switching nodes. The first solution to wavelength division multiplexing (WDM) based WAN networks that overcomes this limitation is presented. The proposed Lightnet architecture takes into account the idiosyncrasies of WDM switching/transmission leading to an efficient and pragmatic solution. The Lightnet architecture trades the ample WDM bandwidth for a reduction in the number of processing stages and a simplification of each switching stage, leading to drastically increased effective network throughputs. The principle of the Lightnet architecture is the construction and use of virtual topology networks, embedded in the original network in the wavelength domain. For this construction Lightnets utilize the new concept of lightpaths which constitute the links of the virtual topology. Lightpaths are all-optical, multihop, paths in the network that allow data to be switched through intermediate nodes using high throughput passive optical switches. The use of the virtual topologies and the associated switching design introduce a number of new ideas, which are discussed in detail

    Analysis and optimization of highly reliable systems

    Get PDF
    In the field of network design, the survivability property enables the network to maintain a certain level of network connectivity and quality of service under failure conditions. In this thesis, survivability aspects of communication systems are studied. Aspects of reliability and vulnerability of network design are also addressed. The contributions are three-fold. First, a Hop Constrained node Survivable Network Design Problem (HCSNDP) with optional (Steiner) nodes is modelled. This kind of problems are N P-Hard. An exact integer linear model is built, focused on networks represented by graphs without rooted demands, considering costs in arcs and in Steiner nodes. In addition to the exact model, the calculation of lower and upper bounds to the optimal solution is included. Models were tested over several graphs and instances, in order to validate it in cases with known solution. An Approximation Algorithm is also developed in order to address a particular case of SNDP: the Two Node Survivable Star Problem (2NCSP) with optional nodes. This problem belongs to the class of N P-Hard computational problems too. Second, the research is focused on cascading failures and target/random attacks. The Graph Fragmentation Problem (GFP) is the result of a worst case analysis of a random attack. A fixed number of individuals for protection can be chosen, and a non-protected target node immediately destroys all reachable nodes. The goal is to minimize the expected number of destroyed nodes in the network. This problem belongs to the N P-Hard class. A mathematical programming formulation is introduced and exact resolution for small instances as well as lower and upper bounds to the optimal solution. In addition to exact methods, we address the GFP by several approaches: metaheuristics, approximation algorithms, polytime methods for specific instances and exact methods in exponential time. Finally, the concept of separability in stochastic binary systems is here introduced. Stochastic Binary Systems (SBS) represent a mathematical model of a multi-component on-off system subject to independent failures. The reliability evaluation of an SBS belongs to the N P-Hard class. Therefore, we fully characterize separable systems using Han-Banach separation theorem for convex sets. Using this new concept of separable systems and Markov inequality, reliability bounds are provided for arbitrary SBS

    Tragedy of the Regulatory Commons: LightSquared and the Missing Spectrum Rights

    Get PDF
    The endemic underuse of radio spectrum constitutes a tragedy of the regulatory commons. Like other common interest tragedies, the outcome results from a legal or market structure that prevents economic actors from executing socially efficient bargains. In wireless markets, innovative applications often provoke claims by incumbent radio users that the new traffic will interfere with existing services. Sometimes these concerns are mitigated via market transactions, a la “Coasian bargaining.” Other times, however, solutions cannot be found even when social gains dominate the cost of spillovers. In the recent “LightSquared debacle,” such spectrum allocation failure played out. GPS interests that access frequencies adjacent to the band hosting LightSquared’s new nationwide mobile network complained that the wireless entrant would harm the operation of locational devices. Based on these complaints, regulators then killed LightSquared’s planned 4G network. Conservative estimates placed the prospective 4G consumer gains at least an order of magnitude above GPS losses. “Win win” bargains were theoretically available, fixing GPS vulnerabilities while welcoming the highly valuable wireless innovation. Yet transaction costs—largely caused by policy choices to issue limited and highly fragmented spectrum usage rights (here in the GPS band)—proved prohibitive. This episode provides a template for understanding market and non-market failure in radio spectrum allocation

    On the Potential of NoC Virtualization for Multicore Chips

    Full text link

    Contrasting Views of Complexity and Their Implications For Network-Centric Infrastructures

    Get PDF
    There exists a widely recognized need to better understand and manage complex “systems of systems,” ranging from biology, ecology, and medicine to network-centric technologies. This is motivating the search for universal laws of highly evolved systems and driving demand for new mathematics and methods that are consistent, integrative, and predictive. However, the theoretical frameworks available today are not merely fragmented but sometimes contradictory and incompatible. We argue that complexity arises in highly evolved biological and technological systems primarily to provide mechanisms to create robustness. However, this complexity itself can be a source of new fragility, leading to “robust yet fragile” tradeoffs in system design. We focus on the role of robustness and architecture in networked infrastructures, and we highlight recent advances in the theory of distributed control driven by network technologies. This view of complexity in highly organized technological and biological systems is fundamentally different from the dominant perspective in the mainstream sciences, which downplays function, constraints, and tradeoffs, and tends to minimize the role of organization and design

    Building a generalized distributed system model

    Get PDF
    The key elements in the second year (1991-92) of our project are: (1) implementation of the distributed system prototype; (2) successful passing of the candidacy examination and a PhD proposal acceptance by the funded student; (3) design of storage efficient schemes for replicated distributed systems; and (4) modeling of gracefully degrading reliable computing systems. In the third year of the project (1992-93), we propose to: (1) complete the testing of the prototype; (2) enhance the functionality of the modules by enabling the experimentation with more complex protocols; (3) use the prototype to verify the theoretically predicted performance of locking protocols, etc.; and (4) work on issues related to real-time distributed systems. This should result in efficient protocols for these systems

    On the variance of the number of occupied boxes

    Full text link
    We consider the occupancy problem where balls are thrown independently at infinitely many boxes with fixed positive frequencies. It is well known that the random number of boxes occupied by the first n balls is asymptotically normal if its variance V_n tends to infinity. In this work, we mainly focus on the opposite case where V_n is bounded, and derive a simple necessary and sufficient condition for convergence of V_n to a finite limit, thus settling a long-standing question raised by Karlin in the seminal paper of 1967. One striking consequence of our result is that the possible limit may only be a positive integer number. Some new conditions for other types of behavior of the variance, like boundedness or convergence to infinity, are also obtained. The proofs are based on the poissonization techniques.Comment: 34 page
    • 

    corecore