1,417 research outputs found
Exploiting AWG Free Spectral Range Periodicity in Distributed Multicast Architectures
Modular optical switch architectures combining wavelength routing based on
arrayed waveguide grating (AWG) devices and multicasting based on star couplers
hold promise for flexibly addressing the exponentially growing traffic demands
in a cost- and power-efficient fashion. In a default switching scenario, an
input port of the AWG is connected to an output port via a single wavelength.
This can severely limit the capacity between broadcast domains, resulting in
interdomain traffic switching bottlenecks. In this paper, we examine the
possibility of resolving capacity bottlenecks by exploiting multiple AWG free
spectral ranges (FSRs), i.e., setting up multiple parallel connections between
each pair of broadcast domains. To this end, we introduce a multi-FSR
scheduling algorithm for interconnecting broadcast domains by fairly
distributing the wavelength resources among them. We develop a general-purpose
analytical framework to study the blocking probabilities in a multistage
switching scenario and compare our results with Monte Carlo simulations. Our
study points to significant improvements with a moderate increase in the number
of FSRs. We show that an FSR count beyond four results in diminishing returns.
Furthermore, to investigate the trade-offs between the network- and
physical-layer effects, we conduct a cross-layer analysis, taking into account
pulse amplitude modulation (PAM) and rate-adaptive forward error correction
(FEC). We illustrate how the effective bit rate per port increases with an
increase in the number of FSRs. %We also look at the advantages of an
impairment-aware scheduling strategy in a multi-FSR switching scenario
A Multi-Floor Arrayed Waveguide Grating Based Architecture with Grid Topology for Datacenter Networks
This paper proposes a grid topology based passive optical interconnect (POI) architecture that is composed of multiple floors of arrayed waveguide grating routers (AWGRs) to offer high connectivity and scalability for datacenter networks. In the proposed POI signal only needs to pass one AWGR, and thus can avoid the crosstalk accumulation and cascaded filtering effects, which exist in many existing POI architectures based on cascaded AWGRs. Meanwhile, due to high connectivity, the proposed grid topology based POI also has the potential advantage of high reliability. Simulation results validate the network performance. With a proper node degree, the proposed grid topology can achieve acceptable blocking probability. Besides, steady performance is kept when the number of floors increases, indicating good scalability of the proposed POI
When Task Scheduling Meets Flexible-bandwidth Optical Interconnects: A Cross-layer Resource Orchestration Design
We propose a cross-layer resource orchestration design for task scheduling in flexible-bandwidth optical data center networks. Results show the proposed design can achieve 8.2 Ă—, 1.9 Ă— and 4.8 Ă— reductions of request blocking probability, end-to-end delay and packet loss rate, compared with the baseline
Optical Networks and Interconnects
The rapid evolution of communication technologies such as 5G and beyond, rely
on optical networks to support the challenging and ambitious requirements that
include both capacity and reliability. This chapter begins by giving an
overview of the evolution of optical access networks, focusing on Passive
Optical Networks (PONs). The development of the different PON standards and
requirements aiming at longer reach, higher client count and delivered
bandwidth are presented. PON virtualization is also introduced as the
flexibility enabler. Triggered by the increase of bandwidth supported by access
and aggregation network segments, core networks have also evolved, as presented
in the second part of the chapter. Scaling the physical infrastructure requires
high investment and hence, operators are considering alternatives to optimize
the use of the existing capacity. This chapter introduces different planning
problems such as Routing and Spectrum Assignment problems, placement problems
for regenerators and wavelength converters, and how to offer resilience to
different failures. An overview of control and management is also provided.
Moreover, motivated by the increasing importance of data storage and data
processing, this chapter also addresses different aspects of optical data
center interconnects. Data centers have become critical infrastructure to
operate any service. They are also forced to take advantage of optical
technology in order to keep up with the growing capacity demand and power
consumption. This chapter gives an overview of different optical data center
network architectures as well as some expected directions to improve the
resource utilization and increase the network capacity
Recommended from our members
Reconfigurable Optically Interconnected Systems
With the immense growth of data consumption in today's data centers and high-performance computing systems driven by the constant influx of new applications, the network infrastructure supporting this demand is under increasing pressure to enable higher bandwidth, latency, and flexibility requirements. Optical interconnects, able to support high bandwidth wavelength division multiplexed signals with extreme energy efficiency, have become the basis for long-haul and metro-scale networks around the world, while photonic components are being rapidly integrated within rack and chip-scale systems. However, optical and photonic interconnects are not a direct replacement for electronic-based components. Rather, the integration of optical interconnects with electronic peripherals allows for unique functionalities that can improve the capacity, compute performance and flexibility of current state-of-the-art computing systems. This requires physical layer methodologies for their integration with electronic components, as well as system level control planes that incorporates the optical layer characteristics. This thesis explores various network architectures and the associated control plane, hardware infrastructure, and other supporting software modules needed to integrate silicon photonics and MEMS based optical switching into conventional datacom network systems ranging from intra-data center and high-performance computing systems to the metro-scale layer networks between data centers. In each of these systems, we demonstrate dynamic bandwidth steering and compute resource allocation capabilities to enable significant performance improvements. The key accomplishments of this thesis are as follows.
In Part 1, we present high-performance computing network architectures that integrate silicon photonic switches for optical bandwidth steering, enabling multiple reconfigurable topologies that results in significant system performance improvements. As high-performance systems rely on increased parallelism by scaling up to greater numbers of processor nodes, communication between these nodes grows rapidly and the interconnection network becomes a bottleneck to the overall performance of the system. It has been observed that many scientific applications operating on high-performance computing systems cause highly skewed traffic over the network, congesting only a small percentage of the total available links while other links are underutilized. This mismatch of the traffic and the bandwidth allocation of the physical layer network presents the opportunity to optimize the bandwidth resource utilization of the system by using silicon photonic switches to perform bandwidth steering. This allows the individual processors to perform at their maximum compute potential and thereby improving the overall system performance. We show various testbeds that integrates both microring resonator and Mach-Zehnder based silicon photonic switches within Dragonfly and Fat-Tree topology networks built with conventional
equipment, and demonstrate 30-60% reduction in execution time of real high-performance benchmark applications.
Part 2 presents a flexible network architecture and control plane that enables autonomous bandwidth steering and IT resource provisioning capabilities between metro-scale geographically distributed data centers. It uses a software-defined control plane to autonomously provision both network and IT resources to support different quality of service requirements and optimizes resource utilization under dynamically changing load variations. By actively monitoring both the bandwidth utilization of the network and CPU or memory resources of the end hosts, the control plane autonomously provisions background or dynamic connections with different levels of quality of service using optical MEMS switching, as well as initializing live migrations of virtual machines to consolidate or distribute workload. Together these functionalities provide flexibility and maximize efficiency in processing and transferring data, and enables energy and cost savings by scaling down the system when resources are not needed. An experimental testbed of three data center nodes was built to demonstrate the feasibility of these capabilities.
Part 3 presents Lightbridge, a communications platform specifically designed to provide a more seamless integration between processor nodes and an optically switched network. It addresses some of the crucial issues faced by the works presented in the previous chapters related to optical switching. When optical switches perform switching operations, they change the physical topology of the network, and they lack the capability to buffer packets, resulting in certain optical circuits being unavailable. This prompts the question of whether it is safe to transmit packets by end hosts at any given time. Lightbridge was developed to coordinate switching and routing of optical circuits across the network, by having the processors gain information about the current state of the optical network before transmitting packets, and being able to buffer packets when the optical circuit is not available. This part describes details of Lightbridge which is constituted by a loadable Linux kernel module along with other supporting modifications to the Linux kernel in order to achieve the necessary functionalities
Enabling Technologies for Optical Data Center Networks: Spatial Division Multiplexing
With the continuously growing popularity of cloud services, the traffic volume inside the\ua0data\ua0centers is dramatically increasing. As a result, a scalable and efficient infrastructure\ua0for\ua0data\ua0center\ua0networks\ua0(DCNs) is required. The current\ua0optical\ua0DCNs using either individual fibers or fiber ribbons are costly, bulky, hard to manage, and not scalable.\ua0Spatial\ua0division\ua0multiplexing\ua0(SDM) based on multicore or multimode (few-mode) fibers is recognized as a promising technology to increase the\ua0spatial\ua0efficiency\ua0for\ua0optical\ua0DCNs, which opens a new way towards high capacity and scalability. This tutorial provides an overview of the components, transmission options, and interconnect architectures\ua0for\ua0SDM-based DCNs, as well as potential technical challenges and future directions. It also covers the co-existence of SDM and other\ua0multiplexing\ua0techniques, such as wavelength-division\ua0multiplexing\ua0and flexible spectrum\ua0multiplexing, in\ua0optical\ua0DCNs
Space-Division Multiplexing in Data Center Networks: On Multi-Core Fiber Solutions and Crosstalk-Suppressed Resource Allocation
The rapid growth of traffic inside data centers caused by the increasing adoption of cloud services necessitates a scalable and cost-efficient networking infrastructure. Space-division multiplexing (SDM) is considered as a promising solution to overcome the optical network capacity crunch and support cost-effective network capacity scaling. Multi-core fiber (MCF) is regarded as the most feasible and efficient way to realize SDM networks, and its deployment inside data centers seems very likely as the issue of inter-core crosstalk (XT) is not severe over short link spans (<1  km ) compared to that in long-haul transmission. However, XT can still have a considerable effect in MCF over short distances, which can limit the transmission reach and in turn the data center’s size. XT can be further reduced by bi-directional transmission of optical signals in adjacent MCF cores. This paper evaluates the benefits of MCF-based SDM solutions in terms of maximizing the capacity and spatial efficiency of data center networks. To this end, we present an analytical model for XT in bi-directional normal step-index and trench-assisted MCFs and propose corresponding XT-aware core prioritization schemes. We further develop XT-aware spectrum resource allocation strategies aimed at relieving the complexity of online XT computation. These strategies divide the available spectrum into disjoint bands and incrementally add them to the pool of accessible resources based on the network conditions. Several combinations of core mapping and spectrum resource allocation algorithms are investigated for eight types of homogeneous MCFs comprising 7–61 cores, three different multiplexing schemes, and three data center network topologies with two traffic scenarios. Extensive simulation results show that combining bi-directional transmission in dense core fibers with tailored resource allocation schemes significantly increases the network capacity. Moreover, a multiplexing scheme that combines SDM and WDM can achieve up to 33 times higher link spatial efficiency and up to 300 times greater capacity compared to a WDM solution
- …