690 research outputs found
Trigger and Timing Distributions using the TTC-PON and GBT Bridge Connection in ALICE for the LHC Run 3 Upgrade
The ALICE experiment at CERN is preparing for a major upgrade for the third
phase of data taking run (Run 3), when the high luminosity phase of the Large
Hadron Collider (LHC) starts. The increase in the beam luminosity will result
in high interaction rate causing the data acquisition rate to exceed 3 TB/sec.
In order to acquire data for all the events and to handle the increased data
rate, a transition in the readout electronics architecture from the triggered
to the trigger-less acquisition mode is required. In this new architecture, a
dedicated electronics block called the Common Readout Unit (CRU) is defined to
act as a nodal communication point for detector data aggregation and as a
distribution point for timing, trigger and control (TTC) information. TTC
information in the upgraded triggerless readout architecture uses two
asynchronous high-speed serial links connections: the TTC-PON and the GBT. We
have carried out a study to evaluate the quality of the embedded timing signals
forwarded by the CRU to the connected electronics using the TTC-PON and GBT
bridge connection. We have used four performance metrics to characterize the
communication bridge: (a)the latency added by the firmware logic, (b)the jitter
cleaning effect of the PLL on the timing signal, (c)BER analysis for
quantitative measurement of signal quality, and (d)the effect of optical
transceivers parameter settings on the signal strength. Reliability study of
the bridge connection in maintaining the phase consistency of timing signals is
conducted by performing multiple iterations of power on/off cycle, firmware
upgrade and reset assertion/de-assertion cycle (PFR cycle). The test results
are presented and discussed concerning the performance of the TTC-PON and GBT
bridge communication chain using the CRU prototype and its compliance with the
ALICE timing requirements
Delay Flip-Flop (DFF) Metastability Impact on Clock and Data Recovery (CDR) and Phase-Locked Loop (PLL) Circuits
Modeling delay flip-flops for binary (e.g., Alexander) phase detectors requires paying close attention to three important timing parameters: setup time, hold time, and clock edge-to-output (or briefly C2Q time). These parameters have a critical role in determining the status of the system on the circuit level. This study provided a guideline for designing an optimum DFF for an Alexander phase detector in a clock and data recovery circuit. Furthermore, it indicated DFF timing requirements for a high-speed phase detector in a clock and data recovery circuit. The CDR was also modeled by Verilog-A, and the results were compared with Simulink model achievements. Eventually designed in 45 nm CMOS technology, for 10 Gbps random sequence, the recovered clock contained 0.136 UI and 0.15 UI peak-to-peak jitter on the falling and rising edges respectively, and the lock time was 125 ns. The overall power dissipation was 21 mW from a 1 V supply voltage. Future work includes layout design and manufacturing of the proposed design
Communications with guaranteed bandwidth and low latency using frequency-referenced multiplexing
Emerging cloud applications such as virtual reality and connected car fleets demand guaranteed connections, as well as low and stable latency, to edge data centres. Currently, user–cloud communications rely on time-scheduled data frames through tree-topology fibre networks, which are incapable of providing guaranteed connections with low or stable latency and cannot be scaled to a larger number of users. Here we show that a frequency-referenced multiplexing method can provide guaranteed bandwidth and low latency for time-critical applications. We use clock and optical frequency synchronization, enabled by frequency comb and signal processing techniques, to provide each user with dedicated optical bandwidth, creating scalable user–cloud upstream communications. As a proof of concept, we demonstrate a frequency-division multiplexing system servicing up to 64 users with an aggregate bandwidth of 160 GHz, exhibiting a data rate of up to 4.3 Gbps per user (240.0 Gbps aggregated capacity considering a 200 GHz wavelength band) with a high receiver sensitivity of –35 dBm
Megabits secure key rate quantum key distribution
Quantum cryptography (QC) can provide unconditional secure communication
between two authorized parties based on the basic principles of quantum
mechanics. However, imperfect practical conditions limit its transmission
distance and communication speed. Here we implemented the differential phase
shift (DPS) quantum key distribution (QKD) with up-conversion assisted hybrid
photon detector (HPD) and achieved 1.3 M bits per second secure key rate over a
10-km fiber, which is tolerant against the photon number splitting (PNS)
attack, general collective attacks on individual photons, and any other known
sequential unambiguous state discrimination (USD) attacks.Comment: 14 pages, 4 figure
Coexistence of high-bit-rate quantum key distribution and data on optical fiber
Quantum key distribution (QKD) uniquely allows distribution of cryptographic
keys with security verified by quantum mechanical limits. Both protocol
execution and subsequent applications require the assistance of classical data
communication channels. While using separate fibers is one option, it is
economically more viable if data and quantum signals are simultaneously
transmitted through a single fiber. However, noise-photon contamination arising
from the intense data signal has severely restricted both the QKD distances and
secure key rates. Here, we exploit a novel temporal-filtering effect for
noise-photon rejection. This allows high-bit-rate QKD over fibers up to 90 km
in length and populated with error-free bidirectional Gb/s data communications.
With high-bit rate and range sufficient for important information
infrastructures, such as smart cities and 10 Gbit Ethernet, QKD is a
significant step closer towards wide-scale deployment in fiber networks.Comment: 7 pages, 5 figure
Recommended from our members
Fully-photonic digital radio over fibre for future super-broadband access network applications
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel UniversityIn this thesis a Fully-Photonic DRoF (FP-DRoF) system is proposed for deploying of future super-broadband access networks. Digital Radio over Fibre (DRoF) is more independent of the fibre network impairments and the length of fibre than the ARoF link. In order for fully optical deployment of the signal conversion techniques in the FP-DRoF architecture, two key components an Analogue-to-Digital Converter (ADC) and a Digital-to-Analogue Converter (DAC)) for data conversion are designed and their performance are investigated whereas the physical functionality is evaluated. The system simulation results of the proposed pipelined Photonic ADC (PADC) show that the PADC has 10 GHz bandwidth around 60 GHz of sampling rate. Furthermore, by
changing the bandwidth of the optical bandpass filter, switching to another band of sampling frequency provides optimised performance condition of the PADC. The PADC has low changes on the Effective Number of Bit (ENOB) response versus analogue RF input from 1 GHz up to 22 GHz for 60 GHz sampling frequency. The proposed 8-Bit pipelined PADC performance in terms of ENOB is evaluated at 60 Gigasample/s which is about 4.1. Recently, different methods have been reported by researchers to implement Photonic DACs
(PDACs), but their aim was to convert digital electrical signals to the corresponding analogue signal by assisting the optical techniques. In this thesis, a Binary Weighted PDAC (BW-PDAC) is proposed. In this BW-PDAC, optical digital signals are fully optically converted to an analogue signal. The spurious free dynamic range at the output of the PDAC in a back-to-back deployment of the PADC and the PDAC was 26.6 dBc. For further improvement in the system performance, a 3R (Retiming, Reshaping and Reamplifying) regeneration system is proposed in this thesis. Simulation results show that for an ultrashort RZ pulse with a 5% duty cycle at 65 Gbit/s using the proposed 3R regeneration system on a link reduces rms timing jitter by 90% while the regenerated pulse eye opening height is improved by 65%. Finally, in this thesis the proposed FP-DRoF functionality is evaluated whereas its performance is investigated through a dedicated and shared fibre links. The simulation results show (in the case of low level signal to noise ratio, in comparison with ARoF through
a dedicated fibre link) that the FP-DRoF has better BER performance than the ARoF in the order of 10-20. Furthermore, in order to realize a BER about 10-25 for the ARoF, the power penalty is about 4 dBm higher than the FP-DRoF link. The simulation results demonstrate that by considering 0.2 dB/km attenuation of a standard single mode fibre, the dedicated fibre length for the FP-DRoF link can be increased to about 20 km more than the ARoF link. Moreover, for performance assessment of the proposed FP-DRoF in a shared fibre link, the BER of the FP-DRoF link is about 10-10 magnitude less than the ARoF link for -19 dBm launched power into the fibre and the power penalty of the ARoF system is 10 dBm more than the FP-DRoF link. It is significant to increase the fibre link’s length of the FP-DRoF access network using common infrastructure. In addition, the simulation results are demonstrated that the FP-DRoF with non-uniform Wavelength Division Multiplexing (WDM) is more robust against four wave mixing impairment than the conventional WDM technique with uniform wavelength allocation and has better performance in terms of BER. It is clearly verified that the lunched power penalty at CS for DRoF link with uniform WDM techniques is about 2 dB higher than non-uniform WDM technique. Furthermore, uniform WDM method requires more bandwidth than non-uniform scheme which depends on the total number of channels and channels spacing
Analysis of jitter impact on high speed transmissions of wavelength-division multiplexing networks
In this study, we conduct a thorough assessment of the effect of jitter occurrence in high speed 10 Gbps and 200 GHz Wavelength-Division Multiplexing (WDM) optical network. First, we present a simulation model to study the effect of jitter presence in the proposed network and then determine the maximum amount of jitter which the network can withstand. The model is then employed to predict the types of jitter received at the end of the transmission line. For the input power level of 0 dBm and Bit Error Rate (BER) of 1E09, the observed total jitter, JT, random jitter, JR and deterministic jitter, JD is 0.2676 UI, 0.1602 UI and 0.1073 UI, respectively
- …