101,474 research outputs found
Throughput-based Design for Polar Coded-Modulation
Typically, forward error correction (FEC) codes are designed based on the
minimization of the error rate for a given code rate. However, for applications
that incorporate hybrid automatic repeat request (HARQ) protocol and adaptive
modulation and coding, the throughput is a more important performance metric
than the error rate. Polar codes, a new class of FEC codes with simple rate
matching, can be optimized efficiently for maximization of the throughput. In
this paper, we aim to design HARQ schemes using multilevel polar
coded-modulation (MLPCM). Thus, we first develop a method to determine a
set-partitioning based bit-to-symbol mapping for high order QAM constellations.
We simplify the LLR estimation of set-partitioned QAM constellations for a
multistage decoder, and we introduce a set of algorithms to design
throughput-maximizing MLPCM for the successive cancellation decoding (SCD).
These codes are specifically useful for non-combining (NC) and Chase-combining
(CC) HARQ protocols. Furthermore, since optimized codes for SCD are not optimal
for SC list decoders (SCLD), we propose a rate matching algorithm to find the
best rate for SCLD while using the polar codes optimized for SCD. The resulting
codes provide throughput close to the capacity with low decoding complexity
when used with NC or CC HARQ
Motion-Compensated Coding and Frame-Rate Up-Conversion: Models and Analysis
Block-based motion estimation (ME) and compensation (MC) techniques are
widely used in modern video processing algorithms and compression systems. The
great variety of video applications and devices results in numerous compression
specifications. Specifically, there is a diversity of frame-rates and
bit-rates. In this paper, we study the effect of frame-rate and compression
bit-rate on block-based ME and MC as commonly utilized in inter-frame coding
and frame-rate up conversion (FRUC). This joint examination yields a
comprehensive foundation for comparing MC procedures in coding and FRUC. First,
the video signal is modeled as a noisy translational motion of an image. Then,
we theoretically model the motion-compensated prediction of an available and
absent frames as in coding and FRUC applications, respectively. The theoretic
MC-prediction error is further analyzed and its autocorrelation function is
calculated for coding and FRUC applications. We show a linear relation between
the variance of the MC-prediction error and temporal-distance. While the
affecting distance in MC-coding is between the predicted and reference frames,
MC-FRUC is affected by the distance between the available frames used for the
interpolation. Moreover, the dependency in temporal-distance implies an inverse
effect of the frame-rate. FRUC performance analysis considers the prediction
error variance, since it equals to the mean-squared-error of the interpolation.
However, MC-coding analysis requires the entire autocorrelation function of the
error; hence, analytic simplicity is beneficial. Therefore, we propose two
constructions of a separable autocorrelation function for prediction error in
MC-coding. We conclude by comparing our estimations with experimental results
Simplified Multiuser Detection for SCMA with Sum-Product Algorithm
Sparse code multiple access (SCMA) is a novel non-orthogonal multiple access
technique, which fully exploits the shaping gain of multi-dimensional
codewords. However, the lack of simplified multiuser detection algorithm
prevents further implementation due to the inherently high computation
complexity. In this paper, general SCMA detector algorithms based on
Sum-product algorithm are elaborated. Then two improved algorithms are
proposed, which simplify the detection structure and curtail exponent
operations quantitatively in logarithm domain. Furthermore, to analyze these
detection algorithms fairly, we derive theoretical expression of the average
mutual information (AMI) of SCMA (SCMA-AMI), and employ a statistical method to
calculate SCMA-AMI based specific detection algorithm. Simulation results show
that the performance is almost as well as the based message passing algorithm
in terms of both BER and AMI while the complexity is significantly decreased,
compared to the traditional Max-Log approximation method
Dissipation-induced continuous quantum error correction for superconducting circuits
Quantum error correction (QEC) is a crucial step towards long coherence times
required for efficient quantum information processing (QIP). One major
challenge in this direction concerns the fast real-time analysis of error
syndrome measurements and the associated feedback control. Recent proposals on
autonomous QEC (AQEC) have opened new perspectives to overcome this difficulty.
Here, we design an AQEC scheme based on quantum reservoir engineering adapted
to superconducting qubits. We focus on a three-qubit bit-flip code, where three
transmon qubits are dispersively coupled to a few low-Q resonator modes. By
applying only continuous-wave drives of fixed but well-chosen frequencies and
amplitudes, we engineer an effective interaction Hamiltonian to evacuate the
entropy created by eventual bit-flip errors. We provide a full analytical and
numerical study of the protocol, while introducing the main limitations on the
achievable error correction rates.Comment: 9 pages, 6 figure
Reconstructing the calibrated strain signal in the Advanced LIGO detectors
Advanced LIGO's raw detector output needs to be calibrated to compute
dimensionless strain h(t). Calibrated strain data is produced in the time
domain using both a low-latency, online procedure and a high-latency, offline
procedure. The low-latency h(t) data stream is produced in two stages, the
first of which is performed on the same computers that operate the detector's
feedback control system. This stage, referred to as the front-end calibration,
uses infinite impulse response (IIR) filtering and performs all operations at a
16384 Hz digital sampling rate. Due to several limitations, this procedure
currently introduces certain systematic errors in the calibrated strain data,
motivating the second stage of the low-latency procedure, known as the
low-latency gstlal calibration pipeline. The gstlal calibration pipeline uses
finite impulse response (FIR) filtering to apply corrections to the output of
the front-end calibration. It applies time-dependent correction factors to the
sensing and actuation components of the calibrated strain to reduce systematic
errors. The gstlal calibration pipeline is also used in high latency to
recalibrate the data, which is necessary due mainly to online dropouts in the
calibrated data and identified improvements to the calibration models or
filters.Comment: 20 pages including appendices and bibliography. 11 Figures. 3 Table
- …