1,916 research outputs found
Iterative Soft Input Soft Output Decoding of Reed-Solomon Codes by Adapting the Parity Check Matrix
An iterative algorithm is presented for soft-input-soft-output (SISO)
decoding of Reed-Solomon (RS) codes. The proposed iterative algorithm uses the
sum product algorithm (SPA) in conjunction with a binary parity check matrix of
the RS code. The novelty is in reducing a submatrix of the binary parity check
matrix that corresponds to less reliable bits to a sparse nature before the SPA
is applied at each iteration. The proposed algorithm can be geometrically
interpreted as a two-stage gradient descent with an adaptive potential
function. This adaptive procedure is crucial to the convergence behavior of the
gradient descent algorithm and, therefore, significantly improves the
performance. Simulation results show that the proposed decoding algorithm and
its variations provide significant gain over hard decision decoding (HDD) and
compare favorably with other popular soft decision decoding methods.Comment: 10 pages, 10 figures, final version accepted by IEEE Trans. on
Information Theor
Roadmap on optical security
Postprint (author's final draft
Convolutional Dictionary Learning through Tensor Factorization
Tensor methods have emerged as a powerful paradigm for consistent learning of
many latent variable models such as topic models, independent component
analysis and dictionary learning. Model parameters are estimated via CP
decomposition of the observed higher order input moments. However, in many
domains, additional invariances such as shift invariances exist, enforced via
models such as convolutional dictionary learning. In this paper, we develop
novel tensor decomposition algorithms for parameter estimation of convolutional
models. Our algorithm is based on the popular alternating least squares method,
but with efficient projections onto the space of stacked circulant matrices.
Our method is embarrassingly parallel and consists of simple operations such as
fast Fourier transforms and matrix multiplications. Our algorithm converges to
the dictionary much faster and more accurately compared to the alternating
minimization over filters and activation maps
- …