6,731 research outputs found
Implementation of the Trigonometric LMS Algorithm using Original Cordic Rotation
The LMS algorithm is one of the most successful adaptive filtering
algorithms. It uses the instantaneous value of the square of the error signal
as an estimate of the mean-square error (MSE). The LMS algorithm changes
(adapts) the filter tap weights so that the error signal is minimized in the
mean square sense. In Trigonometric LMS (TLMS) and Hyperbolic LMS (HLMS), two
new versions of LMS algorithms, same formulations are performed as in the LMS
algorithm with the exception that filter tap weights are now expressed using
trigonometric and hyperbolic formulations, in cases for TLMS and HLMS
respectively. Hence appears the CORDIC algorithm as it can efficiently perform
trigonometric, hyperbolic, linear and logarithmic functions. While
hardware-efficient algorithms often exist, the dominance of the software
systems has kept those algorithms out of the spotlight. Among these hardware-
efficient algorithms, CORDIC is an iterative solution for trigonometric and
other transcendental functions. Former researches worked on CORDIC algorithm to
observe the convergence behavior of Trigonometric LMS (TLMS) algorithm and
obtained a satisfactory result in the context of convergence performance of
TLMS algorithm. But revious researches directly used the CORDIC block output in
their simulation ignoring the internal step-by-step rotations of the CORDIC
processor. This gives rise to a need for verification of the convergence
performance of the TLMS algorithm to investigate if it actually performs
satisfactorily if implemented with step-by-step CORDIC rotation. This research
work has done this job. It focuses on the internal operations of the CORDIC
hardware, implements the Trigonometric LMS (TLMS) and Hyperbolic LMS (HLMS)
algorithms using actual CORDIC rotations. The obtained simulation results are
highly satisfactory and also it shows that convergence behavior of HLMS is much
better than TLMS.Comment: 12 pages, 5 figures, 1 table. Published in IJCNC;
http://airccse.org/journal/cnc/0710ijcnc08.pdf,
http://airccse.org/journal/ijc2010.htm
Fast and Accurate Bilateral Filtering using Gauss-Polynomial Decomposition
The bilateral filter is a versatile non-linear filter that has found diverse
applications in image processing, computer vision, computer graphics, and
computational photography. A widely-used form of the filter is the Gaussian
bilateral filter in which both the spatial and range kernels are Gaussian. A
direct implementation of this filter requires operations per
pixel, where is the standard deviation of the spatial Gaussian. In
this paper, we propose an accurate approximation algorithm that can cut down
the computational complexity to per pixel for any arbitrary
(constant-time implementation). This is based on the observation that the range
kernel operates via the translations of a fixed Gaussian over the range space,
and that these translated Gaussians can be accurately approximated using the
so-called Gauss-polynomials. The overall algorithm emerging from this
approximation involves a series of spatial Gaussian filtering, which can be
implemented in constant-time using separability and recursion. We present some
preliminary results to demonstrate that the proposed algorithm compares
favorably with some of the existing fast algorithms in terms of speed and
accuracy.Comment: To appear in the IEEE International Conference on Image Processing
(ICIP 2015
Efficient implementation of the Hardy-Ramanujan-Rademacher formula
We describe how the Hardy-Ramanujan-Rademacher formula can be implemented to
allow the partition function to be computed with softly optimal
complexity and very little overhead. A new implementation
based on these techniques achieves speedups in excess of a factor 500 over
previously published software and has been used by the author to calculate
, an exponent twice as large as in previously reported
computations.
We also investigate performance for multi-evaluation of , where our
implementation of the Hardy-Ramanujan-Rademacher formula becomes superior to
power series methods on far denser sets of indices than previous
implementations. As an application, we determine over 22 billion new
congruences for the partition function, extending Weaver's tabulation of 76,065
congruences.Comment: updated version containing an unconditional complexity proof;
accepted for publication in LMS Journal of Computation and Mathematic
Wildcard dimensions, coding theory and fault-tolerant meshes and hypercubes
Hypercubes, meshes and tori are well known interconnection networks for parallel computers. The sets of edges in those graphs can be partitioned to dimensions. It is well known that the hypercube can be extended by adding a wildcard dimension resulting in a folded hypercube that has better fault-tolerant and communication capabilities. First we prove that the folded hypercube is optimal in the sense that only a single wildcard dimension can be added to the hypercube. We then investigate the idea of adding wildcard dimensions to d-dimensional meshes and tori. Using techniques from error correcting codes we construct d-dimensional meshes and tori with wildcard dimensions. Finally, we show how these constructions can be used to tolerate edge and node faults in mesh and torus networks
Accuracy and Stability of Computing High-Order Derivatives of Analytic Functions by Cauchy Integrals
High-order derivatives of analytic functions are expressible as Cauchy
integrals over circular contours, which can very effectively be approximated,
e.g., by trapezoidal sums. Whereas analytically each radius r up to the radius
of convergence is equal, numerical stability strongly depends on r. We give a
comprehensive study of this effect; in particular we show that there is a
unique radius that minimizes the loss of accuracy caused by round-off errors.
For large classes of functions, though not for all, this radius actually gives
about full accuracy; a remarkable fact that we explain by the theory of Hardy
spaces, by the Wiman-Valiron and Levin-Pfluger theory of entire functions, and
by the saddle-point method of asymptotic analysis. Many examples and
non-trivial applications are discussed in detail.Comment: Version 4 has some references and a discussion of other quadrature
rules added; 57 pages, 7 figures, 6 tables; to appear in Found. Comput. Mat
- …