7 research outputs found
Simple and Nearly Optimal Polynomial Root-finding by Means of Root Radii Approximation
We propose a new simple but nearly optimal algorithm for the approximation of
all sufficiently well isolated complex roots and root clusters of a univariate
polynomial. Quite typically the known root-finders at first compute some crude
but reasonably good approximations to well-conditioned roots (that is, those
isolated from the other roots) and then refine the approximations very fast, by
using Boolean time which is nearly optimal, up to a polylogarithmic factor. By
combining and extending some old root-finding techniques, the geometry of the
complex plane, and randomized parametrization, we accelerate the initial stage
of obtaining crude to all well-conditioned simple and multiple roots as well as
isolated root clusters. Our algorithm performs this stage at a Boolean cost
dominated by the nearly optimal cost of subsequent refinement of these
approximations, which we can perform concurrently, with minimum processor
communication and synchronization. Our techniques are quite simple and
elementary; their power and application range may increase in their combination
with the known efficient root-finding methods.Comment: 12 pages, 1 figur
Non-Negative Local Sparse Coding for Subspace Clustering
Subspace sparse coding (SSC) algorithms have proven to be beneficial to
clustering problems. They provide an alternative data representation in which
the underlying structure of the clusters can be better captured. However, most
of the research in this area is mainly focused on enhancing the sparse coding
part of the problem. In contrast, we introduce a novel objective term in our
proposed SSC framework which focuses on the separability of data points in the
coding space. We also provide mathematical insights into how this
local-separability term improves the clustering result of the SSC framework.
Our proposed non-linear local SSC algorithm (NLSSC) also benefits from the
efficient choice of its sparsity terms and constraints. The NLSSC algorithm is
also formulated in the kernel-based framework (NLKSSC) which can represent the
nonlinear structure of data. In addition, we address the possibility of having
redundancies in sparse coding results and its negative effect on graph-based
clustering problems. We introduce the link-restore post-processing step to
improve the representation graph of non-negative SSC algorithms such as ours.
Empirical evaluations on well-known clustering benchmarks show that our
proposed NLSSC framework results in better clusterings compared to the
state-of-the-art baselines and demonstrate the effectiveness of the
link-restore post-processing in improving the clustering accuracy via
correcting the broken links of the representation graph.Comment: 15 pages, IDA 2018 conferenc
Non-Negative Local Sparse Coding for Subspace Clustering
Hosseini B, Hammer B. Non-Negative Local Sparse Coding for Subspace Clustering. Advances in Intelligent Data Analysis XVII. IDA 2018. 2018.Subspace sparse coding (SSC) algorithms have proven to be beneficial to the clustering problems. They provide an alternative data representation in which the underlying structure of the clusters can be better captured. However, most of the research in this area is mainly focused on enhancing the sparse coding part of the problem. In contrast, we introduce a novel objective term in our proposed SSC framework which focuses on the separability of data points in the coding space. We also provide mathematical insights into how this local-separability term can improve the clustering result of the SSC framework. Our proposed non-linear local SSC algorithm (NLSSC) also benefits from the efficient choice of its sparsity terms and constraints. The NLSSC algorithm is also formulated in the kernel-based framework (NLKSSC) which can represent the nonlinear structure of data. In addition, we address the possibility of having redundancies in sparse coding results and its negative effect on graph-based clustering problems. Accordingly, we introduce the link-restore post-processing step to improve the representation graph of non-negative SSC algorithms such as ours. Empirical evaluations on well-known clustering benchmarks show that our proposed NLSSC framework results in better clusterings compared to the state-of-the-art baselines, and demonstrate the effectiveness of the link-restore post-processing in improving the clustering accuracy via correcting the broken links of the representation graph
Polynomial Evaluation and Interpolation and Transformations of Matrix Structures
Abstract. Multipoint polynomial evaluation and interpolation are fundamental for modern numerical and symbolic computing. The known algorithms solve both problems over any field of constants in nearly linear arithmetic time, but the cost grows to quadratic for numerical solution. We decrease this cost dramatically and for a large class of inputs yield nearly linear time as well. We first restate our tasks as multiplication of a Vandermonde matrix and its inverse by a vector, then transform this matrix into other structured matrices, and finally apply a variant of the Multipole celebrated techniques to achieve the desired speedup for the computations with polynomials, Vandermonde matrices and their transposes. An important impact of our work is a new demonstration of the power of the method of the transformation of matrix structures, which we proposed in [P90]. At the end we comment on further applications and extension of this method to computations with structured matrices, polynomials, and rational functions
Root-Refining for a Polynomial Equation
Abstract. Polynomial root-finders usually consist of two stages. At first a crude approximation to a root is slowly computed; then it is much faster refined by means of the same or distinct iteration. The efficiency of com-puting an initial approximation resists formal study, and the users rely on empirical data. In contrast, the efficiency of refinement is formally measured by the classical concept q1/d where q denotes the convergence order, whereas d denotes the number of function evaluations per iter-ation. In our case of a polynomial of a degree n we use 2n arithmetic operations per its evaluation of at a point. Noting this we extend the def-inition to cover iterations that are not reduced to function evaluations alone, including iterations that simultaneously refine n approximations to all n roots of a degree n polynomial. By employing two approaches to the latter task, both based on recursive polynomial factorization, we yield refinement with the efficiency 2d, d = cn / log2 n for a positive constant c. For large n this is a dramatic increase versus the record efficiency 2 of refining an approximation to a single root of a polynomial. The advance could motivate practical use of the proposed root-refiners