7 research outputs found

    Simple and Nearly Optimal Polynomial Root-finding by Means of Root Radii Approximation

    Full text link
    We propose a new simple but nearly optimal algorithm for the approximation of all sufficiently well isolated complex roots and root clusters of a univariate polynomial. Quite typically the known root-finders at first compute some crude but reasonably good approximations to well-conditioned roots (that is, those isolated from the other roots) and then refine the approximations very fast, by using Boolean time which is nearly optimal, up to a polylogarithmic factor. By combining and extending some old root-finding techniques, the geometry of the complex plane, and randomized parametrization, we accelerate the initial stage of obtaining crude to all well-conditioned simple and multiple roots as well as isolated root clusters. Our algorithm performs this stage at a Boolean cost dominated by the nearly optimal cost of subsequent refinement of these approximations, which we can perform concurrently, with minimum processor communication and synchronization. Our techniques are quite simple and elementary; their power and application range may increase in their combination with the known efficient root-finding methods.Comment: 12 pages, 1 figur

    Non-Negative Local Sparse Coding for Subspace Clustering

    Full text link
    Subspace sparse coding (SSC) algorithms have proven to be beneficial to clustering problems. They provide an alternative data representation in which the underlying structure of the clusters can be better captured. However, most of the research in this area is mainly focused on enhancing the sparse coding part of the problem. In contrast, we introduce a novel objective term in our proposed SSC framework which focuses on the separability of data points in the coding space. We also provide mathematical insights into how this local-separability term improves the clustering result of the SSC framework. Our proposed non-linear local SSC algorithm (NLSSC) also benefits from the efficient choice of its sparsity terms and constraints. The NLSSC algorithm is also formulated in the kernel-based framework (NLKSSC) which can represent the nonlinear structure of data. In addition, we address the possibility of having redundancies in sparse coding results and its negative effect on graph-based clustering problems. We introduce the link-restore post-processing step to improve the representation graph of non-negative SSC algorithms such as ours. Empirical evaluations on well-known clustering benchmarks show that our proposed NLSSC framework results in better clusterings compared to the state-of-the-art baselines and demonstrate the effectiveness of the link-restore post-processing in improving the clustering accuracy via correcting the broken links of the representation graph.Comment: 15 pages, IDA 2018 conferenc

    Non-Negative Local Sparse Coding for Subspace Clustering

    No full text
    Hosseini B, Hammer B. Non-Negative Local Sparse Coding for Subspace Clustering. Advances in Intelligent Data Analysis XVII. IDA 2018. 2018.Subspace sparse coding (SSC) algorithms have proven to be beneficial to the clustering problems. They provide an alternative data representation in which the underlying structure of the clusters can be better captured. However, most of the research in this area is mainly focused on enhancing the sparse coding part of the problem. In contrast, we introduce a novel objective term in our proposed SSC framework which focuses on the separability of data points in the coding space. We also provide mathematical insights into how this local-separability term can improve the clustering result of the SSC framework. Our proposed non-linear local SSC algorithm (NLSSC) also benefits from the efficient choice of its sparsity terms and constraints. The NLSSC algorithm is also formulated in the kernel-based framework (NLKSSC) which can represent the nonlinear structure of data. In addition, we address the possibility of having redundancies in sparse coding results and its negative effect on graph-based clustering problems. Accordingly, we introduce the link-restore post-processing step to improve the representation graph of non-negative SSC algorithms such as ours. Empirical evaluations on well-known clustering benchmarks show that our proposed NLSSC framework results in better clusterings compared to the state-of-the-art baselines, and demonstrate the effectiveness of the link-restore post-processing in improving the clustering accuracy via correcting the broken links of the representation graph

    Polynomial Evaluation and Interpolation and Transformations of Matrix Structures

    No full text
    Abstract. Multipoint polynomial evaluation and interpolation are fundamental for modern numerical and symbolic computing. The known algorithms solve both problems over any field of constants in nearly linear arithmetic time, but the cost grows to quadratic for numerical solution. We decrease this cost dramatically and for a large class of inputs yield nearly linear time as well. We first restate our tasks as multiplication of a Vandermonde matrix and its inverse by a vector, then transform this matrix into other structured matrices, and finally apply a variant of the Multipole celebrated techniques to achieve the desired speedup for the computations with polynomials, Vandermonde matrices and their transposes. An important impact of our work is a new demonstration of the power of the method of the transformation of matrix structures, which we proposed in [P90]. At the end we comment on further applications and extension of this method to computations with structured matrices, polynomials, and rational functions

    Root-Refining for a Polynomial Equation

    No full text
    Abstract. Polynomial root-finders usually consist of two stages. At first a crude approximation to a root is slowly computed; then it is much faster refined by means of the same or distinct iteration. The efficiency of com-puting an initial approximation resists formal study, and the users rely on empirical data. In contrast, the efficiency of refinement is formally measured by the classical concept q1/d where q denotes the convergence order, whereas d denotes the number of function evaluations per iter-ation. In our case of a polynomial of a degree n we use 2n arithmetic operations per its evaluation of at a point. Noting this we extend the def-inition to cover iterations that are not reduced to function evaluations alone, including iterations that simultaneously refine n approximations to all n roots of a degree n polynomial. By employing two approaches to the latter task, both based on recursive polynomial factorization, we yield refinement with the efficiency 2d, d = cn / log2 n for a positive constant c. For large n this is a dramatic increase versus the record efficiency 2 of refining an approximation to a single root of a polynomial. The advance could motivate practical use of the proposed root-refiners

    Fast matrix multiplication and its algebraic neighborhood

    No full text
    corecore