3,722 research outputs found

    CSD: Discriminance with Conic Section for Improving Reverse k Nearest Neighbors Queries

    Full text link
    The reverse kk nearest neighbor (RkkNN) query finds all points that have the query point as one of their kk nearest neighbors (kkNN), where the kkNN query finds the kk closest points to its query point. Based on the characteristics of conic section, we propose a discriminance, named CSD (Conic Section Discriminance), to determine points whether belong to the RkkNN set without issuing any queries with non-constant computational complexity. By using CSD, we also implement an efficient RkkNN algorithm CSD-RkkNN with a computational complexity at O(k1.5⋅log k)O(k^{1.5}\cdot log\,k). The comparative experiments are conducted between CSD-RkkNN and other two state-of-the-art RkNN algorithms, SLICE and VR-RkkNN. The experimental results indicate that the efficiency of CSD-RkkNN is significantly higher than its competitors

    A Graph-Based Semi-Supervised k Nearest-Neighbor Method for Nonlinear Manifold Distributed Data Classification

    Get PDF
    kk Nearest Neighbors (kkNN) is one of the most widely used supervised learning algorithms to classify Gaussian distributed data, but it does not achieve good results when it is applied to nonlinear manifold distributed data, especially when a very limited amount of labeled samples are available. In this paper, we propose a new graph-based kkNN algorithm which can effectively handle both Gaussian distributed data and nonlinear manifold distributed data. To achieve this goal, we first propose a constrained Tired Random Walk (TRW) by constructing an RR-level nearest-neighbor strengthened tree over the graph, and then compute a TRW matrix for similarity measurement purposes. After this, the nearest neighbors are identified according to the TRW matrix and the class label of a query point is determined by the sum of all the TRW weights of its nearest neighbors. To deal with online situations, we also propose a new algorithm to handle sequential samples based a local neighborhood reconstruction. Comparison experiments are conducted on both synthetic data sets and real-world data sets to demonstrate the validity of the proposed new kkNN algorithm and its improvements to other version of kkNN algorithms. Given the widespread appearance of manifold structures in real-world problems and the popularity of the traditional kkNN algorithm, the proposed manifold version kkNN shows promising potential for classifying manifold-distributed data.Comment: 32 pages, 12 figures, 7 table
    • …
    corecore