14,705 research outputs found

    Randomized Tensor Ring Decomposition and Its Application to Large-scale Data Reconstruction

    Full text link
    Dimensionality reduction is an essential technique for multi-way large-scale data, i.e., tensor. Tensor ring (TR) decomposition has become popular due to its high representation ability and flexibility. However, the traditional TR decomposition algorithms suffer from high computational cost when facing large-scale data. In this paper, taking advantages of the recently proposed tensor random projection method, we propose two TR decomposition algorithms. By employing random projection on every mode of the large-scale tensor, the TR decomposition can be processed at a much smaller scale. The simulation experiment shows that the proposed algorithms are 4−254-25 times faster than traditional algorithms without loss of accuracy, and our algorithms show superior performance in deep learning dataset compression and hyperspectral image reconstruction experiments compared to other randomized algorithms.Comment: ICASSP submissio

    QCD and Relativistic O(αsv2)O(\alpha_{s}v^2) Corrections to Hadronic Decays of Spin-Singlet Heavy Quarkonia hc,hbh_c, h_b and ηb\eta_b

    Full text link
    We calculate the annihilation decay widths of spin-singlet heavy quarkonia hc,hbh_c, h_b and ηb\eta_b} into light hadrons with both QCD and relativistic corrections at order O(αsv2)O(\alpha_{s}v^{2}) in nonrelativistic QCD. With appropriate estimates for the long-distance matrix elements by using the potential model and operator evolution method, we find that our predictions of these decay widths are consistent with recent experimental measurements. We also find that the O(αsv2)O(\alpha_{s}v^{2}) corrections are small for bbˉb\bar{b} states but substantial for ccˉc\bar{c} states. In particular, the negative contribution of O(αsv2)O(\alpha_{s}v^{2}) correction to the hch_{c} decay can lower the decay width, as compared with previous predictions without the O(αsv2)O(\alpha_{s}v^{2}) correction, and thus result in a good agreement with the recent BESIII measurement.Comment: version published in PRD, 30 pages, 8 figures, more discussions on LDMEs adde

    Generative Adversarial Positive-Unlabelled Learning

    Full text link
    In this work, we consider the task of classifying binary positive-unlabeled (PU) data. The existing discriminative learning based PU models attempt to seek an optimal reweighting strategy for U data, so that a decent decision boundary can be found. However, given limited P data, the conventional PU models tend to suffer from overfitting when adapted to very flexible deep neural networks. In contrast, we are the first to innovate a totally new paradigm to attack the binary PU task, from perspective of generative learning by leveraging the powerful generative adversarial networks (GAN). Our generative positive-unlabeled (GenPU) framework incorporates an array of discriminators and generators that are endowed with different roles in simultaneously producing positive and negative realistic samples. We provide theoretical analysis to justify that, at equilibrium, GenPU is capable of recovering both positive and negative data distributions. Moreover, we show GenPU is generalizable and closely related to the semi-supervised classification. Given rather limited P data, experiments on both synthetic and real-world dataset demonstrate the effectiveness of our proposed framework. With infinite realistic and diverse sample streams generated from GenPU, a very flexible classifier can then be trained using deep neural networks.Comment: 8 page

    Somoclu: An Efficient Parallel Library for Self-Organizing Maps

    Get PDF
    Somoclu is a massively parallel tool for training self-organizing maps on large data sets written in C++. It builds on OpenMP for multicore execution, and on MPI for distributing the workload across the nodes in a cluster. It is also able to boost training by using CUDA if graphics processing units are available. A sparse kernel is included, which is useful for high-dimensional but sparse data, such as the vector spaces common in text mining workflows. Python, R and MATLAB interfaces facilitate interactive use. Apart from fast execution, memory use is highly optimized, enabling training large emergent maps even on a single computer.Comment: 26 pages, 9 figures. The code is available at https://peterwittek.github.io/somoclu
    • …
    corecore