3,852 research outputs found

    Adaptive multichannel control of time-varying broadband noise and vibrations

    Get PDF
    This paper presents results obtained from a number of applications in which a recent adaptive algorithm for broadband multichannel active noise control is used. The core of the algorithm uses the inverse of the minimum-phase part of the secondary path for improvement of the speed of convergence. A further improvement of the speed of convergence is obtained by using double control filters for elimination of adaptation loop delay. Regularization was found to be necessary for robust operation. The regularization technique which is used preserves the structure to eliminate the adaptation loop delay. Depending on the application at hand, a number of extensions are used for this algorithm. For an application with rapidly changing disturbance spectra, the core algorithm was extended with an iterative affine projection scheme, leading to improved convergence rates as compared to the standard nomalized lms update rules. In another application, in which the influence of the parametric uncertainties was critical, the core algorithm was extended with low authority control loops operating at high sample rates. In addition, results of other applications are given, such as control of acoustic energy density and control of time-varying periodic and non-periodic vibrations

    Cycles in adversarial regularized learning

    Get PDF
    Regularized learning is a fundamental technique in online optimization, machine learning and many other fields of computer science. A natural question that arises in these settings is how regularized learning algorithms behave when faced against each other. We study a natural formulation of this problem by coupling regularized learning dynamics in zero-sum games. We show that the system's behavior is Poincar\'e recurrent, implying that almost every trajectory revisits any (arbitrarily small) neighborhood of its starting point infinitely often. This cycling behavior is robust to the agents' choice of regularization mechanism (each agent could be using a different regularizer), to positive-affine transformations of the agents' utilities, and it also persists in the case of networked competition, i.e., for zero-sum polymatrix games.Comment: 22 pages, 4 figure

    Disturbance Grassmann Kernels for Subspace-Based Learning

    Full text link
    In this paper, we focus on subspace-based learning problems, where data elements are linear subspaces instead of vectors. To handle this kind of data, Grassmann kernels were proposed to measure the space structure and used with classifiers, e.g., Support Vector Machines (SVMs). However, the existing discriminative algorithms mostly ignore the instability of subspaces, which would cause the classifiers misled by disturbed instances. Thus we propose considering all potential disturbance of subspaces in learning processes to obtain more robust classifiers. Firstly, we derive the dual optimization of linear classifiers with disturbance subject to a known distribution, resulting in a new kernel, Disturbance Grassmann (DG) kernel. Secondly, we research into two kinds of disturbance, relevant to the subspace matrix and singular values of bases, with which we extend the Projection kernel on Grassmann manifolds to two new kernels. Experiments on action data indicate that the proposed kernels perform better compared to state-of-the-art subspace-based methods, even in a worse environment.Comment: This paper include 3 figures, 10 pages, and has been accpeted to SIGKDD'1
    • ā€¦
    corecore