1,052,969 research outputs found
An Effective Feature Selection Method Based on Pair-Wise Feature Proximity for High Dimensional Low Sample Size Data
Feature selection has been studied widely in the literature. However, the
efficacy of the selection criteria for low sample size applications is
neglected in most cases. Most of the existing feature selection criteria are
based on the sample similarity. However, the distance measures become
insignificant for high dimensional low sample size (HDLSS) data. Moreover, the
variance of a feature with a few samples is pointless unless it represents the
data distribution efficiently. Instead of looking at the samples in groups, we
evaluate their efficiency based on pairwise fashion. In our investigation, we
noticed that considering a pair of samples at a time and selecting the features
that bring them closer or put them far away is a better choice for feature
selection. Experimental results on benchmark data sets demonstrate the
effectiveness of the proposed method with low sample size, which outperforms
many other state-of-the-art feature selection methods.Comment: European Signal Processing Conference 201
New Identities for small hyperbolic surfaces
Luo and Tan gave a new identity for hyperbolic surfaces with/without geodesic
boundary in terms of dilogarithms of the lengths of simple closed geodesics on
embedded three-holed spheres or one-holed tori. However, the identity was
trivial for a hyperbolic one-holed torus with geodesic boundary. In this paper
we adapt the argument from Luo and Tan to give an identity for hyperbolic tori
with one geodesic boundary or cusp in terms of dilogarithm functions on the set
of lengths of simple closed geodesics on the torus. As a corollary, we are also
able to express the Luo-Tan identity as a sum over all immersed three-holed
spheres which are embeddings when restricted to the interior of .Comment: 11 pages, 4 figure
The Socialist Dilemma: A Brief Study of the History of Socialism and the Dilemma It Presented the Allied Powers During World War II
LUO Remote Online Presenter
Graduate
Textual or Investigativ
Scheduling Dimension Reduction of LPV Models -- A Deep Neural Network Approach
In this paper, the existing Scheduling Dimension Reduction (SDR) methods for
Linear Parameter-Varying (LPV) models are reviewed and a Deep Neural Network
(DNN) approach is developed that achieves higher model accuracy under
scheduling dimension reduction. The proposed DNN method and existing SDR
methods are compared on a two-link robotic manipulator, both in terms of model
accuracy and performance of controllers synthesized with the reduced models.
The methods compared include SDR for state-space models using Principal
Component Analysis (PCA), Kernel PCA (KPCA) and Autoencoders (AE). On the
robotic manipulator example, the DNN method achieves improved representation of
the matrix variations of the original LPV model in terms of the Frobenius norm
compared to the current methods. Moreover, when the resulting model is used to
accommodate synthesis, improved closed-loop performance is obtained compared to
the current methods.Comment: Accepted to American Control Conference (ACC) 2020, Denve
Rigors of Righteousness: The Puritan Prescription for Spiritual Formation
LUO Remote Online Presenter
Graduate
Textual or Investigativ
- …