116 research outputs found
Feature Reduction Based on Sum-of-SNR (SOSNR) Optimization
Dimensionality reduction plays an important role in machine learning techniques. In classification, data transformation aims to reduce the number of feature dimensions, whereas attempts to enhance the class separability. To this end, we propose a new classifier-independent criterion called 'Sum-of-Signal-to-Noise-Ratio' (SoSNR). A framework designed for maximization with respect to this criterion is presented and three types of algorithms, respectively based on (1) gradient, (2) deflation and (3) sparsity, are proposed. The techniques are conducted on standard UCI databases and compared to other related methods. Results show trade-offs between computational complexity and classification accuracy among different approaches
Ridge-Adjusted Slack Variable Optimization for Supervised Classification
This paper presents an iterative classification algorithm called Ridge-adjusted Slack Variable Optimization (RiSVO). RiSVO is an iterative procedure with two steps: (1) A working subset of the training data is selected so as to reject "extreme" patterns. (2) the decision vector and threshold value are obtained by minimizing the energy function associated with the slack variables. From a computational perspective, we have established a sufficient condition for the "inclusion property" among successive working sets, which allows us to save computation time. Most importantly, under the inclusion property, the monotonic reduction of the energy function can be assured in both substeps at each iteration, thus assuring the convergence of the algorithm. Moreover, ridge regularization is incorporated to improve the robustness and better cope with over-fitting and ill-conditioned problems. To verify the proposed algorithm, we conducted simulations on three data sets from the UCI database: adult, shuttle and bank. Our simulation shows stability and convergence of the RiSVO method. The results also show improvement of performance over the SVM classifier
Feature Selection for Genomic Signal Processing: Unsupervised, Supervised, and Self-Supervised Scenarios
effective data mining system lies in the representation of pattern vectors. For many bioinformatic applications, data are represented as vectors of extremely high dimension. This motivates the research on feature selection. In the literature, there are plenty of reports on feature selection methods. In terms of training data types, they are divided into the unsupervised and supervised categories. In terms of selection methods, they fall into filter and wrapper categories. This paper will provide a brief overview on the state-of-the-arts feature selection methods on all these categories. Sample applications of these methods for genomic signal processing will be highlighted. This paper also describes a notion of self-supervision. A special method called vector index adaptive SVM (VIA-SVM) is described for selecting features under the self-supervision scenario. Furthermore, the paper makes use of a more powerful symmetric doubly supervised formulation, for which VIA-SVM is particularly useful. Based on several subcellular localization experiments, and microarray time course experiments, the VIA-SVM algorithm when combined with some filter-type metrics appears to deliver a substantial dimension reduction (one-order of magnitude) with only little degradation on accuracy
On Using High-Definition Body Worn Cameras for Face Recognition from a Distance
Recognition of human faces from a distance is highly desirable for law-enforcement. This paper evaluates the use of low-cost, high-definition (HD) body worn video cameras for face recognition from a distance. A comparison of HD vs. Standard-definition (SD) video for face recognition from a distance is presented. HD and SD videos of 20 subjects were acquired in different conditions and at varying distances. The evaluation uses three benchmark algorithms: Eigenfaces, Fisherfaces and Wavelet Transforms. The study indicates when gallery and probe images consist of faces captured from a distance, HD video result in better recognition accuracy, compared to SD video. This scenario resembles real-life conditions of video surveillance and law-enforcement activities. However, at a close range, face data obtained from SD video result in similar, if not better recognition accuracy than using HD face data of the same range
Modified f(G) gravity models with curvature-matter coupling
A modified f(G) gravity model with coupling between matter and geometry is
proposed, which is described by the product of the Lagrange density of the
matter and an arbitrary function of the Gauss-Bonnet term. The field equations
and the equations of motion corresponding to this model show the
non-conservation of the energy-momentum tensor, the presence of an extra-force
acting on test particles and the non-geodesic motion. Moreover, the energy
conditions and the stability criterion at de Sitter point in the modified f(G)
gravity models with curvature-matter coupling are derived, which can degenerate
to the well-known energy conditions in general relativity. Furthermore, in
order to get some insight on the meaning of these energy conditions, we apply
them to the specific models of f(G) gravity and the corresponding constraints
on the models are given. In addition, the conditions and the candidate for
late-time cosmic accelerated expansion in the modified f(G) gravity are studied
by means of conditions of power-law expansion and the equation of state of
matter less than -1/ 3 .Comment: 13 pages, 4 figure
Measurement of the Bottom-Strange Meson Mixing Phase in the Full CDF Data Set
We report a measurement of the bottom-strange meson mixing phase \beta_s
using the time evolution of B0_s -> J/\psi (->\mu+\mu-) \phi (-> K+ K-) decays
in which the quark-flavor content of the bottom-strange meson is identified at
production. This measurement uses the full data set of proton-antiproton
collisions at sqrt(s)= 1.96 TeV collected by the Collider Detector experiment
at the Fermilab Tevatron, corresponding to 9.6 fb-1 of integrated luminosity.
We report confidence regions in the two-dimensional space of \beta_s and the
B0_s decay-width difference \Delta\Gamma_s, and measure \beta_s in [-\pi/2,
-1.51] U [-0.06, 0.30] U [1.26, \pi/2] at the 68% confidence level, in
agreement with the standard model expectation. Assuming the standard model
value of \beta_s, we also determine \Delta\Gamma_s = 0.068 +- 0.026 (stat) +-
0.009 (syst) ps-1 and the mean B0_s lifetime, \tau_s = 1.528 +- 0.019 (stat) +-
0.009 (syst) ps, which are consistent and competitive with determinations by
other experiments.Comment: 8 pages, 2 figures, Phys. Rev. Lett 109, 171802 (2012
The Public Repository of Xenografts enables discovery and randomized phase II-like trials in mice
More than 90% of drugs with preclinical activity fail in human trials, largely due to insufficient efficacy. We hypothesized that adequately powered trials of patient-derived xenografts (PDX) in mice could efficiently define therapeutic activity across heterogeneous tumors. To address this hypothesis, we established a large, publicly available repository of well-characterized leukemia and lymphoma PDXs that undergo orthotopic engraftment, called the Public Repository of Xenografts (PRoXe). PRoXe includes all de-identified information relevant to the primary specimens and the PDXs derived from them. Using this repository, we demonstrate that large studies of acute leukemia PDXs that mimic human randomized clinical trials can characterize drug efficacy and generate transcriptional, functional, and proteomic biomarkers in both treatment-naive and relapsed/refractory disease
- …