262 research outputs found

    Nonparallel support vector machines for pattern classification

    Get PDF
    We propose a novel nonparallel classifier, called nonparallel support vector machine (NPSVM), for binary classification. Our NPSVM that is fully different from the existing nonparallel classifiers, such as the generalized eigenvalue proximal support vector machine (GEPSVM) and the twin support vector machine (TWSVM), has several incomparable advantages: 1) two primal problems are constructed implementing the structural risk minimization principle; 2) the dual problems of these two primal problems have the same advantages as that of the standard SVMs, so that the kernel trick can be applied directly, while existing TWSVMs have to construct another two primal problems for nonlinear cases based on the approximate kernel-generated surfaces, furthermore, their nonlinear problems cannot degenerate to the linear case even the linear kernel is used; 3) the dual problems have the same elegant formulation with that of standard SVMs and can certainly be solved efficiently by sequential minimization optimization algorithm, while existing GEPSVM or TWSVMs are not suitable for large scale problems; 4) it has the inherent sparseness as standard SVMs; 5) existing TWSVMs are only the special cases of the NPSVM when the parameters of which are appropriately chosen. Experimental results on lots of datasets show the effectiveness of our method in both sparseness and classification accuracy, and therefore, confirm the above conclusion further. In some sense, our NPSVM is a new starting point of nonparallel classifiers

    Multi-view Facial Landmark Detection

    Get PDF
    In this thesis, we tackle the problem of designing a multi-view facial landmark detector which is robust and works in real-time on low-end hardware. Our landmark detector is an instance of the structured output classi ers describing the face by a mixture of tree based Deformable Part Models (DPM). We propose to learn parameters of the detector by the Structured Output Support Vector Machine algorithm which, in contrast to existing methods, directly optimizes a loss function closely related to the standard evaluation metrics used in landmark detection. We also propose a novel two-stage approach to learn the multi-view landmark detectors, which provides better localization accuracy and signi cantly reduces the overall learning time. We propose several speedups that enable to use the globally optimal prediction strategy based on the dynamic programming in real time even for dense landmark sets. The empirical evaluation shows that the proposed detector is competitive with the current state-ofthe- art both regarding the accuracy and speed. We also propose two improvements of the Bundle Method for Regularized Risk Minimization (BMRM) algorithm which is among the most popular batch solvers used in structured output learning. First, we propose to augment the objective function by a quadratic prox-center whose strength is controlled by a novel adaptive strategy preventing zig-zag behavior in the cases when the genuine regularization term is weak. Second, we propose to speed up convergence by using multiple cutting plane models which better approximate the objective function with minimal increase in the computational cost. Experimental evaluation shows that the new BMRM algorithm which uses both improvements speeds up learning up to an order of magnitude on standard computer vision benchmarks, and 3 to 4 times when applied to the learning of the DPM based landmark detector. vKatedra kybernetik

    A kernel-free L1 norm regularized ν-support vector machine model with application

    Get PDF
    With a view to overcoming a few shortcomings resulting from the kernel-based SVM models, these kernel-free support vector machine (SVM) models are newly promoted and researched. With the aim of deeply enhancing the classification accuracy of present kernel-free quadratic surface support vector machine (QSSVM) models while avoiding computational complexity, an emerging kernel-free ν-fuzzy reduced QSSVM with L1 norm regularization model is proposed. The model has well-developed sparsity to avoid computational complexity and overfitting and has been simplified as these standard linear models on condition that the data points are (nearly) linearly separable. Computational tests are implemented on several public benchmark datasets for the purpose of showing the better performance of the presented model compared with a few known binary classification models. Similarly, the numerical consequences support the more elevated training effectiveness of the presented model in comparison with those of other kernel-free SVM models. What`s more, the presented model is smoothly employed in lung cancer subtype diagnosis with good performance, by using the gene expression RNAseq-based lung cancer subtype (LUAD/LUSC) dataset in the TCGA database

    Data-Driven and Hybrid Methods for Naval Applications

    Get PDF
    The goal of this PhD thesis is to study, design and develop data analysis methods for naval applications. Data analysis is improving our ways to understand complex phenomena by profitably taking advantage of the information laying behind a collection of data. In fact, by adopting algorithms coming from the world of statistics and machine learning it is possible to extract valuable information, without requiring specific domain knowledge of the system generating the data. The application of such methods to marine contexts opens new research scenarios, since typical naval problems can now be solved with higher accuracy rates with respect to more classical techniques, based on the physical equations governing the naval system. During this study, some major naval problems have been addressed adopting state-of-the-art and novel data analysis techniques: condition-based maintenance, consisting in assets monitoring, maintenance planning, and real-time anomaly detection; energy and consumption monitoring, in order to reduce vessel consumption and gas emissions; system safety for maneuvering control and collision avoidance; components design, in order to detect possible defects at design stage. A review of the state-of-the-art of data analysis and machine learning techniques together with the preliminary results of the application of such methods to the aforementioned problems show a growing interest in these research topics and that effective data-driven solutions can be applied to the naval context. Moreover, for some applications, data-driven models have been used in conjunction with domain-dependent methods, modelling physical phenomena, in order to exploit both mechanistic knowledge of the system and available measurements. These hybrid methods are proved to provide more accurate and interpretable results with respect to both the pure physical or data-driven approaches taken singularly, thus showing that in the naval context it is possible to offer new valuable methodologies by either providing novel statistical methods or improving the state-of-the-art ones

    The Effect of Malaysia General Election on Financial Network: An Evidence from Shariah-Compliant Stocks on Bursa Malaysia

    Get PDF
    Instead of focusing the volatility of the market, the market participants should consider on how the general election affects the correlation between the stocks during 14th general election Malaysia. The 14th general election of Malaysia was held on 9th May 2018. This event has a great impact towards the stocks listed on Bursa Malaysia. Thus, this study investigates the effect of 14th general election Malaysia towards the correlation between stock in Bursa Malaysia specifically the shariah-compliant stock. In addition, this paper examines the changes in terms of network topology for the duration, sixth months before and after the general election. The minimum spanning tree was used to visualize the correlation between the stocks. Also, the centrality measure, namely degree, closeness and betweenness were computed to identify if any changes of stocks that plays a crucial role in the network for the duration of before and after 14th general election Malaysia

    Towards Learning Representations in Visual Computing Tasks

    Get PDF
    abstract: The performance of most of the visual computing tasks depends on the quality of the features extracted from the raw data. Insightful feature representation increases the performance of many learning algorithms by exposing the underlying explanatory factors of the output for the unobserved input. A good representation should also handle anomalies in the data such as missing samples and noisy input caused by the undesired, external factors of variation. It should also reduce the data redundancy. Over the years, many feature extraction processes have been invented to produce good representations of raw images and videos. The feature extraction processes can be categorized into three groups. The first group contains processes that are hand-crafted for a specific task. Hand-engineering features requires the knowledge of domain experts and manual labor. However, the feature extraction process is interpretable and explainable. Next group contains the latent-feature extraction processes. While the original feature lies in a high-dimensional space, the relevant factors for a task often lie on a lower dimensional manifold. The latent-feature extraction employs hidden variables to expose the underlying data properties that cannot be directly measured from the input. Latent features seek a specific structure such as sparsity or low-rank into the derived representation through sophisticated optimization techniques. The last category is that of deep features. These are obtained by passing raw input data with minimal pre-processing through a deep network. Its parameters are computed by iteratively minimizing a task-based loss. In this dissertation, I present four pieces of work where I create and learn suitable data representations. The first task employs hand-crafted features to perform clinically-relevant retrieval of diabetic retinopathy images. The second task uses latent features to perform content-adaptive image enhancement. The third task ranks a pair of images based on their aestheticism. The goal of the last task is to capture localized image artifacts in small datasets with patch-level labels. For both these tasks, I propose novel deep architectures and show significant improvement over the previous state-of-art approaches. A suitable combination of feature representations augmented with an appropriate learning approach can increase performance for most visual computing tasks.Dissertation/ThesisDoctoral Dissertation Computer Science 201

    Human Metaphase Chromosome Analysis using Image Processing

    Get PDF
    Development of an effective human metaphase chromosome analysis algorithm can optimize expert time usage by increasing the efficiency of many clinical diagnosis processes. Although many methods exist in the literature, they are only applicable for limited morphological variations and are specific to the staining method used during cell preparation. They are also highly influenced by irregular chromosome boundaries as well as the presence of artifacts such as premature sister chromatid separation. Therefore an algorithm is proposed in this research which can operate with any morphological variation of the chromosome across images from multiple staining methods. The proposed algorithm is capable of calculating the segmentation outline, the centerline (which gives the chromosome length), partitioning of the telomere regions and the centromere location of a given chromosome. The algorithm also detects and corrects for the sister chromatid separation artifact in metaphase cell images. A metric termed the Candidate Based Centromere Confidence (CBCC) is proposed to accompany each centromere detection result of the proposed method, giving an indication of the confidence the algorithm has on a given localization. The proposed method was first tested for the ability of calculating an accurate width profile against a centerline based method [1] using 226 chromosomes. A statistical analysis of the centromere detection error values proved that the proposed method can accurately locate centromere locations with statistical significance. Furthermore, the proposed method performed more consistently across different staining methods in comparison to the centerline based approach. When tested with a larger data set of 1400 chromosomes collected from a set of DAPI (4\u27,6-diamidino-2-phenylindole) and Giemsa stained cell images, the proposed candidate based centromere detection algorithm was able to accurately localize 1220 centromere locations yielding a detection accuracy of 87%
    corecore