300,211 research outputs found

    Analisis Kinerja Algoritma C4.5 Pada Sistem Pendukung Keputusan Penentuan Jenis Pelatihan

    Full text link
    - This study describes the application of the algorithm C4.5 on decision support systems to support trainees in PPTIK STIKI Malang in choosing the appropriate type of training. Decision support system based on several criteria derived from the data filled out by participants prior to register as a participant. Further analysis using an algorithm that is used to form a C4.5 decision tree. The decision tree is a method of classification and prediction that represent rules. the rule is then developed using RGFDT (Rule Generation From Decision Tree). Results of testing done by comparing the system with Weka and showed an accuracy of 90%.Keywords—Algorithm C4.5, Decision Support System, RGFD

    Dielectric spectra analysis: reliable parameter estimation using interval analysis

    Get PDF
    Dielectric spectroscopy is an extremely versatile method for characterizing the molecular dynamics over a large range of time scales. Unfortunately, the extraction of model parameters by data fitting is still a crucial problem which is now solved by our program S.A.D.E. S.A.D.E. is based on the algorithm S.I.V.I.A. which was proposed and implemented by Jaulin in order to solve constraint satisfaction problems. The problem of dielectric data analysis is reduced to a problem of choosing the appropriate physical model. In this article, Debye relaxations were used and validated to fit the relaxations of a DGEBA prepolymer and the polarization of the spectrometer electrodes. The conductivity was evaluated too

    Second Order Differences of Cyclic Data and Applications in Variational Denoising

    Full text link
    In many image and signal processing applications, as interferometric synthetic aperture radar (SAR), electroencephalogram (EEG) data analysis or color image restoration in HSV or LCh spaces the data has its range on the one-dimensional sphere S1\mathbb S^1. Although the minimization of total variation (TV) regularized functionals is among the most popular methods for edge-preserving image restoration such methods were only very recently applied to cyclic structures. However, as for Euclidean data, TV regularized variational methods suffer from the so called staircasing effect. This effect can be avoided by involving higher order derivatives into the functional. This is the first paper which uses higher order differences of cyclic data in regularization terms of energy functionals for image restoration. We introduce absolute higher order differences for S1\mathbb S^1-valued data in a sound way which is independent of the chosen representation system on the circle. Our absolute cyclic first order difference is just the geodesic distance between points. Similar to the geodesic distances the absolute cyclic second order differences have only values in [0,{\pi}]. We update the cyclic variational TV approach by our new cyclic second order differences. To minimize the corresponding functional we apply a cyclic proximal point method which was recently successfully proposed for Hadamard manifolds. Choosing appropriate cycles this algorithm can be implemented in an efficient way. The main steps require the evaluation of proximal mappings of our cyclic differences for which we provide analytical expressions. Under certain conditions we prove the convergence of our algorithm. Various numerical examples with artificial as well as real-world data demonstrate the advantageous performance of our algorithm.Comment: 32 pages, 16 figures, shortened version of submitted manuscrip

    Recommendation Subgraphs for Web Discovery

    Full text link
    Recommendations are central to the utility of many websites including YouTube, Quora as well as popular e-commerce stores. Such sites typically contain a set of recommendations on every product page that enables visitors to easily navigate the website. Choosing an appropriate set of recommendations at each page is one of the key features of backend engines that have been deployed at several e-commerce sites. Specifically at BloomReach, an engine consisting of several independent components analyzes and optimizes its clients' websites. This paper focuses on the structure optimizer component which improves the website navigation experience that enables the discovery of novel content. We begin by formalizing the concept of recommendations used for discovery. We formulate this as a natural graph optimization problem which in its simplest case, reduces to a bipartite matching problem. In practice, solving these matching problems requires superlinear time and is not scalable. Also, implementing simple algorithms is critical in practice because they are significantly easier to maintain in production. This motivated us to analyze three methods for solving the problem in increasing order of sophistication: a sampling algorithm, a greedy algorithm and a more involved partitioning based algorithm. We first theoretically analyze the performance of these three methods on random graph models characterizing when each method will yield a solution of sufficient quality and the parameter ranges when more sophistication is needed. We complement this by providing an empirical analysis of these algorithms on simulated and real-world production data. Our results confirm that it is not always necessary to implement complicated algorithms in the real-world and that very good practical results can be obtained by using heuristics that are backed by the confidence of concrete theoretical guarantees

    Metrics reloaded: Pitfalls and recommendations for image analysis validation

    Full text link
    The field of automatic biomedical image analysis crucially depends on robust and meaningful performance metrics for algorithm validation. Current metric usage, however, is often ill-informed and does not reflect the underlying domain interest. Here, we present a comprehensive framework that guides researchers towards choosing performance metrics in a problem-aware manner. Specifically, we focus on biomedical image analysis problems that can be interpreted as a classification task at image, object or pixel level. The framework first compiles domain interest-, target structure-, data set- and algorithm output-related properties of a given problem into a problem fingerprint, while also mapping it to the appropriate problem category, namely image-level classification, semantic segmentation, instance segmentation, or object detection. It then guides users through the process of selecting and applying a set of appropriate validation metrics while making them aware of potential pitfalls related to individual choices. In this paper, we describe the current status of the Metrics Reloaded recommendation framework, with the goal of obtaining constructive feedback from the image analysis community. The current version has been developed within an international consortium of more than 60 image analysis experts and will be made openly available as a user-friendly toolkit after community-driven optimization

    Modelling Pricing Behavior with Weak A‐Priori Information: Exploratory Approach

    Get PDF
    In the absence of reliable a priori information, choosing the appropriate theoretical model to describe an industry’s behavior is a critical issue for empirical studies about market power. A wrong choice may result in model misspecification and the conclusions of the empirical analysis may be driven by the wrong assumption about the behavioral model.This paper develops a methodology aimed to reduce the risk of misspecification bias. The approach is based on the sequential application of a sliced inverse regression (SIR) and a nonparametric Nadaraya/ Watson regression (NW). The SIR‐NW algorithm identifies the factors affecting pricing behavior in an industry and provides a nonparametric characterization of the function linking these variables to price. This information may be used to guide the choice of the model specification for a parametric estimation of market power.The SIR NW algorithm is designed to complement the estimation of structural models of market behavior, rather than to replace it. The value of this methodology for empirical industrial organization studies lies in its data driven approach that does not rely on prior knowledge of the industry. The method reverses the usual hypothesis testing approach. Instead of first choosing the model based on a priori information and then testing if it is compatible with the data, the econometrician selects a theoretical model based on the observed data. Thus, the methodology is particularly suited for those cases where the researcher has no a priori information about the behavioral model, or little confidence in the information that is available
    corecore