300,211 research outputs found
Analisis Kinerja Algoritma C4.5 Pada Sistem Pendukung Keputusan Penentuan Jenis Pelatihan
- This study describes the application of the algorithm C4.5 on decision support systems to support trainees in PPTIK STIKI Malang in choosing the appropriate type of training. Decision support system based on several criteria derived from the data filled out by participants prior to register as a participant. Further analysis using an algorithm that is used to form a C4.5 decision tree. The decision tree is a method of classification and prediction that represent rules. the rule is then developed using RGFDT (Rule Generation From Decision Tree). Results of testing done by comparing the system with Weka and showed an accuracy of 90%.Keywords—Algorithm C4.5, Decision Support System, RGFD
Dielectric spectra analysis: reliable parameter estimation using interval analysis
Dielectric spectroscopy is an extremely versatile method for characterizing the molecular dynamics over a large range of time scales. Unfortunately, the extraction of model parameters by data fitting is still a crucial problem which is now solved by our program S.A.D.E. S.A.D.E. is based on the algorithm S.I.V.I.A. which was proposed and implemented by Jaulin in order to solve constraint satisfaction problems. The problem of dielectric data analysis is reduced to a problem of choosing the appropriate physical model. In this article, Debye relaxations were used and validated to fit the relaxations of a DGEBA prepolymer and the polarization of the spectrometer electrodes. The conductivity was evaluated too
Second Order Differences of Cyclic Data and Applications in Variational Denoising
In many image and signal processing applications, as interferometric
synthetic aperture radar (SAR), electroencephalogram (EEG) data analysis or
color image restoration in HSV or LCh spaces the data has its range on the
one-dimensional sphere . Although the minimization of total
variation (TV) regularized functionals is among the most popular methods for
edge-preserving image restoration such methods were only very recently applied
to cyclic structures. However, as for Euclidean data, TV regularized
variational methods suffer from the so called staircasing effect. This effect
can be avoided by involving higher order derivatives into the functional.
This is the first paper which uses higher order differences of cyclic data in
regularization terms of energy functionals for image restoration. We introduce
absolute higher order differences for -valued data in a sound way
which is independent of the chosen representation system on the circle. Our
absolute cyclic first order difference is just the geodesic distance between
points. Similar to the geodesic distances the absolute cyclic second order
differences have only values in [0,{\pi}]. We update the cyclic variational TV
approach by our new cyclic second order differences. To minimize the
corresponding functional we apply a cyclic proximal point method which was
recently successfully proposed for Hadamard manifolds. Choosing appropriate
cycles this algorithm can be implemented in an efficient way. The main steps
require the evaluation of proximal mappings of our cyclic differences for which
we provide analytical expressions. Under certain conditions we prove the
convergence of our algorithm. Various numerical examples with artificial as
well as real-world data demonstrate the advantageous performance of our
algorithm.Comment: 32 pages, 16 figures, shortened version of submitted manuscrip
Recommendation Subgraphs for Web Discovery
Recommendations are central to the utility of many websites including
YouTube, Quora as well as popular e-commerce stores. Such sites typically
contain a set of recommendations on every product page that enables visitors to
easily navigate the website. Choosing an appropriate set of recommendations at
each page is one of the key features of backend engines that have been deployed
at several e-commerce sites.
Specifically at BloomReach, an engine consisting of several independent
components analyzes and optimizes its clients' websites. This paper focuses on
the structure optimizer component which improves the website navigation
experience that enables the discovery of novel content.
We begin by formalizing the concept of recommendations used for discovery. We
formulate this as a natural graph optimization problem which in its simplest
case, reduces to a bipartite matching problem. In practice, solving these
matching problems requires superlinear time and is not scalable. Also,
implementing simple algorithms is critical in practice because they are
significantly easier to maintain in production. This motivated us to analyze
three methods for solving the problem in increasing order of sophistication: a
sampling algorithm, a greedy algorithm and a more involved partitioning based
algorithm.
We first theoretically analyze the performance of these three methods on
random graph models characterizing when each method will yield a solution of
sufficient quality and the parameter ranges when more sophistication is needed.
We complement this by providing an empirical analysis of these algorithms on
simulated and real-world production data. Our results confirm that it is not
always necessary to implement complicated algorithms in the real-world and that
very good practical results can be obtained by using heuristics that are backed
by the confidence of concrete theoretical guarantees
Metrics reloaded: Pitfalls and recommendations for image analysis validation
The field of automatic biomedical image analysis crucially depends on robust and meaningful performance metrics for algorithm validation. Current metric usage, however, is often ill-informed and does not reflect the underlying domain interest. Here, we present a comprehensive framework that guides researchers towards choosing performance metrics in a problem-aware manner. Specifically, we focus on biomedical image analysis problems that can be interpreted as a classification task at image, object or pixel level. The framework first compiles domain interest-, target structure-, data set- and algorithm output-related properties of a given problem into a problem fingerprint, while also mapping it to the appropriate problem category, namely image-level classification, semantic segmentation, instance segmentation, or object detection. It then guides users through the process of selecting and applying a set of appropriate validation metrics while making them aware of potential pitfalls related to individual choices. In this paper, we describe the current status of the Metrics Reloaded recommendation framework, with the goal of obtaining constructive feedback from the image analysis community. The current version has been developed within an international consortium of more than 60 image analysis experts and will be made openly available as a user-friendly toolkit after community-driven optimization
Modelling Pricing Behavior with Weak A‐Priori Information: Exploratory Approach
In the absence of reliable a priori information, choosing the appropriate theoretical model to describe an industry’s behavior is a critical issue for empirical studies about market power. A wrong choice may result in model misspecification and the conclusions of the empirical analysis may be driven by the wrong assumption about the behavioral model.This paper develops a methodology aimed to reduce the risk of misspecification bias. The approach is based on the sequential application of a sliced inverse regression (SIR) and a nonparametric Nadaraya/ Watson regression (NW). The SIR‐NW algorithm identifies the factors affecting pricing behavior in an industry and provides a nonparametric characterization of the function linking these variables to price. This information may be used to guide the choice of the model specification for a parametric estimation of market power.The SIR NW algorithm is designed to complement the estimation of structural models of market behavior, rather than to replace it. The value of this methodology for empirical industrial organization studies lies in its data driven approach that does not rely on prior knowledge of the industry. The method reverses the usual hypothesis testing approach. Instead of first choosing the model based on a priori information and then testing if it is compatible with the data, the econometrician selects a theoretical model based on the observed data. Thus, the methodology is particularly suited for those cases where the researcher has no a priori information about the behavioral model, or little confidence in the information that is available
- …