87,609 research outputs found
Robust artificial neural networks and outlier detection. Technical report
Large outliers break down linear and nonlinear regression models. Robust
regression methods allow one to filter out the outliers when building a model.
By replacing the traditional least squares criterion with the least trimmed
squares criterion, in which half of data is treated as potential outliers, one
can fit accurate regression models to strongly contaminated data.
High-breakdown methods have become very well established in linear regression,
but have started being applied for non-linear regression only recently. In this
work, we examine the problem of fitting artificial neural networks to
contaminated data using least trimmed squares criterion. We introduce a
penalized least trimmed squares criterion which prevents unnecessary removal of
valid data. Training of ANNs leads to a challenging non-smooth global
optimization problem. We compare the efficiency of several derivative-free
optimization methods in solving it, and show that our approach identifies the
outliers correctly when ANNs are used for nonlinear regression
Derivative-Free Global Minimization in One Dimension: Relaxation, Monte Carlo, and Sampling
We introduce a derivative-free global optimization algorithm that efficiently
computes minima for various classes of one-dimensional functions, including
non-convex, and non-smooth functions.This algorithm numerically approximates
the gradient flow of a relaxed functional, integrating strategies such as Monte
Carlos methods, rejection sampling, and adaptive techniques. These strategies
enhance performance in solving a diverse range of optimization problems while
significantly reducing the number of required function evaluations compared to
established methods. We present a proof of the convergence of the algorithm and
illustrate its performance by comprehensive benchmarking. The proposed
algorithm offers a substantial potential for real-world models. It is
particularly advantageous in situations requiring computationally intensive
objective function evaluations
Adaptive Penalty and Barrier function based on Fuzzy Logic
Optimization methods have been used in many areas of knowledge, such as Engineering, Statistics, Chemistry, among others, to solve optimization problems. In many cases it is not possible to use derivative methods, due to the characteristics of the problem to be solved and/or its constraints, for example if the involved functions are non-smooth and/or their derivatives are not know. To solve this type of problems a Java based API has been implemented, which includes only derivative-free optimization methods, and that can be used to solve both constrained and unconstrained problems. For solving constrained problems, the classic Penalty and Barrier functions were included in the API. In this paper a new approach to Penalty and Barrier functions, based on Fuzzy Logic, is proposed. Two penalty functions, that impose a progressive penalization to solutions that violate the constraints, are discussed. The implemented functions impose a low penalization when the violation of the constraints is low and a heavy penalty when the violation is high. Numerical results, obtained using twenty-eight test problems, comparing the proposed Fuzzy Logic based functions to six of the classic Penalty and Barrier functions are presented. Considering the achieved results, it can be concluded that the proposed penalty functions besides being very robust also have a very good performance
Application of Optimization Methods for Solving Clustering and Classification Problems
Cluster and classification analysis are very interesting data mining topics that can be applied in many fields. Clustering includes the identification of subsets of the data that are similar. Intuitively, samples within a valid cluster are more similar to each other than they are to a sample belonging to a different cluster. Samples in the same cluster have the same label. The aim of data classification is to set up rules for the classification of
some observations that the classes of data are supposed to be known. Here, there is a collection of classes with labels and the problem is to label a new observation or data
point belonging to one or more classes of data. The focus of this thesis is on solvingclustering and classification problems. Specifically, we will focus on new optimization
methods for solving clustering and classification problems. First we briefly give some data analysis background. Then a review of different methods currently available that
can be used to solve clustering and classification problems is also given.
Clustering problem is discussed as a problem of non-smooth, non-convex optimization and a new method for solving this optimization problem is developed. This optimization
problem has a number of characteristics that make it challenging: it has many local minimum, the optimization variables can be either continuous or categorical, and there
are no exact analytical derivatives. In this study we show how to apply a particular class of optimization methods known as pattern search methods to address these challenges.
This method does not explicitly use derivatives, and is particularly appropriate when functions are non-smooth. Also a new algorithm for finding the initial point is proposed.
We have established that our proposed method can produce excellent results compared to those previously known methods. Results of computational experiments on real data
sets present the robustness and advantage of the new method. Next the problem of data classification is studied as a problem of global, non-smooth and non-convex
optimization; this approach consists of describing clusters for the given training sets.
The data vectors are assigned to the closest cluster and correspondingly to the set, which contains this cluster and an algorithm based on a derivative-free method is applied to the solution of this problem. The proposed method has been tested on real-world datasets.
Results of numerical experiments have been presented which demonstrate the effectiveness of the proposed algorithm
- …