235 research outputs found

    Naive possibilistic classifiers for imprecise or uncertain numerical data

    Get PDF
    International audienceIn real-world problems, input data may be pervaded with uncertainty. In this paper, we investigate the behavior of naive possibilistic classifiers, as a counterpart to naive Bayesian ones, for dealing with classification tasks in the presence of uncertainty. For this purpose, we extend possibilistic classifiers, which have been recently adapted to numerical data, in order to cope with uncertainty in data representation. Here the possibility distributions that are used are supposed to encode the family of Gaussian probabilistic distributions that are compatible with the considered dataset. We consider two types of uncertainty: (i) the uncertainty associated with the class in the training set, which is modeled by a possibility distribution over class labels, and (ii) the imprecision pervading attribute values in the testing set represented under the form of intervals for continuous data. Moreover, the approach takes into account the uncertainty about the estimation of the Gaussian distribution parameters due to the limited amount of data available. We first adapt the possibilistic classification model, previously proposed for the certain case, in order to accommodate the uncertainty about class labels. Then, we propose an algorithm based on the extension principle to deal with imprecise attribute values. The experiments reported show the interest of possibilistic classifiers for handling uncertainty in data. In particular, the probability-to-possibility transform-based classifier shows a robust behavior when dealing with imperfect data

    An overview of flexibility and generalized uncertainty in optimization

    Full text link

    Possibilistic classifiers for numerical data

    Get PDF
    International audienceNaive Bayesian Classifiers, which rely on independence hypotheses, together with a normality assumption to estimate densities for numerical data, are known for their simplicity and their effectiveness. However, estimating densities, even under the normality assumption, may be problematic in case of poor data. In such a situation, possibility distributions may provide a more faithful representation of these data. Naive Possibilistic Classifiers (NPC), based on possibility theory, have been recently proposed as a counterpart of Bayesian classifiers to deal with classification tasks. There are only few works that treat possibilistic classification and most of existing NPC deal only with categorical attributes. This work focuses on the estimation of possibility distributions for continuous data. In this paper we investigate two kinds of possibilistic classifiers. The first one is derived from classical or flexible Bayesian classifiers by applying a probability–possibility transformation to Gaussian distributions, which introduces some further tolerance in the description of classes. The second one is based on a direct interpretation of data in possibilistic formats that exploit an idea of proximity between data values in different ways, which provides a less constrained representation of them. We show that possibilistic classifiers have a better capability to detect new instances for which the classification is ambiguous than Bayesian classifiers, where probabilities may be poorly estimated and illusorily precise. Moreover, we propose, in this case, an hybrid possibilistic classification approach based on a nearest-neighbour heuristics to improve the accuracy of the proposed possibilistic classifiers when the available information is insufficient to choose between classes. Possibilistic classifiers are compared with classical or flexible Bayesian classifiers on a collection of benchmarks databases. The experiments reported show the interest of possibilistic classifiers. In particular, flexible possibilistic classifiers perform well for data agreeing with the normality assumption, while proximity-based possibilistic classifiers outperform others in the other cases. The hybrid possibilistic classification exhibits a good ability for improving accuracy

    Informational Paradigm, management of uncertainty and theoretical formalisms in the clustering framework: A review

    Get PDF
    Fifty years have gone by since the publication of the first paper on clustering based on fuzzy sets theory. In 1965, L.A. Zadeh had published “Fuzzy Sets” [335]. After only one year, the first effects of this seminal paper began to emerge, with the pioneering paper on clustering by Bellman, Kalaba, Zadeh [33], in which they proposed a prototypal of clustering algorithm based on the fuzzy sets theory

    Observation of temporary accommodation for construction workers according to the code of practice for temporary construction site workers amenities and accommodation (ms2593:2015) in Johor, Malaysia

    Get PDF
    The Malaysian government is currently improving the quality of workers temporary accommodation by introducing MS2593:2015 (Code of Practice for Temporary Site Workers Amenities and Accommodation) in 2015. It is in line with the initiative in the Construction Industry Transformation Programme (2016-2020) to increase the quality and well-being of construction workers in Malaysia. Thus, to gauge the current practice of temporary accommodation on complying with the particular guideline, this paper has put forth the observation of such accommodation towards elements in Section 3 within MS2593:2015. A total of seventeen (17) temporary accommodation provided by Grade 6 and Grade 7 contractors in Johor were selected and assessed. The results disclosed that most of the temporary accommodation was not complying with the guideline, where only thirteen (13) out of fifty-eight (58) elements have recorded full compliance (100%), and the lowest compliance percentage (5.9%) are discovered in the Section 3.12 (Signage). In a nutshell, given the significant gap of compliance between current practices of temporary accommodation and MS2593:2015, a holistic initiative need to be in place for the guideline to be worthwhile

    Numerical treatment of imprecise random fields in non-linear solid mechanics

    Get PDF
    The quantification and propagation of mixed uncertain material parameters in the context of solid mechanical finite element simulations is studied. While aleatory uncertainties appear in terms of spatial varying parameters, i.e. random fields, the epistemic character is induced by a lack of knowledge regarding the correlation length, which is therefore described by interval values. The concept and description of the resulting imprecise random fields is introduced in detail. The challenges occurring from interval valued correlation lengths are clarified. These include mainly the stochastic dimension, which can become very high under some circumstances, as well as the comparability of different correlation length scenarios with regard to the underlying truncation error of the applied Karhunen-LoĂšve expansion. Additionally, the computation time can increase drastically, if the straightforward and robust double loop approach is applied. Sparse stochastic collocation method and sparse polynomial chaos expansion are studied to reduce the number of required sample evaluations, i.e. the computational cost. To keep the stochastic dimension as low as possible, the random fields are described by Karhunen-LoĂšve expansion, using a modified exponential correlation kernel, which is advantageous in terms of a fast convergence while providing an analytic solution. Still, for small correlation lengths, the investigated approaches are limited by the curse of dimensionality. Furthermore, they turn out to be not suited for non-linear material models. As a straightforward alternative, a decoupled interpolation approach is proposed, offering a practical engineering estimate. For this purpose, the uncertain quantities only need to be propagated as a random variable and deterministically in terms of the mean values. From these results, the so-called absolutely no idea probability box (ani-p-box) can be obtained, bounding the results of the interval valued correlation length being between zero and infinity. The idea is, to interpolate the result of any arbitrary correlation length within this ani-p-box, exploiting prior knowledge about the statistical behaviour of the input random field corresponding to the correlation length. The new approach is studied for one- and two-dimensional random fields. Furthermore, linear and non-linear finite element models are used in terms of linear-elastic or elasto-plastic material laws, the latter including linear hardening. It appears that the approach only works satisfyingly for sufficiently smooth responses but an improvement by considering also higher order statistics is motivated for future research.DFG/SPP 1886/NA330/12-1/E

    Clustering of nonstationary data streams: a survey of fuzzy partitional methods

    Get PDF
    YesData streams have arisen as a relevant research topic during the past decade. They are real‐time, incremental in nature, temporally ordered, massive, contain outliers, and the objects in a data stream may evolve over time (concept drift). Clustering is often one of the earliest and most important steps in the streaming data analysis workflow. A comprehensive literature is available about stream data clustering; however, less attention is devoted to the fuzzy clustering approach, even though the nonstationary nature of many data streams makes it especially appealing. This survey discusses relevant data stream clustering algorithms focusing mainly on fuzzy methods, including their treatment of outliers and concept drift and shift.Ministero dell‘Istruzione, dell‘Universitá e della Ricerca
    • 

    corecore