28,339 research outputs found
Magnetic operations: a little fuzzy physics?
We examine the behaviour of charged particles in homogeneous, constant and/or
oscillating magnetic fields in the non-relativistic approximation. A special
role of the geometric center of the particle trajectory is elucidated. In
quantum case it becomes a 'fuzzy point' with non-commuting coordinates, an
element of non-commutative geometry which enters into the traditional control
problems. We show that its application extends beyond the usually considered
time independent magnetic fields of the quantum Hall effect. Some simple cases
of magnetic control by oscillating fields lead to the stability maps differing
from the traditional Strutt diagram.Comment: 28 pages, 8 figure
On the incorporation of interval-valued fuzzy sets into the Bousi-Prolog system: declarative semantics, implementation and applications
In this paper we analyse the benefits of incorporating interval-valued fuzzy
sets into the Bousi-Prolog system. A syntax, declarative semantics and im-
plementation for this extension is presented and formalised. We show, by using
potential applications, that fuzzy logic programming frameworks enhanced with
them can correctly work together with lexical resources and ontologies in order
to improve their capabilities for knowledge representation and reasoning
Learning Opposites Using Neural Networks
Many research works have successfully extended algorithms such as
evolutionary algorithms, reinforcement agents and neural networks using
"opposition-based learning" (OBL). Two types of the "opposites" have been
defined in the literature, namely \textit{type-I} and \textit{type-II}. The
former are linear in nature and applicable to the variable space, hence easy to
calculate. On the other hand, type-II opposites capture the "oppositeness" in
the output space. In fact, type-I opposites are considered a special case of
type-II opposites where inputs and outputs have a linear relationship. However,
in many real-world problems, inputs and outputs do in fact exhibit a nonlinear
relationship. Therefore, type-II opposites are expected to be better in
capturing the sense of "opposition" in terms of the input-output relation. In
the absence of any knowledge about the problem at hand, there seems to be no
intuitive way to calculate the type-II opposites. In this paper, we introduce
an approach to learn type-II opposites from the given inputs and their outputs
using the artificial neural networks (ANNs). We first perform \emph{opposition
mining} on the sample data, and then use the mined data to learn the
relationship between input and its opposite . We have validated
our algorithm using various benchmark functions to compare it against an
evolving fuzzy inference approach that has been recently introduced. The
results show the better performance of a neural approach to learn the
opposites. This will create new possibilities for integrating oppositional
schemes within existing algorithms promising a potential increase in
convergence speed and/or accuracy.Comment: To appear in proceedings of the 23rd International Conference on
Pattern Recognition (ICPR 2016), Cancun, Mexico, December 201
Fuzzy ARTMAP: A Neural Network Architecture for Incremental Supervised Learning of Analog Multidimensional Maps
A new neural network architecture is introduced for incremental supervised learning of recognition categories and multidimensional maps in response to arbitrary sequences of analog or binary input vectors. The architecture, called Fuzzy ARTMAP, achieves a synthesis of fuzzy logic and Adaptive Resonance Theory (ART) neural networks by exploiting a close formal similarity between the computations of fuzzy subsethood and ART category choice, resonance, and learning. Fuzzy ARTMAP also realizes a new Minimax Learning Rule that conjointly minimizes predictive error and maximizes code compression, or generalization. This is achieved by a match tracking process that increases the ART vigilance parameter by the minimum amount needed to correct a predictive error. As a result, the system automatically learns a minimal number of recognition categories, or "hidden units", to met accuracy criteria. Category proliferation is prevented by normalizing input vectors at a preprocessing stage. A normalization procedure called complement coding leads to a symmetric theory in which the MIN operator (Λ) and the MAX operator (v) of fuzzy logic play complementary roles. Complement coding uses on-cells and off-cells to represent the input pattern, and preserves individual feature amplitudes while normalizing the total on-cell/off-cell vector. Learning is stable because all adaptive weights can only decrease in time. Decreasing weights correspond to increasing sizes of category "boxes". Smaller vigilance values lead to larger category boxes. Improved prediction is achieved by training the system several times using different orderings of the input set. This voting strategy can also be used to assign probability estimates to competing predictions given small, noisy, or incomplete training sets. Four classes of simulations illustrate Fuzzy ARTMAP performance as compared to benchmark back propagation and genetic algorithm systems. These simulations include (i) finding points inside vs. outside a circle; (ii) learning to tell two spirals apart; (iii) incremental approximation of a piecewise continuous function; and (iv) a letter recognition database. The Fuzzy ARTMAP system is also compared to Salzberg's NGE system and to Simpson's FMMC system.British Petroleum (89-A-1204); Defense Advanced Research Projects Agency (90-0083); National Science Foundation (IRI 90-00530); Office of Naval Research (N00014-91-J-4100); Air Force Office of Scientific Research (90-0175
Unifying Practical Uncertainty Representations: II. Clouds
There exist many simple tools for jointly capturing variability and
incomplete information by means of uncertainty representations. Among them are
random sets, possibility distributions, probability intervals, and the more
recent Ferson's p-boxes and Neumaier's clouds, both defined by pairs of
possibility distributions. In the companion paper, we have extensively studied
a generalized form of p-box and situated it with respect to other models . This
paper focuses on the links between clouds and other representations.
Generalized p-boxes are shown to be clouds with comonotonic distributions. In
general, clouds cannot always be represented by random sets, in fact not even
by 2-monotone (convex) capacities.Comment: 30 pages, 7 figures, Pre-print of journal paper to be published in
International Journal of Approximate Reasoning (with expanded section
concerning clouds and probability intervals
- …