11 research outputs found
Fuzzy ARTMAP: A Neural Network Architecture for Incremental Supervised Learning of Analog Multidimensional Maps
A new neural network architecture is introduced for incremental supervised learning of recognition categories and multidimensional maps in response to arbitrary sequences of analog or binary input vectors. The architecture, called Fuzzy ARTMAP, achieves a synthesis of fuzzy logic and Adaptive Resonance Theory (ART) neural networks by exploiting a close formal similarity between the computations of fuzzy subsethood and ART category choice, resonance, and learning. Fuzzy ARTMAP also realizes a new Minimax Learning Rule that conjointly minimizes predictive error and maximizes code compression, or generalization. This is achieved by a match tracking process that increases the ART vigilance parameter by the minimum amount needed to correct a predictive error. As a result, the system automatically learns a minimal number of recognition categories, or "hidden units", to met accuracy criteria. Category proliferation is prevented by normalizing input vectors at a preprocessing stage. A normalization procedure called complement coding leads to a symmetric theory in which the MIN operator (Λ) and the MAX operator (v) of fuzzy logic play complementary roles. Complement coding uses on-cells and off-cells to represent the input pattern, and preserves individual feature amplitudes while normalizing the total on-cell/off-cell vector. Learning is stable because all adaptive weights can only decrease in time. Decreasing weights correspond to increasing sizes of category "boxes". Smaller vigilance values lead to larger category boxes. Improved prediction is achieved by training the system several times using different orderings of the input set. This voting strategy can also be used to assign probability estimates to competing predictions given small, noisy, or incomplete training sets. Four classes of simulations illustrate Fuzzy ARTMAP performance as compared to benchmark back propagation and genetic algorithm systems. These simulations include (i) finding points inside vs. outside a circle; (ii) learning to tell two spirals apart; (iii) incremental approximation of a piecewise continuous function; and (iv) a letter recognition database. The Fuzzy ARTMAP system is also compared to Salzberg's NGE system and to Simpson's FMMC system.British Petroleum (89-A-1204); Defense Advanced Research Projects Agency (90-0083); National Science Foundation (IRI 90-00530); Office of Naval Research (N00014-91-J-4100); Air Force Office of Scientific Research (90-0175
Recommended from our members
A study of distance-based machine learning algorithms
Distance-based algorithms are machine learning algorithms that classify queries
by computing distances between these queries and a number of internally stored
exemplars. Exemplars that are closest to the query have the largest in
uence on
the classi cation assigned to the query. Two speci c distance-based algorithms, the
nearest neighbor algorithm and the nearest-hyperrectangle algorithm, are studied in
detail.
It is shown that the k-nearest neighbor algorithm (kNN) outperforms the rst-
nearest neighbor algorithm only under certain conditions. Data sets must contain
moderate amounts of noise. Training examples from the di erent classes must belong
to clusters that allow an increase in the value of k without reaching into clusters of
other classes. Methods for choosing the value of k for kNN are investigated. It is
shown that one-fold cross-validation on a restricted number of values for k su ces
for best performance. It is also shown that for best performance the votes of the
k-nearest neighbors of a query should be weighted in inverse proportion to their
distances from the query.
Principal component analysis is shown to reduce the number of relevant dimen-
sions substantially in several domains. Two methods for learning feature weights
for a weighted Euclidean distance metric are proposed. These methods improve the
performance of kNN and NN in a variety of domains.
The nearest-hyperrectangle algorithm (NGE) is found to give predictions that are
substantially inferior to those given by kNN in a variety of domains. Experiments performed to understand this inferior performance led to the discovery of several
improvements to NGE. Foremost of these is BNGE, a batch algorithm that avoids
construction of overlapping hyperrectangles from di erent classes. Although it is
generally superior to NGE, BNGE is still signi cantly inferior to kNN in a variety
of domains. Hence, a hybrid algorithm (KBNGE), that uses BNGE in parts of the
input space that can be represented by a single hyperrectangle and kNN otherwise,
is introduced.
The primary contributions of this dissertation are (a) several improvements to
existing distance-based algorithms, (b) several new distance-based algorithms, and
(c) an experimentally supported understanding of the conditions under which various
distance-based algorithms are likely to give good performance
Designing similarity functions
The concept of similarity is important in many areas of cognitive science, computer science, and statistics. In machine learning, functions that measure similarity between two instances form the core of instance-based classifiers. Past similarity measures have been primarily based on simple Euclidean distance. As machine learning has matured, it has become obvious that a simple numeric instance representation is insufficient for most domains. Similarity functions for symbolic attributes have been developed, and simple methods for combining these functions with numeric similarity functions were devised. This sequence of events has revealed three important issues, which this thesis addresses.
The first issue is concerned with combining multiple measures of similarity. There is no equivalence between units of numeric similarity and units of symbolic similarity. Existing similarity functions for numeric and symbolic attributes have no common foundation, and so various schemes have been devised to avoid biasing the overall similarity towards one type of attribute. The similarity function design framework proposed by this thesis produces probability distributions that describe the likelihood of transforming between two attribute values. Because common units of probability are employed, similarities may be combined using standard methods. It is empirically shown that the resulting similarity functions treat different attribute types coherently.
The second issue relates to the instance representation itself. The current choice of numeric and symbolic attribute types is insufficient for many domains, in which more complicated representations are required. For example, a domain may require varying numbers of features, or features with structural information. The framework proposed by this thesis is sufficiently general to permit virtually any type of instance representation-all that is required is that a set of basic transformations that operate on the instances be defined. To illustrate the framework’s applicability to different instance representations, several example similarity functions are developed.
The third, and perhaps most important, issue concerns the ability to incorporate domain knowledge within similarity functions. Domain information plays an important part in choosing an instance representation. However, even given an adequate instance representation, domain information is often lost. For example, numeric features that are modulo (such as the time of day) can be perfectly represented as a numeric attribute, but simple linear similarity functions ignore the modulo nature of the attribute. Similarly, symbolic attributes may have inter-symbol relationships that should be captured in the similarity function. The design framework proposed by this thesis allows domain information to be captured in the similarity function, both in the transformation model and in the probability assigned to basic transformations. Empirical results indicate that such domain information improves classifier performance, particularly when training data is limited
Stylistic atructures: a computational approach to text classification
The problem of authorship attribution has received attention both in the academic world (e.g. did Shakespeare or Marlowe write Edward III?) and outside (e.g. is this confession really the words of the accused or was it made up by someone else?). Previous studies by statisticians and literary scholars have sought "verbal habits" that characterize particular authors consistently. By and large, this has meant looking for distinctive rates of usage of specific marker words -- as in the classic study by Mosteller and Wallace of the Federalist Papers.
The present study is based on the premiss that authorship attribution is just one type of text classification and that advances in this area can be made by applying and adapting techniques from the field of machine learning.
Five different trainable text-classification systems are described, which differ from current stylometric practice in a number of ways, in particular by using a wider variety of marker patterns than customary and by seeking such markers automatically, without being told what to look for. A comparison of the strengths and weaknesses of these systems, when tested on a representative range of text-classification problems, confirms the importance of paying more attention than usual to alternative methods of representing distinctive differences between types of text.
The thesis concludes with suggestions on how to make further progress towards the goal of a fully automatic, trainable text-classification system
Stylistic atructures: a computational approach to text classification
The problem of authorship attribution has received attention both in the academic world (e.g. did Shakespeare or Marlowe write Edward III?) and outside (e.g. is this confession really the words of the accused or was it made up by someone else?). Previous studies by statisticians and literary scholars have sought "verbal habits" that characterize particular authors consistently. By and large, this has meant looking for distinctive rates of usage of specific marker words -- as in the classic study by Mosteller and Wallace of the Federalist Papers.
The present study is based on the premiss that authorship attribution is just one type of text classification and that advances in this area can be made by applying and adapting techniques from the field of machine learning.
Five different trainable text-classification systems are described, which differ from current stylometric practice in a number of ways, in particular by using a wider variety of marker patterns than customary and by seeking such markers automatically, without being told what to look for. A comparison of the strengths and weaknesses of these systems, when tested on a representative range of text-classification problems, confirms the importance of paying more attention than usual to alternative methods of representing distinctive differences between types of text.
The thesis concludes with suggestions on how to make further progress towards the goal of a fully automatic, trainable text-classification system
Newspaper Clippings: March 29, 1950 to April 11, 1951
A collection of newspaper clippings with dates ranging from March 29, 1950, to April 11, 1951. The clippings are arranged in chronological order. During this time Michael DeCiantis was serving as the West Warwick Town Solicitor. Among other things, Scrapbook 5 contains clippings featuring DeCiantis\u27 exploits as a town solicitor, town developments, politics, and a murder trial and investigation. A few pages of the scrapbook have been dedicated to Ann DeCiantis, Michael DeCiantis\u27 daughter, who was selected by television agencies for a role in a new series. Some pages have articles that obscure other items on the same page. In this case, the same page has been shot multiple times to show all content.https://digitalcommons.ric.edu/mss-0001_series_06/1005/thumbnail.jp
Bowdoin Alumnus Volume 33 (1958-1959)
https://digitalcommons.bowdoin.edu/alumni-magazines/1031/thumbnail.jp
The Whitworthian 1974-1975
The Whitworthian student newspaper. September 1974-May 1975.https://digitalcommons.whitworth.edu/whitworthian/1058/thumbnail.jp
Newspaper Clippings: July 1, 1941 to March 29, 1950
A collection of newspaper clippings with dates ranging from July 1, 1941, to March 29, 1950. The clippings are arranged in chronological order. During this time Michael DeCiantis resigned as a member of the State Unemployment Compensation Board to deeper pursue a growing interest in politics. Much of Scrapbook 4 contains clippings following DeCiantis\u27 exploits in politics, starting with his support of Superior Judge Robert E. Quinn for a Democratic Senate Nomination and later his successes as West Warwick Town Solicitor and President of the Kent County Bar Association. Some pages have articles that obscure other items on the same page. In this case, the same page has been shot multiple times to show all content.https://digitalcommons.ric.edu/mss-0001_series_06/1002/thumbnail.jp
The Iowa Official Register, 1981-1982
The Iowa Official Register, commonly known as the "Redbook," serves as a biographical and historical record of Iowa's leaders, government and people