CORE
🇺🇦
make metadata, not war
Services
Services overview
Explore all CORE services
Access to raw data
API
Dataset
FastSync
Content discovery
Recommender
Discovery
OAI identifiers
OAI Resolver
Managing content
Dashboard
Bespoke contracts
Consultancy services
Support us
Support us
Membership
Sponsorship
Community governance
Advisory Board
Board of supporters
Research network
About
About us
Our mission
Team
Blog
FAQs
Contact us
Cross-Validation In Fuzzy Artmap For Large Databases
Authors
George Anagnostopoulos
Michael Georgiopoulos
Anna Koufakou
Publication date
10 October 2001
Publisher
'Information Bulletin on Variable Stars (IBVS)'
Abstract
In this paper we are examining the issue of overtraining in Fuzzy ARTMAP. Over-training in Fuzzy ARTMAP manifests itself in two different ways: (a) it degrades the generalization performance of Fuzzy ARTMAP as training progresses; and (b) it creates unnecessarily large Fuzzy ARTMAP neural network architectures. In this work, we are demonstrating that overtraining happens in Fuzzy ARTMAP and we propose an old remedy for its cure: cross-validation. In our experiments, we compare the performance of Fuzzy ARTMAP that is trained (i) until the completion of training, (ii) for one epoch, and (iii) until its performance on a validation set is maximized. The experiments were performed on artificial and real databases. The conclusion derived from those experiments is that cross-validation is a useful procedure in Fuzzy ARTMAP, because it produces smaller Fuzzy ARTMAP architectures with improved generalization performance. The trade-off is that cross-validation introduces additional computational complexity in the training phase of Fuzzy ARTMAP. Copyright © 2001 Elsevier Science Ltd
Similar works
Full text
Available Versions
University of Central Florida (UCF): STARS (Showcase of Text, Archives, Research & Scholarship)
See this paper in CORE
Go to the repository landing page
Download from data provider
oai:stars.library.ucf.edu:scop...
Last time updated on 19/07/2022