18,166 research outputs found
Distributed computing methodology for training neural networks in an image-guided diagnostic application
Distributed computing is a process through which a set of computers connected by a network is used collectively to solve a single problem. In this paper, we propose a distributed computing methodology for training neural networks for the detection of lesions in colonoscopy. Our approach is based on partitioning the training set across multiple processors using a parallel virtual machine. In this way, interconnected computers of varied architectures can be used for the distributed evaluation of the error function and gradient values, and, thus, training neural networks utilizing various learning methods. The proposed methodology has large granularity and low synchronization, and has been implemented and tested. Our results indicate that the parallel virtual machine implementation of the training algorithms developed leads to considerable speedup, especially when large network architectures and training sets are used
Practical recommendations for gradient-based training of deep architectures
Learning algorithms related to artificial neural networks and in particular
for Deep Learning may seem to involve many bells and whistles, called
hyper-parameters. This chapter is meant as a practical guide with
recommendations for some of the most commonly used hyper-parameters, in
particular in the context of learning algorithms based on back-propagated
gradient and gradient-based optimization. It also discusses how to deal with
the fact that more interesting results can be obtained when allowing one to
adjust many hyper-parameters. Overall, it describes elements of the practice
used to successfully and efficiently train and debug large-scale and often deep
multi-layer neural networks. It closes with open questions about the training
difficulties observed with deeper architectures
Artificial Neural Network-based error compensation procedure for low-cost encoders
An Artificial Neural Network-based error compensation method is proposed for
improving the accuracy of resolver-based 16-bit encoders by compensating for
their respective systematic error profiles. The error compensation procedure,
for a particular encoder, involves obtaining its error profile by calibrating
it on a precision rotary table, training the neural network by using a part of
this data and then determining the corrected encoder angle by subtracting the
ANN-predicted error from the measured value of the encoder angle. Since it is
not guaranteed that all the resolvers will have exactly similar error profiles
because of the inherent differences in their construction on a micro scale, the
ANN has been trained on one error profile at a time and the corresponding
weight file is then used only for compensating the systematic error of this
particular encoder. The systematic nature of the error profile for each of the
encoders has also been validated by repeated calibration of the encoders over a
period of time and it was found that the error profiles of a particular encoder
recorded at different epochs show near reproducible behavior. The ANN-based
error compensation procedure has been implemented for 4 encoders by training
the ANN with their respective error profiles and the results indicate that the
accuracy of encoders can be improved by nearly an order of magnitude from
quoted values of ~6 arc-min to ~0.65 arc-min when their corresponding
ANN-generated weight files are used for determining the corrected encoder
angle.Comment: 16 pages, 4 figures. Accepted for Publication in Measurement Science
and Technology (MST
- …