70 research outputs found

    Learning and generalization in feed-forward neural networks

    Get PDF

    PhysicsGP: A Genetic Programming Approach to Event Selection

    Full text link
    We present a novel multivariate classification technique based on Genetic Programming. The technique is distinct from Genetic Algorithms and offers several advantages compared to Neural Networks and Support Vector Machines. The technique optimizes a set of human-readable classifiers with respect to some user-defined performance measure. We calculate the Vapnik-Chervonenkis dimension of this class of learning machines and consider a practical example: the search for the Standard Model Higgs Boson at the LHC. The resulting classifier is very fast to evaluate, human-readable, and easily portable. The software may be downloaded at: http://cern.ch/~cranmer/PhysicsGP.htmlComment: 16 pages 9 figures, 1 table. Submitted to Comput. Phys. Commu

    Parallelizing Feed-Forward Artificial Neural Networks on Transputers

    Get PDF
    This thesis is about parallelizing the training phase of a feed-forward, artificial neural network. More specifically, we develop and analyze a number of parallelizations of the widely used neural net learning algorithm called back-propagation. We describe two different strategies for parallelizing the back-propagation algorithm. A number of parallelizations employing these strategies have been implemented on a system of 48 transputers, permitting us to evaluate and analyze their performances based on the results of actual runs

    Mechatronics of systems with undetermined configurations

    Get PDF
    This work is submitted for the award of a PhD by published works. It deals with some of the efforts of the author over the last ten years in the field of Mechatronics. Mechatronics is a new area invented by the Japanese in the late 1970's, it consists of a synthesis of computers and electronics to improve mechanical systems. To control any mechanical event three fundamental features must be brought together: the sensors used to observe the process, the control software, including the control algorithm used and thirdly the actuator that provides the stimulus to achieve the end result. Simulation, which plays such an important part in the Mechatronics process, is used in both in continuous and discrete forms. The author has spent some considerable time developing skills in all these areas. The author was certainly the first at Middlesex to appreciate the new developments in Mechatronics and their significance for manufacturing. The author was one of the first mechanical engineers to recognise the significance of the new transputer chip. This was applied to the LQG optimal control of a cinefilm copying process. A 300% improvement in operating speed was achieved, together with tension control. To make more efficient use of robots they have to be made both faster and cheaper. The author found extremely low natural frequencies of vibration, ranging from 3 to 25 Hz. This limits the speed of response of existing robots. The vibration data was some of the earliest available in this field, certainly in the UK. Several schemes have been devised to control the flexible robot and maintain the required precision. Actuator technology is one area where mechatronic systems have been the subject of intense development. At Middlesex we have improved on the Aexator pneumatic muscle actuator, enabling it to be used with a precision of about 2 mm. New control challenges have been undertaken now in the field of machine tool chatter and the prevention of slip. A variety of novel and traditional control algorithms have been investigated in order to find out the best approach to solve this problem

    From Conventional to Cl-Based Spatial Analysis

    Get PDF
    Series: Discussion Papers of the Institute for Economic Geography and GIScienc

    Autonomy in the real real-world: A behaviour based view of autonomous systems control in an industrial product inspection system

    Get PDF
    The thesis presented in this dissertation appears in two sequential parts that arose from an exploration of the use of Behaviour Based Artificial Intelligence (BBAI) techniques in a domain outside that of robotics, where BBAI is most frequently used. The work details a real-world physical implementation of the control and interactions of an industrial product inspection system from a BBAI perspective. It concentrates particularly on the control of a number of active laser scanning sensor systems (each a subsystem of a larger main inspection system), using a subsumption architecture. This industrial implementation is in itself a new direction for BBAI control and an important aspect of this thesis. However, the work has also led on to the development of a number of key ideas which contribute to the field of BBAI in general. The second part of the thesis concerns the nature of physical and temporal constraints on a distributed control system and the desirability of utilising mechanisms to provide continuous, low-level learning and adaptation of domain knowledge on a sub-behavioural basis. Techniques used include artificial neural networks and hill-climbing state-space search algorithms. Discussion is supported with examples from experiments with the laser scanning inspection system. Encouraging results suggest that concerted design effort at this low level of activity will benefit the whole system in terms of behavioural robustness and reliability. Relevant aspects of the design process that should be of value in similar real-world projects are identified and emphasised. These issues are particularly important in providing a firm foundation for artificial intelligence based control systems

    Efficient Mapping of Neural Network Models on a Class of Parallel Architectures.

    Get PDF
    This dissertation develops a formal and systematic methodology for efficient mapping of several contemporary artificial neural network (ANN) models on k-ary n-cube parallel architectures (KNC\u27s). We apply the general mapping to several important ANN models including feedforward ANN\u27s trained with backpropagation algorithm, radial basis function networks, cascade correlation learning, and adaptive resonance theory networks. Our approach utilizes a parallel task graph representing concurrent operations of the ANN model during training. The mapping of the ANN is performed in two steps. First, the parallel task graph of the ANN is mapped to a virtual KNC of compatible dimensionality. This involves decomposing each operation into its atomic tasks. Second, the dimensionality of the virtual KNC architecture is recursively reduced through a sequence of transformations until a desired metric is optimized. We refer to this process as folding the virtual architecture. The optimization criteria we consider in this dissertation are defined in terms of the iteration time of the algorithm on the folded architecture. If necessary, the mapping scheme may utilize a subset of the processors of a given KNC architecture if it results in the most efficient simulation. A unique feature of our mapping is that it systematically selects an appropriate degree of parallelism leading to a highly efficient realization of the ANN model on KNC architectures. A novel feature of our work is its ability to efficiently map unit-allocating ANN\u27s. These networks possess a dynamic structure which grows during training. We present a highly efficient scheme for simulating such networks on existing KNC parallel architectures. We assume an upper bound on size of the neural network We perform the folding such that the iteration time of the largest network is minimized. We show that our mapping leads to near-optimal simulation of smaller instances of the neural network. In addition, based on our mapping no data migration or task rescheduling is needed as the size of network grows

    Applications of neural networks to control systems

    Get PDF
    Tese de dout., Engenharia Electrónica, School of Electronic Engineering Science, Univ. of Wales, Bangor, 1992This work investigates the applicability of artificial neural networks to control systems. The following properties of neural networks are identified as of major interest to this field: their ability to implement nonlinear mappings, their massively parallel structure and their capacity to adapt. Exploiting the first feature, a new method is proposed for PID autotuning. Based on integral measures of the open or closed loop step response, multilayer perceptrons (MLPs) are used to supply PID parameter values to a standard PID controller. Before being used on-line, the MLPs are trained offline, to provide PID parameter values based on integral performance criteria. Off-line simulations, where a plant with time-varying parameters and time varying transfer function is considered, show that well damped responses are obtained. The neural PID autotuner is subsequently implemented in real-time. Extensive experimentation confirms the good results obtained in the off-line simulations. To reduce the training time incurred when using the error back-propagation algorithm, three possibilities are investigated. A comparative study of higherorder methods of optimization identifies the Levenberg-Marquardt (LM)algorithm as the best method. When used for function approximation purposes, the neurons in the output layer of the MLPs have a linear activation function. Exploiting this linearity, the standard training criterion can be replaced by a new, yet equivalent, criterion. Using the LM algorithm to minimize this new criterion, together with an alternative form of Jacobian matrix, a new learning algorithm is obtained. This algorithm is subsequently parallelized. Its main blocks of computation are identified, separately parallelized, and finally connected together. The training time of MLPs is reduced by a factor greater than 70 executing the new learning algorithm on 7 Inmos transputers

    An instruction systolic array architecture for multiple neural network types

    Get PDF
    Modern electronic systems, especially sensor and imaging systems, are beginning to incorporate their own neural network subsystems. In order for these neural systems to learn in real-time they must be implemented using VLSI technology, with as much of the learning processes incorporated on-chip as is possible. The majority of current VLSI implementations literally implement a series of neural processing cells, which can be connected together in an arbitrary fashion. Many do not perform the entire neural learning process on-chip, instead relying on other external systems to carry out part of the computation requirements of the algorithm. The work presented here utilises two dimensional instruction systolic arrays in an attempt to define a general neural architecture which is closer to the biological basis of neural networks - it is the synapses themselves, rather than the neurons, that have dedicated processing units. A unified architecture is described which can be programmed at the microcode level in order to facilitate the processing of multiple neural network types. An essential part of neural network processing is the neuron activation function, which can range from a sequential algorithm to a discrete mathematical expression. The architecture presented can easily carry out the sequential functions, and introduces a fast method of mathematical approximation for the more complex functions. This can be evaluated on-chip, thus implementing the entire neural process within a single system. VHDL circuit descriptions for the chip have been generated, and the systolic processing algorithms and associated microcode instruction set for three different neural paradigms have been designed. A software simulator of the architecture has been written, giving results for several common applications in the field
    • …
    corecore