1,435 research outputs found

    Tropospheric range error parameters: Further studies

    Get PDF
    Improved parameters are presented for predicting the tropospheric effect on electromagnetic range measurements from surface meteorological data. More geographic locations have been added to the earlier list. Parameters are given for computing the dry component of the zenith radio range effect from surface pressure alone with an rms error of 1 to 2 mm, or the total range effect from the dry and wet components of the surface refractivity and a two-part quartic profile model. The new parameters are obtained, as before, from meteorological balloon data but with improved procedures, including the conversion of the geopotential heights of the balloon data to actual or geometric heights before using the data. The revised values of the parameter k (dry component of vertical radio range effect per unit pressure at the surface) show more latitude variation than is accounted for by the variation of g, the acceleration of gravity

    Intelligent search for distributed information sources using heterogeneous neural networks

    Get PDF
    As the number and diversity of distributed information sources on the Internet exponentially increase, various search services are developed to help the users to locate relevant information. But they still exist some drawbacks such as the difficulty of mathematically modeling retrieval process, the lack of adaptivity and the indiscrimination of search. This paper shows how heteroge-neous neural networks can be used in the design of an intelligent distributed in-formation retrieval (DIR) system. In particular, three typical neural network models - Kohoren's SOFM Network, Hopfield Network, and Feed Forward Network with Back Propagation algorithm are introduced to overcome the above drawbacks in current research of DIR by using their unique properties. This preliminary investigation suggests that Neural Networks are useful tools for intelligent search for distributed information sources

    Storage Capacity of Two-dimensional Neural Networks

    Get PDF
    We investigate the maximum number of embedded patterns in the two-dimensional Hopfield model. The grand state energies of two specific network states, namely, the energies of the pure-ferromagnetic state and the state of specific one stored pattern are calculated exactly in terms of the correlation function of the ferromagnetic Ising model. We also investigate the energy landscape around them by computer simulations. Taking into account the qualitative features of the phase diagrams obtained by Nishimori, Whyte and Sherrington [Phys. Rev. E {\bf 51}, 3628 (1995)], we conclude that the network cannot retrieve more than three patterns.Comment: 13pages, 7figures, revtex

    On the equivalence of the Ashkin-Teller and the four-state Potts-glass models of neural networks

    Full text link
    We show that for a particular choice of the coupling parameters the Ashkin-Teller spin-glass neural network model with the Hebb learning rule and one condensed pattern yields the same thermodynamic properties as the four-state anisotropic Potts-glass neural network model. This equivalence is not seen at the level of the Hamiltonians.Comment: 3 pages, revtex, additional arguments presente

    Rhythmic inhibition allows neural networks to search for maximally consistent states

    Full text link
    Gamma-band rhythmic inhibition is a ubiquitous phenomenon in neural circuits yet its computational role still remains elusive. We show that a model of Gamma-band rhythmic inhibition allows networks of coupled cortical circuit motifs to search for network configurations that best reconcile external inputs with an internal consistency model encoded in the network connectivity. We show that Hebbian plasticity allows the networks to learn the consistency model by example. The search dynamics driven by rhythmic inhibition enable the described networks to solve difficult constraint satisfaction problems without making assumptions about the form of stochastic fluctuations in the network. We show that the search dynamics are well approximated by a stochastic sampling process. We use the described networks to reproduce perceptual multi-stability phenomena with switching times that are a good match to experimental data and show that they provide a general neural framework which can be used to model other 'perceptual inference' phenomena

    A statistical mechanics of an oscillator associative memory with scattered natural frequencies

    Full text link
    Analytic treatment of a non-equilibrium random system with large degrees of freedoms is one of most important problems of physics. However, little research has been done on this problem as far as we know. In this paper, we propose a new mean field theory that can treat a general class of a non-equilibrium random system. We apply the present theory to an analysis for an associative memory with oscillatory elements, which is a well-known typical random system with large degrees of freedoms.Comment: 8 pages, 4 figure

    Naive mean field approximation for image restoration

    Full text link
    We attempt image restoration in the framework of the Baysian inference. Recently, it has been shown that under a certain criterion the MAP (Maximum A Posterior) estimate, which corresponds to the minimization of energy, can be outperformed by the MPM (Maximizer of the Posterior Marginals) estimate, which is equivalent to a finite-temperature decoding method. Since a lot of computational time is needed for the MPM estimate to calculate the thermal averages, the mean field method, which is a deterministic algorithm, is often utilized to avoid this difficulty. We present a statistical-mechanical analysis of naive mean field approximation in the framework of image restoration. We compare our theoretical results with those of computer simulation, and investigate the potential of naive mean field approximation.Comment: 9 pages, 11 figure

    Phase Transitions of an Oscillator Neural Network with a Standard Hebb Learning Rule

    Full text link
    Studies have been made on the phase transition phenomena of an oscillator network model based on a standard Hebb learning rule like the Hopfield model. The relative phase informations---the in-phase and anti-phase, can be embedded in the network. By self-consistent signal-to-noise analysis (SCSNA), it was found that the storage capacity is given by αc=0.042\alpha_c = 0.042, which is better than that of Cook's model. However, the retrieval quality is worse. In addition, an investigation was made into an acceleration effect caused by asymmetry of the phase dynamics. Finally, it was numerically shown that the storage capacity can be improved by modifying the shape of the coupling function.Comment: 10 pages, 6 figure

    Quantum computers can search rapidly by using almost any transformation

    Get PDF
    A quantum computer has a clear advantage over a classical computer for exhaustive search. The quantum mechanical algorithm for exhaustive search was originally derived by using subtle properties of a particular quantum mechanical operation called the Walsh-Hadamard (W-H) transform. This paper shows that this algorithm can be implemented by replacing the W-H transform by almost any quantum mechanical operation. This leads to several new applications where it improves the number of steps by a square-root. It also broadens the scope for implementation since it demonstrates quantum mechanical algorithms that can readily adapt to available technology.Comment: This paper is an adapted version of quant-ph/9711043. It has been modified to make it more readable for physicists. 9 pages, postscrip

    Maladaptation and the paradox of robustness in evolution

    Get PDF
    Background. Organisms use a variety of mechanisms to protect themselves against perturbations. For example, repair mechanisms fix damage, feedback loops keep homeostatic systems at their setpoints, and biochemical filters distinguish signal from noise. Such buffering mechanisms are often discussed in terms of robustness, which may be measured by reduced sensitivity of performance to perturbations. Methodology/Principal Findings. I use a mathematical model to analyze the evolutionary dynamics of robustness in order to understand aspects of organismal design by natural selection. I focus on two characters: one character performs an adaptive task; the other character buffers the performance of the first character against perturbations. Increased perturbations favor enhanced buffering and robustness, which in turn decreases sensitivity and reduces the intensity of natural selection on the adaptive character. Reduced selective pressure on the adaptive character often leads to a less costly, lower performance trait. Conclusions/Significance. The paradox of robustness arises from evolutionary dynamics: enhanced robustness causes an evolutionary reduction in the adaptive performance of the target character, leading to a degree of maladaptation compared to what could be achieved by natural selection in the absence of robustness mechanisms. Over evolutionary time, buffering traits may become layered on top of each other, while the underlying adaptive traits become replaced by cheaper, lower performance components. The paradox of robustness has widespread implications for understanding organismal design
    corecore