61,750 research outputs found

    A divide and conquer method for symbolic regression

    Full text link
    Symbolic regression aims to find a function that best explains the relationship between independent variables and the objective value based on a given set of sample data. Genetic programming (GP) is usually considered as an appropriate method for the problem since it can optimize functional structure and coefficients simultaneously. However, the convergence speed of GP might be too slow for large scale problems that involve a large number of variables. Fortunately, in many applications, the target function is separable or partially separable. This feature motivated us to develop a new method, divide and conquer (D&C), for symbolic regression, in which the target function is divided into a number of sub-functions and the sub-functions are then determined by any of a GP algorithm. The separability is probed by a new proposed technique, Bi-Correlation test (BiCT). D&C powered GP has been tested on some real-world applications, and the study shows that D&C can help GP to get the target function much more rapidly

    Spectral Cauchy Characteristic Extraction of strain, news and gravitational radiation flux

    Get PDF
    We present a new approach for the Cauchy-characteristic extraction of gravitational radiation strain, news function, and the flux of the energy-momentum, supermomentum and angular momentum associated with the Bondi-Metzner-Sachs asymptotic symmetries. In Cauchy-characteristic extraction, a characteristic evolution code takes numerical data on an inner worldtube supplied by a Cauchy evolution code, and propagates it outwards to obtain the space-time metric in a neighborhood of null infinity. The metric is first determined in a scrambled form in terms of coordinates determined by the Cauchy formalism. In prior treatments, the waveform is first extracted from this metric and then transformed into an asymptotic inertial coordinate system. This procedure provides the physically proper description of the waveform and the radiated energy but it does not generalize to determine the flux of angular momentum or supermomentum. Here we formulate and implement a new approach which transforms the full metric into an asymptotic inertial frame and provides a uniform treatment of all the radiation fluxes associated with the asymptotic symmetries. Computations are performed and calibrated using the Spectral Einstein Code (SpEC).Comment: 30 pages, 17 figure

    Binding and Normalization of Binary Sparse Distributed Representations by Context-Dependent Thinning

    Get PDF
    Distributed representations were often criticized as inappropriate for encoding of data with a complex structure. However Plate's Holographic Reduced Representations and Kanerva's Binary Spatter Codes are recent schemes that allow on-the-fly encoding of nested compositional structures by real-valued or dense binary vectors of fixed dimensionality. In this paper we consider procedures of the Context-Dependent Thinning which were developed for representation of complex hierarchical items in the architecture of Associative-Projective Neural Networks. These procedures provide binding of items represented by sparse binary codevectors (with low probability of 1s). Such an encoding is biologically plausible and allows a high storage capacity of distributed associative memory where the codevectors may be stored. In contrast to known binding procedures, Context-Dependent Thinning preserves the same low density (or sparseness) of the bound codevector for varied number of component codevectors. Besides, a bound codevector is not only similar to another one with similar component codevectors (as in other schemes), but it is also similar to the component codevectors themselves. This allows the similarity of structures to be estimated just by the overlap of their codevectors, without retrieval of the component codevectors. This also allows an easy retrieval of the component codevectors. Examples of algorithmic and neural-network implementations of the thinning procedures are considered. We also present representation examples for various types of nested structured data (propositions using role-filler and predicate-arguments representation schemes, trees, directed acyclic graphs) using sparse codevectors of fixed dimension. Such representations may provide a fruitful alternative to the symbolic representations of traditional AI, as well as to the localist and microfeature-based connectionist representations

    Symmetric Regularization, Reduction and Blow-Up of the Planar Three-Body Problem

    Full text link
    We carry out a sequence of coordinate changes for the planar three-body problem which successively eliminate the translation and rotation symmetries, regularize all three double collision singularities and blow-up the triple collision. Parametrizing the configurations by the three relative position vectors maintains the symmetry among the masses and simplifies the regularization of binary collisions. Using size and shape coordinates facilitates the reduction by rotations and the blow-up of triple collision while emphasizing the role of the shape sphere. By using homogeneous coordinates to describe Hamiltonian systems whose configurations spaces are spheres or projective spaces, we are able to take a modern, global approach to these familiar problems. We also show how to obtain the reduced and regularized differential equations in several convenient local coordinates systems.Comment: 51 pages, 4 figure

    Post-Newtonian Theory for Precision Doppler Measurements of Binary Star Orbits

    Get PDF
    The determination of velocities of stars from precise Doppler measurements is described here using relativistic theory of astronomical reference frames so as to determine the Keplerian and post-Keplerian parameters of binary systems. We apply successive Lorentz transformations and the relativistic equation of light propagation to establish the exact treatment of Doppler effect in binary systems both in special and general relativity theories. As a result, the Doppler shift is a sum of (1) linear in c−1c^{-1} terms, which include the ordinary Doppler effect and its variation due to the secular radial acceleration of the binary with respect to observer; (2) terms proportional to c−2c^{-2}, which include the contributions from the quadratic Doppler effect caused by the relative motion of binary star with respect to the Solar system, motion of the particle emitting light and diurnal rotational motion of observer, orbital motion of the star around the binary's barycenter, and orbital motion of the Earth; and (3) terms proportional to c−2c^{-2}, which include the contributions from redshifts due to gravitational fields of the star, star's companion, Galaxy, Solar system, and the Earth. After parameterization of the binary's orbit we find that the presence of periodically changing terms in the Doppler schift enables us disentangling different terms and measuring, along with the well known Keplerian parameters of the binary, four additional post-Keplerian parameters, including the inclination angle of the binary's orbit, ii. We briefly discuss feasibility of practical implementation of these theoretical results, which crucially depends on further progress in the technique of precision Doppler measurements.Comment: Minor changes, 1 Figure included, submitted to Astrophys.

    ART 2-A: An Adaptive Resonance Algorithm for Rapid Category Learning and Recognition

    Full text link
    This article introduces ART 2-A, an efficient algorithm that emulates the self-organizing pattern recognition and hypothesis testing properties of the ART 2 neural network architecture, but at a speed two to three orders of magnitude faster. Analysis and simulations show how the ART 2-A systems correspond to ART 2 dynamics at both the fast-learn limit and at intermediate learning rates. Intermediate learning rates permit fast commitment of category nodes but slow recoding, analogous to properties of word frequency effects, encoding specificity effects, and episodic memory. Better noise tolerance is hereby achieved without a loss of learning stability. The ART 2 and ART 2-A systems are contrasted with the leader algorithm. The speed of ART 2-A makes practical the use of ART 2 modules in large-scale neural computation.BP (89-A-1204); Defense Advanced Research Projects Agency (90-0083); National Science Foundation (IRI-90-00530); Air Force Office of Scientific Research (90-0175, 90-0128); Army Research Office (DAAL-03-88-K0088

    Herding as a Learning System with Edge-of-Chaos Dynamics

    Full text link
    Herding defines a deterministic dynamical system at the edge of chaos. It generates a sequence of model states and parameters by alternating parameter perturbations with state maximizations, where the sequence of states can be interpreted as "samples" from an associated MRF model. Herding differs from maximum likelihood estimation in that the sequence of parameters does not converge to a fixed point and differs from an MCMC posterior sampling approach in that the sequence of states is generated deterministically. Herding may be interpreted as a"perturb and map" method where the parameter perturbations are generated using a deterministic nonlinear dynamical system rather than randomly from a Gumbel distribution. This chapter studies the distinct statistical characteristics of the herding algorithm and shows that the fast convergence rate of the controlled moments may be attributed to edge of chaos dynamics. The herding algorithm can also be generalized to models with latent variables and to a discriminative learning setting. The perceptron cycling theorem ensures that the fast moment matching property is preserved in the more general framework
    • …
    corecore