80 research outputs found
Contrastive learning and neural oscillations
The concept of Contrastive Learning (CL) is developed as a family of possible learning algorithms for neural networks. CL is an extension of Deterministic Boltzmann Machines to more general dynamical systems. During learning, the network oscillates between two phases. One phase has a teacher signal and one phase has no teacher signal. The weights are updated using a learning rule that corresponds to gradient descent on a contrast function that measures the discrepancy between the free network and the network with a teacher signal. The CL approach provides a general unified framework for developing new learning algorithms. It also shows that many different types of clamping and teacher signals are possible. Several examples are given and an analysis of the landscape of the contrast function is proposed with some relevant predictions for the CL curves. An approach that may be suitable for collective analog implementations is described. Simulation results and possible extensions are briefly discussed together with a new conjecture regarding the function of certain oscillations in the brain. In the appendix, we also examine two extensions of contrastive learning to time-dependent trajectories
A survey of visual preprocessing and shape representation techniques
Many recent theories and methods proposed for visual preprocessing and shape representation are summarized. The survey brings together research from the fields of biology, psychology, computer science, electrical engineering, and most recently, neural networks. It was motivated by the need to preprocess images for a sparse distributed memory (SDM), but the techniques presented may also prove useful for applying other associative memories to visual pattern recognition. The material of this survey is divided into three sections: an overview of biological visual processing; methods of preprocessing (extracting parts of shape, texture, motion, and depth); and shape representation and recognition (form invariance, primitives and structural descriptions, and theories of attention)
Learning and Inference in Massive Social Networks
Researchers and practitioners increasingly are gaining access
to data on explicit social networks. For example, telecommunications
and technology firms record data on consumer
networks (via phone calls, emails, voice-over-IP, instant messaging),
and social-network portal sites such as MySpace,
Friendster and Facebook record consumer-generated data
on social networks. Inference for fraud detection [5, 3, 8],
marketing [9], and other tasks can be improved with learned
models that take social networks into account and with collective
inference [12], which allows inferences about nodes
in the network to affect each other. However, these socialnetwork
graphs can be huge, comprising millions to billions
of nodes and one or two orders of magnitude more links.
This paper studies the application of collective inference
to improve prediction over a massive graph. Faced initially
with a social network comprising hundreds of millions of
nodes and a few billion edges, our goal is: to produce an
approximate consumer network that is orders of magnitude
smaller, but still facilitates improved performance via collective
inference. We introduce a sampling technique designed
to reduce the size of the network by many orders of magnitude,
but to keep linkages that facilitate improved prediction
via collective inference.
In short, the sampling scheme operates as follows: (1)
choose a set of nodes of interest; (2) then, in analogy to
snowball sampling [14], grow local graphs around these nodes,
adding their social networks, their neighbors’ social networks,
and so on; (3) next, prune these local graphs of edges
which are expected to contribute little to the collective inference;
(4) finally, connect the local graphs together to form
a graph with (hopefully) useful inference connectivity.
We apply this sampling method to assess whether collective
inference can improve learned targeted-marketing models
for a social network of consumers of telecommunication
services. Prior work [9] has shown improvement to the learning
of targeting models by including social-neighborhood
information—in particular, information on existing customers
in the immediate social network of a potential target. However,
the improvement was restricted to the “network neighbors”,
those targets linked to a prior customer thought to
be good candidates for the new service. Collective inference
techniques may extend the predictive influence of existing
customers beyond their immediate neighborhoods. For the
present work, our motivating conjecture has been that this
influence can improve prediction for consumers who are not
strongly connected to existing customers. Our results show
that this is indeed the case: collective inference on the approximate
network enables significantly improved predictive
performance for non-network-neighbor consumers, and for
consumers who have few links to existing customers.
In the rest of this extended abstract we motivate our approach,
describe our sampling method, present results on
applying our approach to a large real-world target marketing
campaign in the telecommunications industry, and finally
discuss our findings.NYU, Stern School of Business, IOMS Department, Center for Digital Economy Researc
Learning and Inference in Massive Social Networks
Researchers and practitioners increasingly are gaining access
to data on explicit social networks. For example, telecommunications
and technology firms record data on consumer
networks (via phone calls, emails, voice-over-IP, instant messaging),
and social-network portal sites such as MySpace,
Friendster and Facebook record consumer-generated data
on social networks. Inference for fraud detection [5, 3, 8],
marketing [9], and other tasks can be improved with learned
models that take social networks into account and with collective
inference [12], which allows inferences about nodes
in the network to affect each other. However, these socialnetwork
graphs can be huge, comprising millions to billions
of nodes and one or two orders of magnitude more links.
This paper studies the application of collective inference
to improve prediction over a massive graph. Faced initially
with a social network comprising hundreds of millions of
nodes and a few billion edges, our goal is: to produce an
approximate consumer network that is orders of magnitude
smaller, but still facilitates improved performance via collective
inference. We introduce a sampling technique designed
to reduce the size of the network by many orders of magnitude,
but to keep linkages that facilitate improved prediction
via collective inference.
In short, the sampling scheme operates as follows: (1)
choose a set of nodes of interest; (2) then, in analogy to
snowball sampling [14], grow local graphs around these nodes,
adding their social networks, their neighbors’ social networks,
and so on; (3) next, prune these local graphs of edges
which are expected to contribute little to the collective inference;
(4) finally, connect the local graphs together to form
a graph with (hopefully) useful inference connectivity.
We apply this sampling method to assess whether collective
inference can improve learned targeted-marketing models
for a social network of consumers of telecommunication
services. Prior work [9] has shown improvement to the learning
of targeting models by including social-neighborhood
information—in particular, information on existing customers
in the immediate social network of a potential target. However,
the improvement was restricted to the “network neighbors”,
those targets linked to a prior customer thought to
be good candidates for the new service. Collective inference
techniques may extend the predictive influence of existing
customers beyond their immediate neighborhoods. For the
present work, our motivating conjecture has been that this
influence can improve prediction for consumers who are not
strongly connected to existing customers. Our results show
that this is indeed the case: collective inference on the approximate
network enables significantly improved predictive
performance for non-network-neighbor consumers, and for
consumers who have few links to existing customers.
In the rest of this extended abstract we motivate our approach,
describe our sampling method, present results on
applying our approach to a large real-world target marketing
campaign in the telecommunications industry, and finally
discuss our findings.NYU, Stern School of Business, IOMS Department, Center for Digital Economy Researc
Recommended from our members
Novel methods to predict solid-state material properties
Solid-state materials find ubiquitous use in modern technology - from semiconductors in electronics to steel in buildings and superconductors in MRI machines. Theoretical understanding of the atomic-scale behaviour of these materials can be leveraged to design new materials with desirable properties. In this thesis, we investigate the challenges that arise when this is attempted in practice.
Accurate and inexpensive methods to tackle the atomic-scale problem are a prerequisite for materials discovery. We begin with a description of existing methods. This is followed by the development of a Monte Carlo method to calculate expectation values from the many-body picture without the need for a trial wavefunction, which is both a fundamental, and practical, limitation in existing techniques.
Having explored first-principles methods, we turn to their use in understanding materials, beginning with an investigation of the structure of Lithium. Structure searching calculations result in a mixed-phase model at low temperatures, in good agreement with previous experimental and theoretical results. The quasi-harmonic treatment of finite-temperature thermodynamics is extended to include anharmonic nuclear vibrations, which are shown to not alter the phase diagram despite the small mass of the Li atoms.
Focus then shifts towards leveraging these same methods to discover novel superconductors. This begins with an investigation of the LaH and YH compounds, where a new hexagonal phase of LaH provides an explanation for recent experimental measurements. Machine-learning techniques and novel screening methods are then employed to discover hydrides of Rb and Cs that exhibit superconductivity at significantly lower pressures than LaH. Optimizations to, and automation of, the workflow then enables the discovery of superconductors on an unprecedented scale, leading to hundreds of new high-temperature superconductors.
Throughout the thesis, the importance of structures that are saddle-points of the energy landscape becomes apparent. The thesis closes with the development of a new algorithm to locate saddle-points that requires no additional information beyond that used by the cheapest existing methods.
This thesis demonstrates that there is progress to be made at every stage of the first-principles materials discovery process and highlights that improving the workflow itself is a non-trivial, but fruitful, pursuit
Flow pattern analysis for magnetic resonance velocity imaging
Blood flow in the heart is highly complex. Although blood flow patterns have been investigated by both computational modelling and invasive/non-invasive imaging techniques, their evolution and intrinsic connection with cardiovascular disease has yet to be explored. Magnetic resonance (MR) velocity imaging provides a comprehensive distribution of multi-directional in vivo flow distribution so that detailed quantitative analysis of flow patterns is now possible. However, direct visualisation or quantification of vector fields is of little clinical use, especially for inter-subject or serial comparison of changes in flow patterns due to the progression of the disease or in response to therapeutic measures. In order to achieve a comprehensive and integrated description of flow in health and disease, it is necessary to characterise and model both normal and abnormal flows and their effects. To accommodate the diversity of flow patterns in relation to morphological and functional changes, we have described in this thesis an approach of detecting salient topological features prior to analytical assessment of dynamical indices of the flow patterns. To improve the accuracy of quantitative analysis of the evolution of topological flow features, it is essential to restore the original flow fields so that critical points associated with salient flow features can be more reliably detected. We propose a novel framework for the restoration, abstraction, extraction and tracking of flow features such that their dynamic indices can be accurately tracked and quantified. The restoration method is formulated as a constrained optimisation problem to remove the effects of noise and to improve the consistency of the MR velocity data. A computational scheme is derived from the First Order Lagrangian Method for solving the optimisation problem. After restoration, flow abstraction is applied to partition the entire flow field into clusters, each of which is represented by a local linear expansion of its velocity components. This process not only greatly reduces the amount of data required to encode the velocity distribution but also permits an analytical representation of the flow field from which critical points associated with salient flow features can be accurately extracted. After the critical points are extracted, phase portrait theory can be applied to separate them into attracting/repelling focuses, attracting/repelling nodes, planar vortex, or saddle. In this thesis, we have focused on vortical flow features formed in diastole. To track the movement of the vortices within a cardiac cycle, a tracking algorithm based on relaxation labelling is employed. The constraints and parameters used in the tracking algorithm are designed using the characteristics of the vortices. The proposed framework is validated with both simulated and in vivo data acquired from patients with sequential MR examination following myocardial infarction. The main contribution of the thesis is in the new vector field restoration and flow feature abstraction method proposed. They allow the accurate tracking and quantification of dynamic indices associated with salient features so that inter- and intra-subject comparisons can be more easily made. This provides further insight into the evolution of blood flow patterns and permits the establishment of links between blood flow patterns and localised genesis and progression of cardiovascular disease.Open acces
- …