5,322 research outputs found
The phase transition in inhomogeneous random graphs
We introduce a very general model of an inhomogenous random graph with
independence between the edges, which scales so that the number of edges is
linear in the number of vertices. This scaling corresponds to the p=c/n scaling
for G(n,p) used to study the phase transition; also, it seems to be a property
of many large real-world graphs. Our model includes as special cases many
models previously studied.
We show that under one very weak assumption (that the expected number of
edges is `what it should be'), many properties of the model can be determined,
in particular the critical point of the phase transition, and the size of the
giant component above the transition. We do this by relating our random graphs
to branching processes, which are much easier to analyze.
We also consider other properties of the model, showing, for example, that
when there is a giant component, it is `stable': for a typical random graph, no
matter how we add or delete o(n) edges, the size of the giant component does
not change by more than o(n).Comment: 135 pages; revised and expanded slightly. To appear in Random
Structures and Algorithm
Identifying networks with common organizational principles
Many complex systems can be represented as networks, and the problem of
network comparison is becoming increasingly relevant. There are many techniques
for network comparison, from simply comparing network summary statistics to
sophisticated but computationally costly alignment-based approaches. Yet it
remains challenging to accurately cluster networks that are of a different size
and density, but hypothesized to be structurally similar. In this paper, we
address this problem by introducing a new network comparison methodology that
is aimed at identifying common organizational principles in networks. The
methodology is simple, intuitive and applicable in a wide variety of settings
ranging from the functional classification of proteins to tracking the
evolution of a world trade network.Comment: 26 pages, 7 figure
Geometric deep learning: going beyond Euclidean data
Many scientific fields study data with an underlying structure that is a
non-Euclidean space. Some examples include social networks in computational
social sciences, sensor networks in communications, functional networks in
brain imaging, regulatory networks in genetics, and meshed surfaces in computer
graphics. In many applications, such geometric data are large and complex (in
the case of social networks, on the scale of billions), and are natural targets
for machine learning techniques. In particular, we would like to use deep
neural networks, which have recently proven to be powerful tools for a broad
range of problems from computer vision, natural language processing, and audio
analysis. However, these tools have been most successful on data with an
underlying Euclidean or grid-like structure, and in cases where the invariances
of these structures are built into networks used to model them. Geometric deep
learning is an umbrella term for emerging techniques attempting to generalize
(structured) deep neural models to non-Euclidean domains such as graphs and
manifolds. The purpose of this paper is to overview different examples of
geometric deep learning problems and present available solutions, key
difficulties, applications, and future research directions in this nascent
field
- …