14,182 research outputs found
Kernel Graph Convolutional Neural Networks
Graph kernels have been successfully applied to many graph classification
problems. Typically, a kernel is first designed, and then an SVM classifier is
trained based on the features defined implicitly by this kernel. This two-stage
approach decouples data representation from learning, which is suboptimal. On
the other hand, Convolutional Neural Networks (CNNs) have the capability to
learn their own features directly from the raw data during training.
Unfortunately, they cannot handle irregular data such as graphs. We address
this challenge by using graph kernels to embed meaningful local neighborhoods
of the graphs in a continuous vector space. A set of filters is then convolved
with these patches, pooled, and the output is then passed to a feedforward
network. With limited parameter tuning, our approach outperforms strong
baselines on 7 out of 10 benchmark datasets.Comment: Accepted at ICANN '1
Local Search in Unstructured Networks
We review a number of message-passing algorithms that can be used to search
through power-law networks. Most of these algorithms are meant to be
improvements for peer-to-peer file sharing systems, and some may also shed some
light on how unstructured social networks with certain topologies might
function relatively efficiently with local information. Like the networks that
they are designed for, these algorithms are completely decentralized, and they
exploit the power-law link distribution in the node degree. We demonstrate that
some of these search algorithms can work well on real Gnutella networks, scale
sub-linearly with the number of nodes, and may help reduce the network search
traffic that tends to cripple such networks.Comment: v2 includes minor revisions: corrections to Fig. 8's caption and
references. 23 pages, 10 figures, a review of local search strategies in
unstructured networks, a contribution to `Handbook of Graphs and Networks:
From the Genome to the Internet', eds. S. Bornholdt and H.G. Schuster
(Wiley-VCH, Berlin, 2002), to be publishe
On the efficiency of estimating penetrating rank on large graphs
P-Rank (Penetrating Rank) has been suggested as a useful measure of structural similarity that takes account of both incoming and outgoing edges in ubiquitous networks. Existing work often utilizes memoization to compute P-Rank similarity in an iterative fashion, which requires cubic time in the worst case. Besides, previous methods mainly focus on the deterministic computation of P-Rank, but lack the probabilistic framework that scales well for large graphs. In this paper, we propose two efficient algorithms for computing P-Rank on large graphs. The first observation is that a large body of objects in a real graph usually share similar neighborhood structures. By merging such objects with an explicit low-rank factorization, we devise a deterministic algorithm to compute P-Rank in quadratic time. The second observation is that by converting the iterative form of P-Rank into a matrix power series form, we can leverage the random sampling approach to probabilistically compute P-Rank in linear time with provable accuracy guarantees. The empirical results on both real and synthetic datasets show that our approaches achieve high time efficiency with controlled error and outperform the baseline algorithms by at least one order of magnitude
DeepInf: Social Influence Prediction with Deep Learning
Social and information networking activities such as on Facebook, Twitter,
WeChat, and Weibo have become an indispensable part of our everyday life, where
we can easily access friends' behaviors and are in turn influenced by them.
Consequently, an effective social influence prediction for each user is
critical for a variety of applications such as online recommendation and
advertising.
Conventional social influence prediction approaches typically design various
hand-crafted rules to extract user- and network-specific features. However,
their effectiveness heavily relies on the knowledge of domain experts. As a
result, it is usually difficult to generalize them into different domains.
Inspired by the recent success of deep neural networks in a wide range of
computing applications, we design an end-to-end framework, DeepInf, to learn
users' latent feature representation for predicting social influence. In
general, DeepInf takes a user's local network as the input to a graph neural
network for learning her latent social representation. We design strategies to
incorporate both network structures and user-specific features into
convolutional neural and attention networks. Extensive experiments on Open
Academic Graph, Twitter, Weibo, and Digg, representing different types of
social and information networks, demonstrate that the proposed end-to-end
model, DeepInf, significantly outperforms traditional feature engineering-based
approaches, suggesting the effectiveness of representation learning for social
applications.Comment: 10 pages, 5 figures, to appear in KDD 2018 proceeding
Harmony in the Small-World
The Small-World phenomenon, popularly known as six degrees of separation, has
been mathematically formalized by Watts and Strogatz in a study of the
topological properties of a network. Small-worlds networks are defined in terms
of two quantities: they have a high clustering coefficient C like regular
lattices and a short characteristic path length L typical of random networks.
Physical distances are of fundamental importance in the applications to real
cases, nevertheless this basic ingredient is missing in the original
formulation. Here we introduce a new concept, the connectivity length D, that
gives harmony to the whole theory. D can be evaluated on a global and on a
local scale and plays in turn the role of L and 1/C. Moreover it can be
computed for any metrical network and not only for the topological cases. D has
a precise meaning in term of information propagation and describes in an
unified way both the structural and the dynamical aspects of a network:
small-worlds are defined by a small global and local D, i.e. by a high
efficiency in propagating information both on a local and on a global scale.
The neural system of the nematode C. elegans, the collaboration graph of film
actors, and the oldest U.S. subway system, can now be studied also as metrical
networks and are shown to be small-worlds.Comment: 16 pages, 3 figures, accepted for publication in Physica
Software tools for conducting bibliometric analysis in science: An up-to-date review
Bibliometrics has become an essential tool for assessing and analyzing the output of scientists, cooperation between
universities, the effect of state-owned science funding on national research and development performance and educational
efficiency, among other applications. Therefore, professionals and scientists need a range of theoretical and practical
tools to measure experimental data. This review aims to provide an up-to-date review of the various tools available
for conducting bibliometric and scientometric analyses, including the sources of data acquisition, performance analysis
and visualization tools. The included tools were divided into three categories: general bibliometric and performance
analysis, science mapping analysis, and libraries; a description of all of them is provided. A comparative analysis of the
database sources support, pre-processing capabilities, analysis and visualization options were also provided in order to
facilitate its understanding. Although there are numerous bibliometric databases to obtain data for bibliometric and
scientometric analysis, they have been developed for a different purpose. The number of exportable records is between
500 and 50,000 and the coverage of the different science fields is unequal in each database. Concerning the analyzed
tools, Bibliometrix contains the more extensive set of techniques and suitable for practitioners through Biblioshiny.
VOSviewer has a fantastic visualization and is capable of loading and exporting information from many sources. SciMAT
is the tool with a powerful pre-processing and export capability. In views of the variability of features, the users need to
decide the desired analysis output and chose the option that better fits into their aims
- …