64,684 research outputs found
CSNE: Conditional Signed Network Embedding
Signed networks are mathematical structures that encode positive and negative
relations between entities such as friend/foe or trust/distrust. Recently,
several papers studied the construction of useful low-dimensional
representations (embeddings) of these networks for the prediction of missing
relations or signs. Existing embedding methods for sign prediction generally
enforce different notions of status or balance theories in their optimization
function. These theories, however, are often inaccurate or incomplete, which
negatively impacts method performance.
In this context, we introduce conditional signed network embedding (CSNE).
Our probabilistic approach models structural information about the signs in the
network separately from fine-grained detail. Structural information is
represented in the form of a prior, while the embedding itself is used for
capturing fine-grained information. These components are then integrated in a
rigorous manner. CSNE's accuracy depends on the existence of sufficiently
powerful structural priors for modelling signed networks, currently unavailable
in the literature. Thus, as a second main contribution, which we find to be
highly valuable in its own right, we also introduce a novel approach to
construct priors based on the Maximum Entropy (MaxEnt) principle. These priors
can model the \emph{polarity} of nodes (degree to which their links are
positive) as well as signed \emph{triangle counts} (a measure of the degree
structural balance holds to in a network).
Experiments on a variety of real-world networks confirm that CSNE outperforms
the state-of-the-art on the task of sign prediction. Moreover, the MaxEnt
priors on their own, while less accurate than full CSNE, achieve accuracies
competitive with the state-of-the-art at very limited computational cost, thus
providing an excellent runtime-accuracy trade-off in resource-constrained
situations
CSNE : Conditional Signed Network Embedding
Signed networks are mathematical structures that encode positive and negative relations between entities such as friend/foe or trust/distrust. Recently, several papers studied the construction of useful low-dimensional representations (embeddings) of these networks for the prediction of missing relations or signs. Existing embedding methods for sign prediction generally enforce different notions of status or balance theories in their optimization function. These theories, however, are often inaccurate or incomplete, which negatively impacts method performance.
In this context, we introduce conditional signed network embedding (CSNE). Our probabilistic approach models structural information about the signs in the network separately from fine-grained detail. Structural information is represented in the form of a prior, while the embedding itself is used for capturing fine-grained information. These components are then integrated in a rigorous manner. CSNE's accuracy depends on the existence of sufficiently powerful structural priors for modelling signed networks, currently unavailable in the literature. Thus, as a second main contribution, which we find to be highly valuable in its own right, we also introduce a novel approach to construct priors based on the Maximum Entropy (MaxEnt) principle. These priors can model the polarity of nodes (degree to which their links are positive) as well as signed triangle counts (a measure of the degree structural balance holds to in a network).
Experiments on a variety of real-world networks confirm that CSNE outperforms the state-of-the-art on the task of sign prediction. Moreover, the MaxEnt priors on their own, while less accurate than full CSNE, achieve accuracies competitive with the state-of-the-art at very limited computational cost, thus providing an excellent runtime-accuracy trade-off in resource-constrained situations
Factorial graphical lasso for dynamic networks
Dynamic networks models describe a growing number of important scientific
processes, from cell biology and epidemiology to sociology and finance. There
are many aspects of dynamical networks that require statistical considerations.
In this paper we focus on determining network structure. Estimating dynamic
networks is a difficult task since the number of components involved in the
system is very large. As a result, the number of parameters to be estimated is
bigger than the number of observations. However, a characteristic of many
networks is that they are sparse. For example, the molecular structure of genes
make interactions with other components a highly-structured and therefore
sparse process.
Penalized Gaussian graphical models have been used to estimate sparse
networks. However, the literature has focussed on static networks, which lack
specific temporal constraints. We propose a structured Gaussian dynamical
graphical model, where structures can consist of specific time dynamics, known
presence or absence of links and block equality constraints on the parameters.
Thus, the number of parameters to be estimated is reduced and accuracy of the
estimates, including the identification of the network, can be tuned up. Here,
we show that the constrained optimization problem can be solved by taking
advantage of an efficient solver, logdetPPA, developed in convex optimization.
Moreover, model selection methods for checking the sensitivity of the inferred
networks are described. Finally, synthetic and real data illustrate the
proposed methodologies.Comment: 30 pp, 5 figure
Incremental construction of LSTM recurrent neural network
Long Short--Term Memory (LSTM) is a recurrent neural network that
uses structures called memory blocks to allow the net remember
significant events distant in the past input sequence in order to
solve long time lag tasks, where other RNN approaches fail.
Throughout this work we have performed experiments using LSTM
networks extended with growing abilities, which we call GLSTM.
Four methods of training growing LSTM has been compared. These
methods include cascade and fully connected hidden layers as well
as two different levels of freezing previous weights in the
cascade case. GLSTM has been applied to a forecasting problem in a biomedical domain, where the input/output behavior of five
controllers of the Central Nervous System control has to be
modelled. We have compared growing LSTM results against other
neural networks approaches, and our work applying conventional
LSTM to the task at hand.Postprint (published version
Foundations and modelling of dynamic networks using Dynamic Graph Neural Networks: A survey
Dynamic networks are used in a wide range of fields, including social network
analysis, recommender systems, and epidemiology. Representing complex networks
as structures changing over time allow network models to leverage not only
structural but also temporal patterns. However, as dynamic network literature
stems from diverse fields and makes use of inconsistent terminology, it is
challenging to navigate. Meanwhile, graph neural networks (GNNs) have gained a
lot of attention in recent years for their ability to perform well on a range
of network science tasks, such as link prediction and node classification.
Despite the popularity of graph neural networks and the proven benefits of
dynamic network models, there has been little focus on graph neural networks
for dynamic networks. To address the challenges resulting from the fact that
this research crosses diverse fields as well as to survey dynamic graph neural
networks, this work is split into two main parts. First, to address the
ambiguity of the dynamic network terminology we establish a foundation of
dynamic networks with consistent, detailed terminology and notation. Second, we
present a comprehensive survey of dynamic graph neural network models using the
proposed terminologyComment: 28 pages, 9 figures, 8 table
Topological Feature Based Classification
There has been a lot of interest in developing algorithms to extract clusters
or communities from networks. This work proposes a method, based on
blockmodelling, for leveraging communities and other topological features for
use in a predictive classification task. Motivated by the issues faced by the
field of community detection and inspired by recent advances in Bayesian topic
modelling, the presented model automatically discovers topological features
relevant to a given classification task. In this way, rather than attempting to
identify some universal best set of clusters for an undefined goal, the aim is
to find the best set of clusters for a particular purpose.
Using this method, topological features can be validated and assessed within
a given context by their predictive performance.
The proposed model differs from other relational and semi-supervised learning
models as it identifies topological features to explain the classification
decision. In a demonstration on a number of real networks the predictive
capability of the topological features are shown to rival the performance of
content based relational learners. Additionally, the model is shown to
outperform graph-based semi-supervised methods on directed and approximately
bipartite networks.Comment: Awarded 3rd Best Student Paper at 14th International Conference on
Information Fusion 201
New Ideas for Brain Modelling
This paper describes some biologically-inspired processes that could be used
to build the sort of networks that we associate with the human brain. New to
this paper, a 'refined' neuron will be proposed. This is a group of neurons
that by joining together can produce a more analogue system, but with the same
level of control and reliability that a binary neuron would have. With this new
structure, it will be possible to think of an essentially binary system in
terms of a more variable set of values. The paper also shows how recent
research associated with the new model, can be combined with established
theories, to produce a more complete picture. The propositions are largely in
line with conventional thinking, but possibly with one or two more radical
suggestions. An earlier cognitive model can be filled in with more specific
details, based on the new research results, where the components appear to fit
together almost seamlessly. The intention of the research has been to describe
plausible 'mechanical' processes that can produce the appropriate brain
structures and mechanisms, but that could be used without the magical
'intelligence' part that is still not fully understood. There are also some
important updates from an earlier version of this paper
- …