13,379 research outputs found
Transfer Learning across Networks for Collective Classification
This paper addresses the problem of transferring useful knowledge from a
source network to predict node labels in a newly formed target network. While
existing transfer learning research has primarily focused on vector-based data,
in which the instances are assumed to be independent and identically
distributed, how to effectively transfer knowledge across different information
networks has not been well studied, mainly because networks may have their
distinct node features and link relationships between nodes. In this paper, we
propose a new transfer learning algorithm that attempts to transfer common
latent structure features across the source and target networks. The proposed
algorithm discovers these latent features by constructing label propagation
matrices in the source and target networks, and mapping them into a shared
latent feature space. The latent features capture common structure patterns
shared by two networks, and serve as domain-independent features to be
transferred between networks. Together with domain-dependent node features, we
thereafter propose an iterative classification algorithm that leverages label
correlations to predict node labels in the target network. Experiments on
real-world networks demonstrate that our proposed algorithm can successfully
achieve knowledge transfer between networks to help improve the accuracy of
classifying nodes in the target network.Comment: Published in the proceedings of IEEE ICDM 201
CSNE: Conditional Signed Network Embedding
Signed networks are mathematical structures that encode positive and negative
relations between entities such as friend/foe or trust/distrust. Recently,
several papers studied the construction of useful low-dimensional
representations (embeddings) of these networks for the prediction of missing
relations or signs. Existing embedding methods for sign prediction generally
enforce different notions of status or balance theories in their optimization
function. These theories, however, are often inaccurate or incomplete, which
negatively impacts method performance.
In this context, we introduce conditional signed network embedding (CSNE).
Our probabilistic approach models structural information about the signs in the
network separately from fine-grained detail. Structural information is
represented in the form of a prior, while the embedding itself is used for
capturing fine-grained information. These components are then integrated in a
rigorous manner. CSNE's accuracy depends on the existence of sufficiently
powerful structural priors for modelling signed networks, currently unavailable
in the literature. Thus, as a second main contribution, which we find to be
highly valuable in its own right, we also introduce a novel approach to
construct priors based on the Maximum Entropy (MaxEnt) principle. These priors
can model the \emph{polarity} of nodes (degree to which their links are
positive) as well as signed \emph{triangle counts} (a measure of the degree
structural balance holds to in a network).
Experiments on a variety of real-world networks confirm that CSNE outperforms
the state-of-the-art on the task of sign prediction. Moreover, the MaxEnt
priors on their own, while less accurate than full CSNE, achieve accuracies
competitive with the state-of-the-art at very limited computational cost, thus
providing an excellent runtime-accuracy trade-off in resource-constrained
situations
CSNE : Conditional Signed Network Embedding
Signed networks are mathematical structures that encode positive and negative relations between entities such as friend/foe or trust/distrust. Recently, several papers studied the construction of useful low-dimensional representations (embeddings) of these networks for the prediction of missing relations or signs. Existing embedding methods for sign prediction generally enforce different notions of status or balance theories in their optimization function. These theories, however, are often inaccurate or incomplete, which negatively impacts method performance.
In this context, we introduce conditional signed network embedding (CSNE). Our probabilistic approach models structural information about the signs in the network separately from fine-grained detail. Structural information is represented in the form of a prior, while the embedding itself is used for capturing fine-grained information. These components are then integrated in a rigorous manner. CSNE's accuracy depends on the existence of sufficiently powerful structural priors for modelling signed networks, currently unavailable in the literature. Thus, as a second main contribution, which we find to be highly valuable in its own right, we also introduce a novel approach to construct priors based on the Maximum Entropy (MaxEnt) principle. These priors can model the polarity of nodes (degree to which their links are positive) as well as signed triangle counts (a measure of the degree structural balance holds to in a network).
Experiments on a variety of real-world networks confirm that CSNE outperforms the state-of-the-art on the task of sign prediction. Moreover, the MaxEnt priors on their own, while less accurate than full CSNE, achieve accuracies competitive with the state-of-the-art at very limited computational cost, thus providing an excellent runtime-accuracy trade-off in resource-constrained situations
- …