1,761 research outputs found
Learning Vertex Representations for Bipartite Networks
Recent years have witnessed a widespread increase of interest in network
representation learning (NRL). By far most research efforts have focused on NRL
for homogeneous networks like social networks where vertices are of the same
type, or heterogeneous networks like knowledge graphs where vertices (and/or
edges) are of different types. There has been relatively little research
dedicated to NRL for bipartite networks. Arguably, generic network embedding
methods like node2vec and LINE can also be applied to learn vertex embeddings
for bipartite networks by ignoring the vertex type information. However, these
methods are suboptimal in doing so, since real-world bipartite networks concern
the relationship between two types of entities, which usually exhibit different
properties and patterns from other types of network data. For example,
E-Commerce recommender systems need to capture the collaborative filtering
patterns between customers and products, and search engines need to consider
the matching signals between queries and webpages
GPSP: Graph Partition and Space Projection based Approach for Heterogeneous Network Embedding
In this paper, we propose GPSP, a novel Graph Partition and Space Projection
based approach, to learn the representation of a heterogeneous network that
consists of multiple types of nodes and links. Concretely, we first partition
the heterogeneous network into homogeneous and bipartite subnetworks. Then, the
projective relations hidden in bipartite subnetworks are extracted by learning
the projective embedding vectors. Finally, we concatenate the projective
vectors from bipartite subnetworks with the ones learned from homogeneous
subnetworks to form the final representation of the heterogeneous network.
Extensive experiments are conducted on a real-life dataset. The results
demonstrate that GPSP outperforms the state-of-the-art baselines in two key
network mining tasks: node classification and clustering.Comment: WWW 2018 Poste
Algebraic shortcuts for leave-one-out cross-validation in supervised network inference
Supervised machine learning techniques have traditionally been very successful at reconstructing biological networks, such as protein-ligand interaction, protein-protein interaction and gene regulatory networks. Many supervised techniques for network prediction use linear models on a possibly nonlinear pairwise feature representation of edges. Recently, much emphasis has been placed on the correct evaluation of such supervised models. It is vital to distinguish between using a model to either predict new interactions in a given network or to predict interactions for a new vertex not present in the original network. This distinction matters because (i) the performance might dramatically differ between the prediction settings and (ii) tuning the model hyperparameters to obtain the best possible model depends on the setting of interest. Specific cross-validation schemes need to be used to assess the performance in such different prediction settings. In this work we discuss a state-of-the-art kernel-based network inference technique called two-step kernel ridge regression. We show that this regression model can be trained efficiently, with a time complexity scaling with the number of vertices rather than the number of edges. Furthermore, this framework leads to a series of cross-validation shortcuts that allow one to rapidly estimate the model performance for any relevant network prediction setting. This allows computational biologists to fully assess the capabilities of their models
- …