409 research outputs found
Transfer Learning across Networks for Collective Classification
This paper addresses the problem of transferring useful knowledge from a
source network to predict node labels in a newly formed target network. While
existing transfer learning research has primarily focused on vector-based data,
in which the instances are assumed to be independent and identically
distributed, how to effectively transfer knowledge across different information
networks has not been well studied, mainly because networks may have their
distinct node features and link relationships between nodes. In this paper, we
propose a new transfer learning algorithm that attempts to transfer common
latent structure features across the source and target networks. The proposed
algorithm discovers these latent features by constructing label propagation
matrices in the source and target networks, and mapping them into a shared
latent feature space. The latent features capture common structure patterns
shared by two networks, and serve as domain-independent features to be
transferred between networks. Together with domain-dependent node features, we
thereafter propose an iterative classification algorithm that leverages label
correlations to predict node labels in the target network. Experiments on
real-world networks demonstrate that our proposed algorithm can successfully
achieve knowledge transfer between networks to help improve the accuracy of
classifying nodes in the target network.Comment: Published in the proceedings of IEEE ICDM 201
Search Efficient Binary Network Embedding
Traditional network embedding primarily focuses on learning a dense vector
representation for each node, which encodes network structure and/or node
content information, such that off-the-shelf machine learning algorithms can be
easily applied to the vector-format node representations for network analysis.
However, the learned dense vector representations are inefficient for
large-scale similarity search, which requires to find the nearest neighbor
measured by Euclidean distance in a continuous vector space. In this paper, we
propose a search efficient binary network embedding algorithm called BinaryNE
to learn a sparse binary code for each node, by simultaneously modeling node
context relations and node attribute relations through a three-layer neural
network. BinaryNE learns binary node representations efficiently through a
stochastic gradient descent based online learning algorithm. The learned binary
encoding not only reduces memory usage to represent each node, but also allows
fast bit-wise comparisons to support much quicker network node search compared
to Euclidean distance or other distance measures. Our experiments and
comparisons show that BinaryNE not only delivers more than 23 times faster
search speed, but also provides comparable or better search quality than
traditional continuous vector based network embedding methods
ATNPA: A Unified View of Oversmoothing Alleviation in Graph Neural Networks
Oversmoothing is a commonly observed challenge in graph neural network (GNN)
learning, where, as layers increase, embedding features learned from GNNs
quickly become similar/indistinguishable, making them incapable of
differentiating network proximity. A GNN with shallow layer architectures can
only learn short-term relation or localized structure information, limiting its
power of learning long-term connection, evidenced by their inferior learning
performance on heterophilous graphs. Tackling oversmoothing is crucial to
harness deep-layer architectures for GNNs. To date, many methods have been
proposed to alleviate oversmoothing. The vast difference behind their design
principles, combined with graph complications, make it difficult to understand
and even compare their difference in tackling the oversmoothing. In this paper,
we propose ATNPA, a unified view with five key steps: Augmentation,
Transformation, Normalization, Propagation, and Aggregation, to summarize GNN
oversmoothing alleviation approaches. We first outline three themes to tackle
oversmoothing, and then separate all methods into six categories, followed by
detailed reviews of representative methods, including their relation to the
ATNPA, and discussion about their niche, strength, and weakness. The review not
only draws in-depth understanding of existing methods in the field, but also
shows a clear road map for future study.Comment: 16 page
Uncovering dynamically critical regions in near-wall turbulence using 3D Convolutional Neural Networks
Near-wall regions in wall-bounded turbulent flows experience strong
intermittent events involving ejections of slow-moving fluid parcels away from
the wall, and `sweeps' of faster moving fluid towards the wall. Here, we train
a three-dimensional Convolutional Neural Network (CNN) to predict the intensity
of ejection events that occur in Direct Numerical Simulation (DNS) of a
periodic channel flow. The trained network is able to predict burst intensities
accurately for flow snaphshots that are sufficiently removed from the training
data so as to be temporally decorrelated. More importantly, we probe the
trained network to reveal regions of the flow where the network focuses its
attention in order to make a prediction. We find that these salient regions
correlate very well with fluid parcels being ejected away from the wall.
Moreover, the CNN is able to keep track of the salient fluid parcels as the
flow evolves in time. This demonstrates that CNNs are capable of discovering
dynamically critical phenomena in turbulent flows without requiring any
a-priori knowledge of the underlying dynamics. Remarkably, the trained CNN is
able to predict ejection intensities accurately for data at different Reynolds
numbers, which highlights its ability to identify physical processes that
persist across varying flow conditions. The results presented here highlight
the immense potential of CNNs for discovering and analyzing nonlinear spatial
correlations in turbulent flows.Comment: 10 pages, 7 figure
Counting Manatee Aggregations using Deep Neural Networks and Anisotropic Gaussian Kernel
Manatees are aquatic mammals with voracious appetites. They rely on sea grass
as the main food source, and often spend up to eight hours a day grazing. They
move slow and frequently stay in group (i.e. aggregations) in shallow water to
search for food, making them vulnerable to environment change and other risks.
Accurate counting manatee aggregations within a region is not only biologically
meaningful in observing their habit, but also crucial for designing safety
rules for human boaters, divers, etc., as well as scheduling nursing,
intervention, and other plans. In this paper, we propose a deep learning based
crowd counting approach to automatically count number of manatees within a
region, by using low quality images as input. Because manatees have unique
shape and they often stay in shallow water in groups, water surface reflection,
occlusion, camouflage etc. making it difficult to accurately count manatee
numbers. To address the challenges, we propose to use Anisotropic Gaussian
Kernel (AGK), with tunable rotation and variances, to ensure that density
functions can maximally capture shapes of individual manatees in different
aggregations. After that, we apply AGK kernel to different types of deep neural
networks primarily designed for crowd counting, including VGG, SANet, Congested
Scene Recognition network (CSRNet), MARUNet etc. to learn manatee densities and
calculate number of manatees in the scene. By using generic low quality images
extracted from surveillance videos, our experiment results and comparison show
that AGK kernel based manatee counting achieves minimum Mean Absolute Error
(MAE) and Root Mean Square Error (RMSE). The proposed method works particularly
well for counting manatee aggregations in environments with complex background.Comment: 18 pages, 8 figures, 2 tables, 3 algorithms, and it has been accepted
for publication in Scientific Report
- …