973 research outputs found
Multi-Source Multi-View Clustering via Discrepancy Penalty
With the advance of technology, entities can be observed in multiple views.
Multiple views containing different types of features can be used for
clustering. Although multi-view clustering has been successfully applied in
many applications, the previous methods usually assume the complete instance
mapping between different views. In many real-world applications, information
can be gathered from multiple sources, while each source can contain multiple
views, which are more cohesive for learning. The views under the same source
are usually fully mapped, but they can be very heterogeneous. Moreover, the
mappings between different sources are usually incomplete and partially
observed, which makes it more difficult to integrate all the views across
different sources. In this paper, we propose MMC (Multi-source Multi-view
Clustering), which is a framework based on collective spectral clustering with
a discrepancy penalty across sources, to tackle these challenges. MMC has
several advantages compared with other existing methods. First, MMC can deal
with incomplete mapping between sources. Second, it considers the disagreements
between sources while treating views in the same source as a cohesive set.
Third, MMC also tries to infer the instance similarities across sources to
enhance the clustering performance. Extensive experiments conducted on
real-world data demonstrate the effectiveness of the proposed approach
Task-Oriented Communication for Edge Video Analytics
With the development of artificial intelligence (AI) techniques and the
increasing popularity of camera-equipped devices, many edge video analytics
applications are emerging, calling for the deployment of computation-intensive
AI models at the network edge. Edge inference is a promising solution to move
the computation-intensive workloads from low-end devices to a powerful edge
server for video analytics, but the device-server communications will remain a
bottleneck due to the limited bandwidth. This paper proposes a task-oriented
communication framework for edge video analytics, where multiple devices
collect the visual sensory data and transmit the informative features to an
edge server for processing. To enable low-latency inference, this framework
removes video redundancy in spatial and temporal domains and transmits minimal
information that is essential for the downstream task, rather than
reconstructing the videos at the edge server. Specifically, it extracts compact
task-relevant features based on the deterministic information bottleneck (IB)
principle, which characterizes a tradeoff between the informativeness of the
features and the communication cost. As the features of consecutive frames are
temporally correlated, we propose a temporal entropy model (TEM) to reduce the
bitrate by taking the previous features as side information in feature
encoding. To further improve the inference performance, we build a
spatial-temporal fusion module at the server to integrate features of the
current and previous frames for joint inference. Extensive experiments on video
analytics tasks evidence that the proposed framework effectively encodes
task-relevant information of video data and achieves a better rate-performance
tradeoff than existing methods
Task-Oriented Communication for Multi-Device Cooperative Edge Inference
This paper investigates task-oriented communication for multi-device
cooperative edge inference, where a group of distributed low-end edge devices
transmit the extracted features of local samples to a powerful edge server for
inference. While cooperative edge inference can overcome the limited sensing
capability of a single device, it substantially increases the communication
overhead and may incur excessive latency. To enable low-latency cooperative
inference, we propose a learning-based communication scheme that optimizes
local feature extraction and distributed feature encoding in a task-oriented
manner, i.e., to remove data redundancy and transmit information that is
essential for the downstream inference task rather than reconstructing the data
samples at the edge server. Specifically, we leverage an information bottleneck
(IB) principle to extract the task-relevant feature at each edge device and
adopt a distributed information bottleneck (DIB) framework to formalize a
single-letter characterization of the optimal rate-relevance tradeoff for
distributed feature encoding. To admit flexible control of the communication
overhead, we extend the DIB framework to a distributed deterministic
information bottleneck (DDIB) objective that explicitly incorporates the
representational costs of the encoded features. As the IB-based objectives are
computationally prohibitive for high-dimensional data, we adopt variational
approximations to make the optimization problems tractable. To compensate the
potential performance loss due to the variational approximations, we also
develop a selective retransmission (SR) mechanism to identify the redundancy in
the encoded features of multiple edge devices to attain additional
communication overhead reduction. Extensive experiments evidence that the
proposed task-oriented communication scheme achieves a better rate-relevance
tradeoff than baseline methods.Comment: This paper was accepted to IEEE Transactions on Wireless
Communicatio
Selective Knowledge Sharing for Privacy-Preserving Federated Distillation without A Good Teacher
While federated learning is promising for privacy-preserving collaborative
learning without revealing local data, it remains vulnerable to white-box
attacks and struggles to adapt to heterogeneous clients. Federated distillation
(FD), built upon knowledge distillation--an effective technique for
transferring knowledge from a teacher model to student models--emerges as an
alternative paradigm, which provides enhanced privacy guarantees and addresses
model heterogeneity. Nevertheless, challenges arise due to variations in local
data distributions and the absence of a well-trained teacher model, which leads
to misleading and ambiguous knowledge sharing that significantly degrades model
performance. To address these issues, this paper proposes a selective knowledge
sharing mechanism for FD, termed Selective-FD. It includes client-side
selectors and a server-side selector to accurately and precisely identify
knowledge from local and ensemble predictions, respectively. Empirical studies,
backed by theoretical insights, demonstrate that our approach enhances the
generalization capabilities of the FD framework and consistently outperforms
baseline methods. This study presents a promising direction for effective
knowledge transfer in privacy-preserving collaborative learning
LDMIC: Learning-based Distributed Multi-view Image Coding
Multi-view image compression plays a critical role in 3D-related
applications. Existing methods adopt a predictive coding architecture, which
requires joint encoding to compress the corresponding disparity as well as
residual information. This demands collaboration among cameras and enforces the
epipolar geometric constraint between different views, which makes it
challenging to deploy these methods in distributed camera systems with randomly
overlapping fields of view. Meanwhile, distributed source coding theory
indicates that efficient data compression of correlated sources can be achieved
by independent encoding and joint decoding, which motivates us to design a
learning-based distributed multi-view image coding (LDMIC) framework. With
independent encoders, LDMIC introduces a simple yet effective joint context
transfer module based on the cross-attention mechanism at the decoder to
effectively capture the global inter-view correlations, which is insensitive to
the geometric relationships between images. Experimental results show that
LDMIC significantly outperforms both traditional and learning-based MIC methods
while enjoying fast encoding speed. Code will be released at
https://github.com/Xinjie-Q/LDMIC.Comment: Accepted by ICLR 202
Branchy-GNN: a Device-Edge Co-Inference Framework for Efficient Point Cloud Processing
The recent advancements of three-dimensional (3D) data acquisition devices
have spurred a new breed of applications that rely on point cloud data
processing. However, processing a large volume of point cloud data brings a
significant workload on resource-constrained mobile devices, prohibiting from
unleashing their full potentials. Built upon the emerging paradigm of
device-edge co-inference, where an edge device extracts and transmits the
intermediate feature to an edge server for further processing, we propose
Branchy-GNN for efficient graph neural network (GNN) based point cloud
processing by leveraging edge computing platforms. In order to reduce the
on-device computational cost, the Branchy-GNN adds branch networks for early
exiting. Besides, it employs learning-based joint source-channel coding (JSCC)
for the intermediate feature compression to reduce the communication overhead.
Our experimental results demonstrate that the proposed Branchy-GNN secures a
significant latency reduction compared with several benchmark methods
- …