60 research outputs found
A method of storing vector data in compressed form using clustering
The development of the machine learning algorithms for information search in recent years made it possible to represent
text and multimodal documents in the form of vectors. These vector representations (embeddings) preserve the semantic content of documents and allow the search to be performed as the calculation of distance between vectors. Compressing
embeddings can reduce the amount of memory they occupy and improve computational efficiency. The article discusses
existing methods for compressing vector representations without loss of accuracy and with loss of accuracy. A method
is proposed to reduce error by clustering vector representations using lossy compression. The essence of the method
is in performing the preliminary clustering of vector representations, saving the centers of each cluster, and saving
the coordinate value of each vector representation relative to the center of its cluster. Then, the centers of each cluster
are compressed without loss of accuracy, and the resulting shifted vector representations are compressed with loss of
accuracy. To restore the original vector representations, the coordinates of the center of the corresponding cluster are
added to the coordinates of the displaced representation. The proposed method was tested on the fashion-mnist-784-
euclidean and NYT-256-angular datasets. A comparison has been made of compressed vector representations with
loss of accuracy by reducing the bit depth with vector representations compressed using the proposed method. With a
slight (around 10 %) increase in the size of the compressed data, the absolute value of the error from loss of accuracy
decreased by four and two times, respectively, for the tested sets. The developed method can be applied in tasks where
it is necessary to store and process vector representations of multimodal documents, for example, in the development
of search engines
Low-Complexity Vector Source Coding for Discrete Long Sequences with Unknown Distributions
In this paper, we propose a source coding scheme that represents data from
unknown distributions through frequency and support information. Existing
encoding schemes often compress data by sacrificing computational efficiency or
by assuming the data follows a known distribution. We take advantage of the
structure that arises within the spatial representation and utilize it to
encode run-lengths within this representation using Golomb coding. Through
theoretical analysis, we show that our scheme yields an overall bit rate that
nears entropy without a computationally complex encoding algorithm and verify
these results through numerical experiments.Comment: This work has been submitted to the IEEE for possible publication.
Copyright may be transferred without notice, after which this version may no
longer be accessibl
SnakeVoxFormer: Transformer-based Single Image\\Voxel Reconstruction with Run Length Encoding
Deep learning-based 3D object reconstruction has achieved unprecedented
results. Among those, the transformer deep neural model showed outstanding
performance in many applications of computer vision. We introduce
SnakeVoxFormer, a novel, 3D object reconstruction in voxel space from a single
image using the transformer. The input to SnakeVoxFormer is a 2D image, and the
result is a 3D voxel model. The key novelty of our approach is in using the
run-length encoding that traverses (like a snake) the voxel space and encodes
wide spatial differences into a 1D structure that is suitable for transformer
encoding. We then use dictionary encoding to convert the discovered RLE blocks
into tokens that are used for the transformer. The 1D representation is a
lossless 3D shape data compression method that converts to 1D data that use
only about 1% of the original data size. We show how different voxel traversing
strategies affect the effect of encoding and reconstruction. We compare our
method with the state-of-the-art for 3D voxel reconstruction from images and
our method improves the state-of-the-art methods by at least 2.8% and up to
19.8%
Communication-Efficient Federated Learning for Wireless Edge Intelligence in IoT
This is the author accepted manuscript. The final version is available from IEEE via the DOI in this recordThe rapidly expanding number of IoT devices is
generating huge quantities of data, but public concern over
data privacy means users are apprehensive to send data to a
central server for Machine Learning (ML) purposes. The easilychanged behaviours of edge infrastructure that Software Defined
Networking provides makes it possible to collate IoT data at
edge servers and gateways, where Federated Learning (FL) can
be performed: building a central model without uploading data to
the server. FedAvg is a FL algorithm which has been the subject
of much study, however it suffers from a large number of rounds
to convergence with non-Independent, Identically Distributed
(non-IID) client datasets and high communication costs per
round. We propose adapting FedAvg to use a distributed form
of Adam optimisation, greatly reducing the number of rounds
to convergence, along with novel compression techniques, to
produce Communication-Efficient FedAvg (CE-FedAvg). We perform extensive experiments with the MNIST/CIFAR-10 datasets,
IID/non-IID client data, varying numbers of clients, client participation rates, and compression rates. These show CE-FedAvg
can converge to a target accuracy in up to 6× less rounds than
similarly compressed FedAvg, while uploading up to 3× less data,
and is more robust to aggressive compression. Experiments on
an edge-computing-like testbed using Raspberry Pi clients also
show CE-FedAvg is able to reach a target accuracy in up to 1.7×
less real time than FedAvg.Engineering and Physical Sciences Research Council (EPSRC
Sparse Binary Compression: Towards Distributed Deep Learning with minimal Communication
Currently, progressively larger deep neural networks are trained on ever
growing data corpora. As this trend is only going to increase in the future,
distributed training schemes are becoming increasingly relevant. A major issue
in distributed training is the limited communication bandwidth between
contributing nodes or prohibitive communication cost in general. These
challenges become even more pressing, as the number of computation nodes
increases. To counteract this development we propose sparse binary compression
(SBC), a compression framework that allows for a drastic reduction of
communication cost for distributed training. SBC combines existing techniques
of communication delay and gradient sparsification with a novel binarization
method and optimal weight update encoding to push compression gains to new
limits. By doing so, our method also allows us to smoothly trade-off gradient
sparsity and temporal sparsity to adapt to the requirements of the learning
task. Our experiments show, that SBC can reduce the upstream communication on a
variety of convolutional and recurrent neural network architectures by more
than four orders of magnitude without significantly harming the convergence
speed in terms of forward-backward passes. For instance, we can train ResNet50
on ImageNet in the same number of iterations to the baseline accuracy, using
less bits or train it to a lower accuracy using less bits. In the latter case, the total upstream communication required
is cut from 125 terabytes to 3.35 gigabytes for every participating client
EO-ALERT: NEXT GENERATION SATELLITE PROCESSING CHAIN FOR RAPID CIVIL ALERTS
In this paper, we provide an overview of the H2020 EU project EO-ALERT. The aim of EO-ALERT is to propose the definition and development of the next generation Earth observation (EO) data and processing chain, based on a novel flight segment architecture moving optimised key EO data processing elements from the ground segment to on-board the satellite. The objective is to address the need for increased throughput in EO data chain, delivering EO products to the end user with very low latency
- …