9 research outputs found
A Principled Hierarchical Deep Learning Approach to Joint Image Compression and Classification
Among applications of deep learning (DL) involving low cost sensors, remote
image classification involves a physical channel that separates edge sensors
and cloud classifiers. Traditional DL models must be divided between an encoder
for the sensor and the decoder + classifier at the edge server. An important
challenge is to effectively train such distributed models when the connecting
channels have limited rate/capacity. Our goal is to optimize DL models such
that the encoder latent requires low channel bandwidth while still delivers
feature information for high classification accuracy. This work proposes a
three-step joint learning strategy to guide encoders to extract features that
are compact, discriminative, and amenable to common
augmentations/transformations. We optimize latent dimension through an initial
screening phase before end-to-end (E2E) training. To obtain an adjustable bit
rate via a single pre-deployed encoder, we apply entropy-based quantization
and/or manual truncation on the latent representations. Tests show that our
proposed method achieves accuracy improvement of up to 1.5% on CIFAR-10 and 3%
on CIFAR-100 over conventional E2E cross-entropy training
Recommended from our members
End-to-End Joint Image Compression and Deep Learning under Bandwidth Constrained Environments
The past decade has witnessed the rising dominance of deep learning (DL) and artificial intelligence (AI) in a wide range of applications. In particular, the ocean of wireless smart phones and IoT devices continue to fuel the tremendous growth of edge/cloud-based machine learning (ML) systems including image/video recognition and classification. To overcome the infrastructural barrier of limited network bandwidth in cloud ML, existing solutions have mainly relied on traditional compression codecs such as JPEG that were historically engineered for human-end users instead of ML algorithms. Traditional codecs do not necessarily preserve features important to ML algorithms under limited bandwidth, leading to potentially inferior performance. This dissertation investigates application-driven optimization of programmable commercial codec settings for networked learning tasks such as image classification.In the first part of this dissertation, we focus on the efficient use and optimization of existing off-the-shelf commercial image compression codecs in bandwidth constrained image classification applications. We consider a cloud-based inference application where a power and memory limited embedded source device transmits the collected images to a powerful cloud server over bandlimited wireless channels. Our main contributions are two folds. Firstly, we show that the reconstruction step of the existing image decoders is unnecessary for cloud-based inference. Deep learning classifiers designed to take intermediate features as inputs, instead of RGB images, can perform inference few times faster with the same or improved classification accuracy. Secondly, we show that redesigning the entropy coders of commercial image codec such as JPEG2000 and learning optimal parameter setting of the entropy coders for a given task in end-to-end manner can significantly improve rate-accuracy performance of the codec.In the second part, we investigate the methods of improving rate-distortion-accuracy performance in cloud-based AI applications for DL-based image compression codecs. Exploring end-to-end optimization of the complete codec, we propose novel classifier architectures based on variational auto-encoders (VAE) that outperform rate-classification accuracy of several conventional codecs. Further investigating DL-based codecs, we discuss how to achieve better rate-distortion-accuracy performance with end-to-end training revisiting the concept of region of interest (ROI).In the third part of this dissertation, we explore recent interpretable information theory based concepts when modeling real world data and their applicability in data constrained deep learning scenarios. In particular, we investigate the use of linear discriminative representations (LDR) of images when designing cloud-based deep learning systems with improved rate- accuracy performance. Further, considering challenging but practical data constrained tasks such as zero-shot and few-shot learning, we investigate the generalization of such linear feature representations learned with rate reduction concepts
Interplay between vertical sectorization and user distribution for urban NB-IoT networks
Publisher Copyright: © 2022 IEEE.Vertical sectorization introduces considerable gain to particular scenarios where user equipment (UE) are distributed in 3D domain, i.e. high-rise buildings, in terms of coverage and network capacity thanks to availability of active antenna systems. However, due to the huge varieties in distribution of UEs and different physical environments, presenting a comprehensive analytical framework is quite challenging. From this aspect, most available studies on vertical sectorization are limited to present only empirical results. In this paper, we introduce a novel methodology to forecast the performance of NB-IoT systems over urban scenarios. In particular, a logistic distribution-based analytical framework is exploited in order to calculate the group probabilities for each available UE. Based on these probabilities, we propose a scheduling framework with beamforming which improves physical resource block (PRB) utilization by over 50% compared to the case with no scheduling.Peer reviewe