16,447 research outputs found
Finite Blocklength Analysis of Gaussian Random coding in AWGN Channels under Covert constraints II: A Viewpoint of Total Variation Distance
Covert communication over an additive white Gaussian noise (AWGN) channel
with finite block length is investigated in this paper. The attention is on the
covert criterion, which has not been considered in finite block length
circumstance. As an accurate quantity metric of discrimination, the variation
distance with given finite block length n and signal-noise ratio (snr) is
obtained. We give both its analytic solution and expansions which can be easily
evaluated. It is shown that K-L distance, which is frequently adopted as the
metric of discrimination at the adversary in asymptotic regime, is not
convincing in finite block length regime compared with the total variation
distance. Moreover, the convergence rate of the total variation with different
snr is analyzed when the block length tends to infinity. The results will be
very helpful for understanding the behavior of the total variation distance and
practical covert communication
Finite Blocklength Analysis of Gaussian Random Coding in AWGN Channels under Covert Constraint
This paper considers the achievability and converse bounds on the maximal
channel coding rate at a given blocklength and error probability over AWGN
channels. The problem stems from covert communication with Gaussian codewords.
By re-visiting [18], we first present new and more general achievability bounds
for random coding schemes under maximal or average probability of error
requirements. Such general bounds are then applied to covert communication in
AWGN channels where codewords are generated from Gaussian distribution while
meeting the maximal power constraint. Further comparison is made between the
new achievability bounds and existing one with deterministic codebooks.Comment: 18 page
Visual Anomaly Detection via Dual-Attention Transformer and Discriminative Flow
In this paper, we introduce the novel state-of-the-art Dual-attention
Transformer and Discriminative Flow (DADF) framework for visual anomaly
detection. Based on only normal knowledge, visual anomaly detection has wide
applications in industrial scenarios and has attracted significant attention.
However, most existing methods fail to meet the requirements. In contrast, the
proposed DTDF presents a new paradigm: it firstly leverages a pre-trained
network to acquire multi-scale prior embeddings, followed by the development of
a vision Transformer with dual attention mechanisms, namely self-attention and
memorial-attention, to achieve two-level reconstruction for prior embeddings
with the sequential and normality association. Additionally, we propose using
normalizing flow to establish discriminative likelihood for the joint
distribution of prior and reconstructions at each scale. The DADF achieves
98.3/98.4 of image/pixel AUROC on Mvtec AD; 83.7 of image AUROC and 67.4 of
pixel sPRO on Mvtec LOCO AD benchmarks, demonstrating the effectiveness of our
proposed approach.Comment: Submission to IEEE Transactions On Industrial Informatic
- …