169 research outputs found
Broadband linearisation of high-efficiency power amplifiers
A feedforward-based amplifier linearization technique is presented which is capable of yielding significant improvements in both linearity and power efficiency over conventional amplifier classes (e.g. class-A or class-AB). Theoretical and practical results are presented showing that class-C stages may be used for both the main and error amplifiers yielding practical efficiencies well in excess of 30 percent, with theoretical efficiencies of much greater than 40 percent being possible. The levels of linearity which may be achieved are required for most satellite systems, however if greater linearity is required, the technique may be used in addition to conventional pre-distortion techniques
Nonlinearity-Tolerant Modulation Formats for Coherent Optical Communications
Fiber nonlinearity is the main factor limiting the transmission distance of coherent optical communications. We overview several modulation formats intrinsically tolerant to fiber nonlinearity. We recently proposed family of 4D modulation formats based on 2-ary amplitude 8-ary phase-shift keying (2A8PSK), covering the spectral efficiency of 5, 6, and 7 bits/4D symbol, which will be explained in detail in this chapter. These coded modulation formats fill the gap of spectral efficiency between DP-QPSK and DP-16QAM, showing superb performance both in linear and nonlinear regimes. Since these modulation formats share the same constellation and use different parity bit expressions only, digital signal processing can accommodate those multiple modulation formats with minimum additional complexity. Nonlinear transmission simulations indicate that these modulation formats outperform the conventional formats at each spectral efficiency. We also review DSP algorithms and experimental results. Their application to time-domain hybrid modulation for 4–8 bits/4D symbol is also reviewed. Furthermore, an overview of an eight-dimensional 2A8PSK-based modulation format based on a Grassmann code is also given. All these results indicate that the 4D-2A8PSK family show great promise of excellent linear and nonlinear performances in the spectral efficiency between 3.5 and 8 bits/4D symbol
Why Does Differential Privacy with Large Epsilon Defend Against Practical Membership Inference Attacks?
For small privacy parameter , -differential privacy (DP)
provides a strong worst-case guarantee that no membership inference attack
(MIA) can succeed at determining whether a person's data was used to train a
machine learning model. The guarantee of DP is worst-case because: a) it holds
even if the attacker already knows the records of all but one person in the
data set; and b) it holds uniformly over all data sets. In practical
applications, such a worst-case guarantee may be overkill: practical attackers
may lack exact knowledge of (nearly all of) the private data, and our data set
might be easier to defend, in some sense, than the worst-case data set. Such
considerations have motivated the industrial deployment of DP models with large
privacy parameter (e.g. ), and it has been observed
empirically that DP with large can successfully defend against
state-of-the-art MIAs. Existing DP theory cannot explain these empirical
findings: e.g., the theoretical privacy guarantees of are
essentially vacuous. In this paper, we aim to close this gap between theory and
practice and understand why a large DP parameter can prevent practical MIAs. To
tackle this problem, we propose a new privacy notion called practical
membership privacy (PMP). PMP models a practical attacker's uncertainty about
the contents of the private data. The PMP parameter has a natural
interpretation in terms of the success rate of a practical MIA on a given data
set. We quantitatively analyze the PMP parameter of two fundamental DP
mechanisms: the exponential mechanism and Gaussian mechanism. Our analysis
reveals that a large DP parameter often translates into a much smaller PMP
parameter, which guarantees strong privacy against practical MIAs. Using our
findings, we offer principled guidance for practitioners in choosing the DP
parameter.Comment: Accepted at PPAI-24: AAAI Workshop on Privacy-Preserving Artificial
Intelligenc
Stabilizing Subject Transfer in EEG Classification with Divergence Estimation
Classification models for electroencephalogram (EEG) data show a large
decrease in performance when evaluated on unseen test sub jects. We reduce this
performance decrease using new regularization techniques during model training.
We propose several graphical models to describe an EEG classification task.
From each model, we identify statistical relationships that should hold true in
an idealized training scenario (with infinite data and a globally-optimal
model) but that may not hold in practice. We design regularization penalties to
enforce these relationships in two stages. First, we identify suitable proxy
quantities (divergences such as Mutual Information and Wasserstein-1) that can
be used to measure statistical independence and dependence relationships.
Second, we provide algorithms to efficiently estimate these quantities during
training using secondary neural network models. We conduct extensive
computational experiments using a large benchmark EEG dataset, comparing our
proposed techniques with a baseline method that uses an adversarial classifier.
We find our proposed methods significantly increase balanced accuracy on test
subjects and decrease overfitting. The proposed methods exhibit a larger
benefit over a greater range of hyperparameters than the baseline method, with
only a small computational cost at training time. These benefits are largest
when used for a fixed training period, though there is still a significant
benefit for a subset of hyperparameters when our techniques are used in
conjunction with early stopping regularization.Comment: 16 pages, 5 figure
- …