21 research outputs found
Smoothed Differential Privacy
Differential privacy (DP) is a widely-accepted and widely-applied notion of
privacy based on worst-case analysis. Often, DP classifies most mechanisms
without additive noise as non-private (Dwork et al., 2014). Thus, additive
noises are added to improve privacy (to achieve DP). However, in many
real-world applications, adding additive noise is undesirable (Bagdasaryan et
al., 2019) and sometimes prohibited (Liu et al., 2020).
In this paper, we propose a natural extension of DP following the worst
average-case idea behind the celebrated smoothed analysis (Spielman & Teng, May
2004). Our notion, smoothed DP, can effectively measure the privacy leakage of
mechanisms without additive noises under realistic settings. We prove that any
discrete mechanism with sampling procedures is more private than what DP
predicts, while many continuous mechanisms with sampling procedures are still
non-private under smoothed DP. In addition, we prove several desirable
properties of smoothed DP, including composition, robustness to
post-processing, and distribution reduction. Based on those properties, we
propose an efficient algorithm to calculate the privacy parameters for smoothed
DP. Experimentally, we verify that, according to smoothed DP, the discrete
sampling mechanisms are private in real-world elections, and some discrete
neural networks can be private without adding any additive noise. We believe
that these results contribute to the theoretical foundation of realistic
privacy measures beyond worst-case analysis.Comment: 16 Page main text + Appendi
How Private Are Commonly-Used Voting Rules?
Differential privacy has been widely applied to provide privacy guarantees by
adding random noise to the function output. However, it inevitably fails in
many high-stakes voting scenarios, where voting rules are required to be
deterministic. In this work, we present the first framework for answering the
question: "How private are commonly-used voting rules?" Our answers are
two-fold. First, we show that deterministic voting rules provide sufficient
privacy in the sense of distributional differential privacy (DDP). We show that
assuming the adversarial observer has uncertainty about individual votes, even
publishing the histogram of votes achieves good DDP. Second, we introduce the
notion of exact privacy to compare the privacy preserved in various
commonly-studied voting rules, and obtain dichotomy theorems of exact DDP
within a large subset of voting rules called generalized scoring rules
Learning the heterogeneous representation of brain's structure from serial SEM images using a masked autoencoder
IntroductionThe exorbitant cost of accurately annotating the large-scale serial scanning electron microscope (SEM) images as the ground truth for training has always been a great challenge for brain map reconstruction by deep learning methods in neural connectome studies. The representation ability of the model is strongly correlated with the number of such high-quality labels. Recently, the masked autoencoder (MAE) has been shown to effectively pre-train Vision Transformers (ViT) to improve their representational capabilities.MethodsIn this paper, we investigated a self-pre-training paradigm for serial SEM images with MAE to implement downstream segmentation tasks. We randomly masked voxels in three-dimensional brain image patches and trained an autoencoder to reconstruct the neuronal structures.Results and discussionWe tested different pre-training and fine-tuning configurations on three different serial SEM datasets of mouse brains, including two public ones, SNEMI3D and MitoEM-R, and one acquired in our lab. A series of masking ratios were examined and the optimal ratio for pre-training efficiency was spotted for 3D segmentation. The MAE pre-training strategy significantly outperformed the supervised learning from scratch. Our work shows that the general framework of can be a unified approach for effective learning of the representation of heterogeneous neural structural features in serial SEM images to greatly facilitate brain connectome reconstruction
PPSP: prediction of PK-specific phosphorylation site with Bayesian decision theory
BACKGROUND: As a reversible and dynamic post-translational modification (PTM) of proteins, phosphorylation plays essential regulatory roles in a broad spectrum of the biological processes. Although many studies have been contributed on the molecular mechanism of phosphorylation dynamics, the intrinsic feature of substrates specificity is still elusive and remains to be delineated. RESULTS: In this work, we present a novel, versatile and comprehensive program, PPSP (Prediction of PK-specific Phosphorylation site), deployed with approach of Bayesian decision theory (BDT). PPSP could predict the potential phosphorylation sites accurately for ~70 PK (Protein Kinase) groups. Compared with four existing tools Scansite, NetPhosK, KinasePhos and GPS, PPSP is more accurate and powerful than these tools. Moreover, PPSP also provides the prediction for many novel PKs, say, TRK, mTOR, SyK and MET/RON, etc. The accuracy of these novel PKs are also satisfying. CONCLUSION: Taken together, we propose that PPSP could be a potentially powerful tool for the experimentalists who are focusing on phosphorylation substrates with their PK-specific sites identification. Moreover, the BDT strategy could also be a ubiquitous approach for PTMs, such as sumoylation and ubiquitination, etc
Differential Privacy for Eye-Tracking Data
As large eye-tracking datasets are created, data privacy is a pressing concern for the eye-tracking community. De-identifying data does not guarantee privacy because multiple datasets can be linked for inferences. A common belief is that aggregating individuals' data into composite representations such as heatmaps protects the individual. However, we analytically examine the privacy of (noise-free) heatmaps and show that they do not guarantee privacy. We further propose two noise mechanisms that guarantee privacy and analyze their privacy-utility tradeoff. Analysis reveals that our Gaussian noise mechanism is an elegant solution to preserve privacy for heatmaps. Our results have implications for interdisciplinary research to create differentially private mechanisms for eye tracking
SpeechUT: Bridging Speech and Text with Hidden-Unit for Encoder-Decoder Based Speech-Text Pre-training
The rapid development of single-modal pre-training has prompted researchers
to pay more attention to cross-modal pre-training methods. In this paper, we
propose a unified-modal speech-unit-text pre-training model, SpeechUT, to
connect the representations of a speech encoder and a text decoder with a
shared unit encoder. Leveraging hidden-unit as an interface to align speech and
text, we can decompose the speech-to-text model into a speech-to-unit model and
a unit-to-text model, which can be jointly pre-trained with unpaired speech and
text data respectively. Our proposed SpeechUT is fine-tuned and evaluated on
automatic speech recognition (ASR) and speech translation (ST) tasks.
Experimental results show that SpeechUT gets substantial improvements over
strong baselines, and achieves state-of-the-art performance on both the
LibriSpeech ASR and MuST-C ST tasks. To better understand the proposed
SpeechUT, detailed analyses are conducted. The code and pre-trained models are
available at https://aka.ms/SpeechUT.Comment: 14 pages, accepted by EMNLP 202
Pre-Training Transformer Decoder for End-to-End ASR Model with Unpaired Speech Data
This paper studies a novel pre-training technique with unpaired speech data,
Speech2C, for encoder-decoder based automatic speech recognition (ASR). Within
a multi-task learning framework, we introduce two pre-training tasks for the
encoder-decoder network using acoustic units, i.e., pseudo codes, derived from
an offline clustering model. One is to predict the pseudo codes via masked
language modeling in encoder output, like HuBERT model, while the other lets
the decoder learn to reconstruct pseudo codes autoregressively instead of
generating textual scripts. In this way, the decoder learns to reconstruct
original speech information with codes before learning to generate correct
text. Comprehensive experiments on the LibriSpeech corpus show that the
proposed Speech2C can relatively reduce the word error rate (WER) by 19.2% over
the method without decoder pre-training, and also outperforms significantly the
state-of-the-art wav2vec 2.0 and HuBERT on fine-tuning subsets of 10h and 100h.Comment: Submitted to INTERSPEECH 202
Near-Neighbor Methods in Random Preference Completion
This paper studies a stylized, yet natural, learning-to-rank problem and points out the critical incorrectness of a widely used nearest neighbor algorithm. We consider a model with n agents (users) {xi}i∈[n] and m alternatives (items) {yl}l∈[m], each of which is associated with a latent feature vector. Agents rank items nondeterministically according to the Plackett-Luce model, where the higher the utility of an item to the agent, the more likely this item will be ranked high by the agent. Our goal is to identify near neighbors of an arbitrary agent in the latent space for prediction.We first show that the Kendall-tau distance based kNN produces incorrect results in our model. Next, we propose a new anchor-based algorithm to find neighbors of an agent. A salient feature of our algorithm is that it leverages the rankings of many other agents (the so-called “anchors”) to determine the closeness/similarities of two agents. We provide a rigorous analysis for one-dimensional latent space, and complement the theoretical results with experiments on synthetic and real datasets. The experiments confirm that the new algorithm is robust and practical