15 research outputs found
Security evaluation of protected template in biometric cryptosystems
制度:新 ; 報告番号:甲3538号 ; 学位の種類:博士(工学) ; 授与年月日:2012/3/15 ; 早大学位記番号:新587
EdgePruner: Poisoned Edge Pruning in Graph Contrastive Learning
Graph Contrastive Learning (GCL) is unsupervised graph representation
learning that can obtain useful representation of unknown nodes. The node
representation can be utilized as features of downstream tasks. However, GCL is
vulnerable to poisoning attacks as with existing learning models. A
state-of-the-art defense cannot sufficiently negate adverse effects by poisoned
graphs although such a defense introduces adversarial training in the GCL. To
achieve further improvement, pruning adversarial edges is important. To the
best of our knowledge, the feasibility remains unexplored in the GCL domain. In
this paper, we propose a simple defense for GCL, EdgePruner. We focus on the
fact that the state-of-the-art poisoning attack on GCL tends to mainly add
adversarial edges to create poisoned graphs, which means that pruning edges is
important to sanitize the graphs. Thus, EdgePruner prunes edges that contribute
to minimizing the contrastive loss based on the node representation obtained
after training on poisoned graphs by GCL. Furthermore, we focus on the fact
that nodes with distinct features are connected by adversarial edges in
poisoned graphs. Thus, we introduce feature similarity between neighboring
nodes to help more appropriately determine adversarial edges. This similarity
is helpful in further eliminating adverse effects from poisoned graphs on
various datasets. Finally, EdgePruner outputs a graph that yields the minimum
contrastive loss as the sanitized graph. Our results demonstrate that pruning
adversarial edges is feasible on six datasets. EdgePruner can improve the
accuracy of node classification under the attack by up to 5.55% compared with
that of the state-of-the-art defense. Moreover, we show that EdgePruner is
immune to an adaptive attack
Detecting Machine-Translated Text using Back Translation
Machine-translated text plays a crucial role in the communication of people
using different languages. However, adversaries can use such text for malicious
purposes such as plagiarism and fake review. The existing methods detected a
machine-translated text only using the text's intrinsic content, but they are
unsuitable for classifying the machine-translated and human-written texts with
the same meanings. We have proposed a method to extract features used to
distinguish machine/human text based on the similarity between the intrinsic
text and its back-translation. The evaluation of detecting translated sentences
with French shows that our method achieves 75.0% of both accuracy and F-score.
It outperforms the existing methods whose the best accuracy is 62.8% and the
F-score is 62.7%. The proposed method even detects more efficiently the
back-translated text with 83.4% of accuracy, which is higher than 66.7% of the
best previous accuracy. We also achieve similar results not only with F-score
but also with similar experiments related to Japanese. Moreover, we prove that
our detector can recognize both machine-translated and machine-back-translated
texts without the language information which is used to generate these machine
texts. It demonstrates the persistence of our method in various applications in
both low- and rich-resource languages.Comment: INLG 2019, 9 page
VoteTRANS: Detecting Adversarial Text without Training by Voting on Hard Labels of Transformations
Adversarial attacks reveal serious flaws in deep learning models. More
dangerously, these attacks preserve the original meaning and escape human
recognition. Existing methods for detecting these attacks need to be trained
using original/adversarial data. In this paper, we propose detection without
training by voting on hard labels from predictions of transformations, namely,
VoteTRANS. Specifically, VoteTRANS detects adversarial text by comparing the
hard labels of input text and its transformation. The evaluation demonstrates
that VoteTRANS effectively detects adversarial text across various
state-of-the-art attacks, models, and datasets.Comment: Findings of ACL 2023 (long paper
Degree-Preserving Randomized Response for Graph Neural Networks under Local Differential Privacy
Differentially private GNNs (Graph Neural Networks) have been recently
studied to provide high accuracy in various tasks on graph data while strongly
protecting user privacy. In particular, a recent study proposes an algorithm to
protect each user's feature vector in an attributed graph with LDP (Local
Differential Privacy), a strong privacy notion without a trusted third party.
However, this algorithm does not protect edges (friendships) in a social graph
or protect user privacy in unattributed graphs. It remains open how to strongly
protect edges with LDP while keeping high accuracy in GNNs.
In this paper, we propose a novel LDP algorithm called the DPRR
(Degree-Preserving Randomized Response) to provide LDP for edges in GNNs. Our
DPRR preserves each user's degree hence a graph structure while providing edge
LDP. Technically, we use Warner's RR (Randomized Response) and strategic edge
sampling, where each user's sampling probability is automatically tuned to
preserve the degree information. We prove that the DPRR approximately preserves
the degree information under edge LDP. We focus on graph classification as a
task of GNNs and evaluate the DPRR using two social graph datasets. Our
experimental results show that the DPRR significantly outperforms Warner's RR
and provides accuracy close to a non-private algorithm with a reasonable
privacy budget, e.g., epsilon=1
TransMIA: Membership Inference Attacks Using Transfer Shadow Training
Transfer learning has been widely studied and gained increasing popularity to
improve the accuracy of machine learning models by transferring some knowledge
acquired in different training. However, no prior work has pointed out that
transfer learning can strengthen privacy attacks on machine learning models. In
this paper, we propose TransMIA (Transfer learning-based Membership Inference
Attacks), which use transfer learning to perform membership inference attacks
on the source model when the adversary is able to access the parameters of the
transferred model. In particular, we propose a transfer shadow training
technique, where an adversary employs the parameters of the transferred model
to construct shadow models, to significantly improve the performance of
membership inference when a limited amount of shadow training data is available
to the adversary. We evaluate our attacks using two real datasets, and show
that our attacks outperform the state-of-the-art that does not use our transfer
shadow training technique. We also compare four combinations of the
learning-based/entropy-based approach and the fine-tuning/freezing approach,
all of which employ our transfer shadow training technique. Then we examine the
performance of these four approaches based on the distributions of confidence
values, and discuss possible countermeasures against our attacks.Comment: IJCNN 2021 conference pape
Anonymization Technique based on SGD Matrix Factrization
Time-sequence data is high dimensional and contains a lot of information, which can be utilized in various fields, such as insurance, finance, and advertising. Personal data including time-sequence data is converted to anonymized datasets, which need to strike a balance between both privacy and utility. In this paper, we consider low-rank matrix factorization as one of anonymization methods and evaluate its efficiency. We convert time-sequence datasets to matrices and evaluate both privacy and utility. The record IDs in time-sequence data are changed at regular intervals to reduce re-identification risk. However, since individuals tend to behave in a similar fashion over periods of time, there remains a risk of record linkage even if record IDs are different. Hence, we evaluate the re-identification and linkage risks as privacy risks of time-sequence data. Our experimental results show that matrix factorization is a viable anonymization method and it can achieve better utility than existing anonymization methods