1,259 research outputs found
Pornographic Image Recognition via Weighted Multiple Instance Learning
In the era of Internet, recognizing pornographic images is of great
significance for protecting children's physical and mental health. However,
this task is very challenging as the key pornographic contents (e.g., breast
and private part) in an image often lie in local regions of small size. In this
paper, we model each image as a bag of regions, and follow a multiple instance
learning (MIL) approach to train a generic region-based recognition model.
Specifically, we take into account the region's degree of pornography, and make
three main contributions. First, we show that based on very few annotations of
the key pornographic contents in a training image, we can generate a bag of
properly sized regions, among which the potential positive regions usually
contain useful contexts that can aid recognition. Second, we present a simple
quantitative measure of a region's degree of pornography, which can be used to
weigh the importance of different regions in a positive image. Third, we
formulate the recognition task as a weighted MIL problem under the
convolutional neural network framework, with a bag probability function
introduced to combine the importance of different regions. Experiments on our
newly collected large scale dataset demonstrate the effectiveness of the
proposed method, achieving an accuracy with 97.52% true positive rate at 1%
false positive rate, tested on 100K pornographic images and 100K normal images.Comment: 9 pages, 3 figure
Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning
Deep Learning has recently become hugely popular in machine learning,
providing significant improvements in classification accuracy in the presence
of highly-structured and large databases.
Researchers have also considered privacy implications of deep learning.
Models are typically trained in a centralized manner with all the data being
processed by the same training algorithm. If the data is a collection of users'
private data, including habits, personal pictures, geographical positions,
interests, and more, the centralized server will have access to sensitive
information that could potentially be mishandled. To tackle this problem,
collaborative deep learning models have recently been proposed where parties
locally train their deep learning structures and only share a subset of the
parameters in the attempt to keep their respective training sets private.
Parameters can also be obfuscated via differential privacy (DP) to make
information extraction even more challenging, as proposed by Shokri and
Shmatikov at CCS'15.
Unfortunately, we show that any privacy-preserving collaborative deep
learning is susceptible to a powerful attack that we devise in this paper. In
particular, we show that a distributed, federated, or decentralized deep
learning approach is fundamentally broken and does not protect the training
sets of honest participants. The attack we developed exploits the real-time
nature of the learning process that allows the adversary to train a Generative
Adversarial Network (GAN) that generates prototypical samples of the targeted
training set that was meant to be private (the samples generated by the GAN are
intended to come from the same distribution as the training data).
Interestingly, we show that record-level DP applied to the shared parameters of
the model, as suggested in previous work, is ineffective (i.e., record-level DP
is not designed to address our attack).Comment: ACM CCS'17, 16 pages, 18 figure
Multi-perspective cost-sensitive context-aware multi-instance sparse coding and its application to sensitive video recognition
With the development of video-sharing websites, P2P, micro-blog, mobile WAP websites, and so on, sensitive videos can be more easily accessed. Effective sensitive video recognition is necessary for web content security. Among web sensitive videos, this paper focuses on violent and horror videos. Based on color emotion and color harmony theories, we extract visual emotional features from videos. A video is viewed as a bag and each shot in the video is represented by a key frame which is treated as an instance in the bag. Then, we combine multi-instance learning (MIL) with sparse coding to recognize violent and horror videos. The resulting MIL-based model can be updated online to adapt to changing web environments. We propose a cost-sensitive context-aware multi- instance sparse coding (MI-SC) method, in which the contextual structure of the key frames is modeled using a graph, and fusion between audio and visual features is carried out by extending the classic sparse coding into cost-sensitive sparse coding. We then propose a multi-perspective multi- instance joint sparse coding (MI-J-SC) method that handles each bag of instances from an independent perspective, a contextual perspective, and a holistic perspective. The experiments demonstrate that the features with an emotional meaning are effective for violent and horror video recognition, and our cost-sensitive context-aware MI-SC and multi-perspective MI-J-SC methods outperform the traditional MIL methods and the traditional SVM and KNN-based methods
Development of automatic obscene images filtering using deep learning
Because of Internet availability in most societies, access to pornography has be-come a severe issue. On the other side, the pornography industry has grown steadily, and its websites are becoming increasingly popular by offering potential users free passes. Filtering obscene images and video frames is essential in the big data era, where all kinds of information are available for everyone. This paper proposes a fully automated method to filter any storage device from obscene vid-eos and images using deep learning algorithms. The whole recognition process can be divided into two stages, including fine detection and focus detection. The fine detection includes skin color detection with YCbCr and HSV color spaces and accurate face detection using the Adaboost algorithm with Haar-like features. Moreover, focus detection uses AlexNet transfer learning to identify the obscene images which passed stage one. Results showed the effectiveness of our pro-posed algorithm in filtering obscene images or videos. The testing accuracy achieved is 95.26% when tested with 3969 testing images
- …