1,122 research outputs found
Fast and Effective Bag-of-Visual-Word Model to Pornographic Images Recognition Using the FREAK Descriptor
Recently, the Bag of Visual Word (BoVW) has gained enormous popularity between researchers to object recognition. Pornographic image recognition with respect to computational complexity, appropriate accuracy, and memory consumption is a major challenge in the applications with time constraints such as the internet pornography filtering. Most of the existing researches based on the Bow, using the very popular SIFT and SURF algorithms to description and match detected keypoints in the image. The main problem of these methods is high computational complexity due to constructing the high dimensional feature vectors. This research proposed a BoVW based model by adopting very fast and simple binary descriptor FREAK to speed-up pornographic recognition process. Meanwhile, the keypoints are detected in the ROI of images which improves the recognition speed due to eliminating many noise keypoints placed in the image background. Finally, in order to find the most representational visual-vocabulary, different vocabularies are generated from size 150 to 500 for BoVW. Compared with the similar works, the experimental results show that the proposed model has gained remarkable improvement in the terms of computational complexity
Pornographic Image Recognition via Weighted Multiple Instance Learning
In the era of Internet, recognizing pornographic images is of great
significance for protecting children's physical and mental health. However,
this task is very challenging as the key pornographic contents (e.g., breast
and private part) in an image often lie in local regions of small size. In this
paper, we model each image as a bag of regions, and follow a multiple instance
learning (MIL) approach to train a generic region-based recognition model.
Specifically, we take into account the region's degree of pornography, and make
three main contributions. First, we show that based on very few annotations of
the key pornographic contents in a training image, we can generate a bag of
properly sized regions, among which the potential positive regions usually
contain useful contexts that can aid recognition. Second, we present a simple
quantitative measure of a region's degree of pornography, which can be used to
weigh the importance of different regions in a positive image. Third, we
formulate the recognition task as a weighted MIL problem under the
convolutional neural network framework, with a bag probability function
introduced to combine the importance of different regions. Experiments on our
newly collected large scale dataset demonstrate the effectiveness of the
proposed method, achieving an accuracy with 97.52% true positive rate at 1%
false positive rate, tested on 100K pornographic images and 100K normal images.Comment: 9 pages, 3 figure
Smart Content Recognition from Images Using a Mixture of Convolutional Neural Networks
With rapid development of the Internet, web contents become huge. Most of the
websites are publicly available, and anyone can access the contents from
anywhere such as workplace, home and even schools. Nevertheless, not all the
web contents are appropriate for all users, especially children. An example of
these contents is pornography images which should be restricted to certain age
group. Besides, these images are not safe for work (NSFW) in which employees
should not be seen accessing such contents during work. Recently, convolutional
neural networks have been successfully applied to many computer vision
problems. Inspired by these successes, we propose a mixture of convolutional
neural networks for adult content recognition. Unlike other works, our method
is formulated on a weighted sum of multiple deep neural network models. The
weights of each CNN models are expressed as a linear regression problem learned
using Ordinary Least Squares (OLS). Experimental results demonstrate that the
proposed model outperforms both single CNN model and the average sum of CNN
models in adult content recognition.Comment: To be published in LNEE, Code: github.com/mundher/NSF
Automating the construction of scene classifiers for content-based video retrieval
This paper introduces a real time automatic scene classifier within content-based video retrieval. In our envisioned approach end users like documentalists, not image processing experts, build classifiers interactively, by simply indicating positive examples of a scene. Classification consists of a two stage procedure. First, small image fragments called patches are classified. Second, frequency vectors of these patch classifications are fed into a second classifier for global scene classification (e.g., city, portraits, or countryside). The first stage classifiers can be seen as a set of highly specialized, learned feature detectors, as an alternative to letting an image processing expert determine features a priori. We present results for experiments on a variety of patch and image classes. The scene classifier has been used successfully within television archives and for Internet porn filtering
Validating Multimedia Content Moderation Software via Semantic Fusion
The exponential growth of social media platforms, such as Facebook and
TikTok, has revolutionized communication and content publication in human
society. Users on these platforms can publish multimedia content that delivers
information via the combination of text, audio, images, and video. Meanwhile,
the multimedia content release facility has been increasingly exploited to
propagate toxic content, such as hate speech, malicious advertisements, and
pornography. To this end, content moderation software has been widely deployed
on these platforms to detect and blocks toxic content. However, due to the
complexity of content moderation models and the difficulty of understanding
information across multiple modalities, existing content moderation software
can fail to detect toxic content, which often leads to extremely negative
impacts.
We introduce Semantic Fusion, a general, effective methodology for validating
multimedia content moderation software. Our key idea is to fuse two or more
existing single-modal inputs (e.g., a textual sentence and an image) into a new
input that combines the semantics of its ancestors in a novel manner and has
toxic nature by construction. This fused input is then used for validating
multimedia content moderation software. We realized Semantic Fusion as DUO, a
practical content moderation software testing tool. In our evaluation, we
employ DUO to test five commercial content moderation software and two
state-of-the-art models against three kinds of toxic content. The results show
that DUO achieves up to 100% error finding rate (EFR) when testing moderation
software. In addition, we leverage the test cases generated by DUO to retrain
the two models we explored, which largely improves model robustness while
maintaining the accuracy on the original test set.Comment: Accepted by ISSTA 202
- …