13 research outputs found

    Fast pornographic image recognition using compact holistic features and multi-layer neural network

    Get PDF
    The paper presents an alternative fast pornographic image recognition using compact holistic features and multi-layer neural network (MNN). The compact holistic features of pornographic images, which are invariant features against pose and scale, is extracted by shape and frequency analysis on pornographic images under skin region of interests (ROIs). The main objective of this work is to design pornographic recognition scheme which not only can improve performances of existing methods (i.e., methods based on skin probability, scale invariant feature transform, eigenporn, and Multilayer-Perceptron and Neuro-Fuzzy (MP-NF)) but also can works fast for recognition. The experimental outcome display that our proposed system can improve 0.3% of accuracy and reduce 6.60% the false negative rate (FNR) of the best existing method (skin probability and eigenporn on YCbCr, SEP), respectively. Additionally, our proposed method also provides almost similar robust performances to the MP-NF on large size dataset. However, our proposed method needs short recognition time by about 0.021 seconds per image for both tested datasets

    Distinguishing Medical Web Pages from Pornographic Ones: An Efficient Pornography Websites Filtering Method

    Get PDF
    Abstract In this paper, we apply the uncomplicated decision tree data mining algorithm to find association rules about pornographic and medical web pages. On the basis of these association rules, we propose a systematized method of filtering pornographic websites with the following major superiorities: 1) Check only contexts of web pages without scanning pictures to avoid the low operating efficiency in analyzing photographs. Moreover, the error rate is lowered and the accuracy of filtering is enhanced simultaneously. 2) While filtering the pornographic web pages accurately, the misjudgments of identifying medical web pages as pornographic ones will be reduced effectively. 3) A re-learning mechanism is designed to improve our filtering method incrementally. Therefore, the revision information learned from the misjudged web pages can incrementally give feedback to our method and improve its effectiveness. The experimental results showed that each efficacy assessment indexes reached a satisfactory value. Therefore, we can conclude that the proposed method is possessed of outstanding performance and effectivity

    Intelligent instance selection techniques for support vector machine speed optimization with application to e-fraud detection.

    Get PDF
    Doctor of Philosophy in Computer Science. University of KwaZulu-Natal, Durban 2017.Decision-making is a very important aspect of many businesses. There are grievous penalties involved in wrong decisions, including financial loss, damage of company reputation and reduction in company productivity. Hence, it is of dire importance that managers make the right decisions. Machine Learning (ML) simplifies the process of decision making: it helps to discover useful patterns from historical data, which can be used for meaningful decision-making. The ability to make strategic and meaningful decisions is dependent on the reliability of data. Currently, many organizations are overwhelmed with vast amounts of data, and unfortunately, ML algorithms cannot effectively handle large datasets. This thesis therefore proposes seven filter-based and five wrapper-based intelligent instance selection techniques for optimizing the speed and predictive accuracy of ML algorithms, with a particular focus on Support Vector Machine (SVM). Also, this thesis proposes a novel fitness function for instance selection. The primary difference between the filter-based and wrapper-based technique is in their method of selection. The filter-based techniques utilizes the proposed fitness function for selection, while the wrapper-based technique utilizes SVM algorithm for selection. The proposed techniques are obtained by fusing SVM algorithm with the following Nature Inspired algorithms: flower pollination algorithm, social spider algorithm, firefly algorithm, cuckoo search algorithm and bat algorithm. Also, two of the filter-based techniques are boundary detection algorithms, inspired by edge detection in image processing and edge selection in ant colony optimization. Two different sets of experiments were performed in order to evaluate the performance of the proposed techniques (wrapper-based and filter-based). All experiments were performed on four datasets containing three popular e-fraud types: credit card fraud, email spam and phishing email. In addition, experiments were performed on 20 datasets provided by the well-known UCI data repository. The results show that the proposed filter-based techniques excellently improved SVM training speed in 100% (24 out of 24) of the datasets used for evaluation, without significantly affecting SVM classification quality. Moreover, experimental results also show that the wrapper-based techniques consistently improved SVM predictive accuracy in 78% (18 out of 23) of the datasets used for evaluation and simultaneously improved SVM training speed in all cases. Furthermore, two different statistical tests were conducted to further validate the credibility of the results: Freidman’s test and Holm’s post-hoc test. The statistical test results reveal that the proposed filter-based and wrapper-based techniques are significantly faster, compared to standard SVM and some existing instance selection techniques, in all cases. Moreover, statistical test results also reveal that Cuckoo Search Instance Selection Algorithm outperform all the proposed techniques, in terms of speed. Overall, the proposed techniques have proven to be fast and accurate ML-based e-fraud detection techniques, with improved training speed, predictive accuracy and storage reduction. In real life application, such as video surveillance and intrusion detection systems, that require a classifier to be trained very quickly for speedy classification of new target concepts, the filter-based techniques provide the best solutions; while the wrapper-based techniques are better suited for applications, such as email filters, that are very sensitive to slight changes in predictive accuracy

    Cyber Security and Critical Infrastructures

    Get PDF
    This book contains the manuscripts that were accepted for publication in the MDPI Special Topic "Cyber Security and Critical Infrastructure" after a rigorous peer-review process. Authors from academia, government and industry contributed their innovative solutions, consistent with the interdisciplinary nature of cybersecurity. The book contains 16 articles: an editorial explaining current challenges, innovative solutions, real-world experiences including critical infrastructure, 15 original papers that present state-of-the-art innovative solutions to attacks on critical systems, and a review of cloud, edge computing, and fog's security and privacy issues

    Classificação de sites a partir das análises estrutural e textual

    Get PDF
    With the wide use of the web nowadays, also with its constant growth, task of automatic classification of websites has gained increasing importance. In many occasions it is necessary to block access to specific sites, such as in the case of access to adult content sites in elementary and secondary schools. In the literature different studies has appeared proposing new methods for classification of sites, with the goal of increasing the rate of pages correctly categorized. This work aims to contribute to the current methods of classification by comparing four aspects involved in the classification process: classification algorithms, dimensionality (amount of selected attributes), attributes evaluation metrics and selection of textual and structural attributes present in webpages. We use the vector model to treat text and an machine learning classical approach according to the classification task. Several metrics are used to make the selection of the most relevant terms, and classification algorithms from different paradigms are compared: probabilistic (Na¨ıve Bayes), decision tree (C4.5), instance-based learning (KNN - K-Nearest Neighbor) and support vector machine (SVM). The experiments were performed on a dataset containing two languages, English and Portuguese. The results show that it is possible to obtain a classifier with good success indexes using only the information from the anchor text in hyperlinks, in the experiments the classifier based on this information achieved 99.59% F-measure.Com a ampla utilização da web nos dias atuais e também com o seu crescimento constante, a tarefa de classificação automática de sítios web têm adquirido importância crescente, pois em diversas ocasiões é necessário bloquear o acesso a sítios específicos, como por exemplo no caso do acesso a sítios de conteúdo adulto em escolas elementares e secundárias. Na literatura diferentes trabalhos têm surgido propondo novos métodos de classificação de sítios, com o objetivo de aumentar o índice de páginas corretamente categorizadas. Este trabalho tem por objetivo contribuir com os métodos atuais de classificação através de comparações de quatro aspectos envolvidos no processo de classificação: algoritmos de classificação, dimensionalidade (número de atributos considerados), métricas de avaliação de atributos e seleção de atributos textuais e estruturais presentes nas páginas web. Utiliza-se o modelo vetorial para o tratamento de textos e uma abordagem de aprendizagem de máquina clássica considerando a tarefa de classificação. Diversas métricas são utilizadas para fazer a seleção dos termos mais relevantes, e algoritmos de classificação de diferentes paradigmas são comparados: probabilista (Naıve Bayes), árvores de decisão (C4.5), aprendizado baseado em instâncias (KNN - K vizinhos mais próximos) e Máquinas de Vetores de Suporte (SVM). Os experimentos foram realizados em um conjunto de dados contendo sítios de dois idiomas, Português e Inglês. Os resultados demonstram que é possível obter um classificador com bons índices de acerto utilizando apenas as informações do texto ˆancora dos hyperlinks. Nos experimentos o classificador baseado nessas informações atingiu uma Medida-F de 99.59%

    Segmentação de pele em imagens digitais para a detecção automática de conteúdo ofensivo

    Get PDF
    This work presents a study of suitable approaches for automatic detection of offensive content (pornography) in digital images. Extensive experiments were conducted for skin pixel segmentation, colour spaces and content descriptors. This work focus its efforts on skin pixel segmentation, since this segmentation is the pre-processing stage for almost every content-based offensive image classification methods in the literature. Four skin skin segmentation methods were tested in six colour spaces. Also, a structured image database was built to help improve studies in skin segmentation, with the possibility of adding meta-information to the images in the database, such as illumination conditions and camera standards. With the help of meta information from the image database, experimets involving illumination conditions and skin colour segmentation were also done. Finally, some feature extraction algorithms were implemented in order to apply content-based image retrieval (CBIR) algorithms to classify offensive images.CAPES; UOLO presente trabalho tem como objetivo estudar meios de efetuar a detecção automática de conteúdo ofensivo (pornografia) em imagens digitais. Para tal estudou-se largamente segmentação de pixels de pele, espaços de cor e descritores de conteúdo. Esse trabalho tem um foco maior na segmentação de pele, pois é a etapa primordial nos trabalhos envolvendo detecção de conteúdo ofensivo. Testou-se quatro métodos de segmentação de pixels de pele e foi construído um banco de dados estruturado para o estudo de segmentação de pele, com meios de anotação de imagens para auxiliar na estruturação e no controle das características das imagens do banco. Com o auxílio das metainformações do banco de imagens, foram conduzidos estudos envolvendo as condições de iluminação e a segmentação de pele. Por fim, foi implementado um algoritmo de extração de características em sistemas de classificação pelo conteúdo de imagens (CBIR) para detecção de conteúdo ofensivo

    Advances in knowledge discovery and data mining Part II

    Get PDF
    19th Pacific-Asia Conference, PAKDD 2015, Ho Chi Minh City, Vietnam, May 19-22, 2015, Proceedings, Part II</p

    Cultural governance in contemporary China: popular culture, digital technology, and the state

    Get PDF
    This dissertation is a study of the historical formation and transformation of the Chinese online audiovisual industry under forces of strategic political calculations, expanding market relations, and growing social participation, and the cultural ramifications of this process, especially the kind of transformations digital technologies have wrought on the state-TV-station-centered mode of cultural production/distribution and regulatory apparatuses. Through this case, the project aims to theorize the changing mode of cultural governance of post-socialist regimes in the context of digital capitalism. Using mixed methods of documentary research, interviews with industry practitioners, participant observations of trade fairs/festivals, and critical discourse analyses of popular cultural texts, the study finds that the traditional broadcasting and the online video sectors are structured along two different political economic mechanisms. While the former is dominated by domestic capital and heavily regulated by state agencies, the latter is supported by transnational capital and less regulated. Digital technologies coupled with transnational capital thus generate new cultural flows, processes, and practices, which produces a heterogeneous and contested cultural sphere in the digital environment that substantially differs from the one created by traditional television. The development of such a sphere in a cultural environment that was historically policed suggests that the Chinese state strategically configures the cultural realm into multiple zones delineated by technological forms. Cultural zoning allows the state to accommodate needs in relation to transnational forces while simultaneously retaining socialist legacies through state media. Zoning technology demonstrates flexibility in cultural governance and thus illuminates the extraordinary resilience of post-socialist regimes amid neoliberal globalization

    2019 Issue

    Get PDF
    corecore