645 research outputs found

    Processing Large Amounts of Images on Hadoop with OpenCV

    Full text link
    Modern image collections cannot be processed efficiently on one computer due to large collection sizes and high computational costs of modern image processing algorithms. Hence, image processing often requires distributed computing. However, distributed computing is a complicated subject that demands deep technical knowledge and often cannot be used by researches who develop image processing algorithms. The framework is needed that allows the researches to concentrate on image processing tasks and hides from them the complicated details of distributed computing. In addition, the framework should provide the researches with the familiar image processing tools. The paper describes the extension to the MapReduce Image Processing (MIPr) framework that provides the ability to use OpenCV in Hadoop cluster for distributed image processing. The modified MIPr framework allows the development of image processing programs in Java using the OpenCV Java binding. The performance testing of created system on the cloud cluster demonstrated near-linear scalability

    Outfit Recommender System

    Get PDF
    The online apparel retail market size in the United States is worth about seventy-two billion US dollars. Recommendation systems on retail websites generate a lot of this revenue. Thus, improving recommendation systems can increase their revenue. Traditional recommendations for clothes consisted of lexical methods. However, visual-based recommendations have gained popularity over the past few years. This involves processing a multitude of images using different image processing techniques. In order to handle such a vast quantity of images, deep neural networks have been used extensively. With the help of fast Graphics Processing Units, these networks provide results which are extremely accurate, within a small amount of time. However, there are still ways in which recommendations for clothes can be improved. We propose an event-based clothing recommendation system which uses object detection. We train a model to identify nine events/scenarios that a user might attend: White Wedding, Indian Wedding, Conference, Funeral, Red Carpet, Pool Party, Birthday, Graduation and Workout. We train another model to detect clothes out of fifty-three categories of clothes worn at the event. Object detection gives a mAP of 84.01. Nearest neighbors of the clothes detected are recommended to the user

    Design Human Object Detection Yolov4-Tiny Algorithm on ARM Cortex-A72 and A53

    Get PDF
    Currently, many object detection systems still use devices with large sizes, such as using PCs, as supporting devices, for object detection. This makes these devices challenging to use as a security system in public facilities based on human object detection. In contrast, many Mini PCs currently use ARM processors with high specifications. In this research, to detect human objects will use the Mini PC Nanopi M4V2 device that has a speed in processing with the support of CPU Dual-Core Cortex-A72 (up to 2.0 GHz) + Cortex A53 (Up to 2.0 GHz) and 4 Gb DDR4 Ram. In addition, for the human object detection system, the author uses the You Only Look Once (YOLO) method with the YoloV4-Tiny type, With these specifications and methods, the detection rate and FPS score are seen which are the feasibility values for use in detecting human objects. The simulation for human object recognition was carried out using recorded video, simulation obtained a detection rate of 0,9845 or 98% with FPS score of 3.81-5.55.  These results are the best when compared with the YOLOV4 and YOLOV5 models. With these results, it can be applied in various human detection applications and of course robustness testing is needed

    CSISE: cloud-based semantic image search engine

    Get PDF
    Title from PDF of title page, viewed on March 27, 2014Thesis advisor: Yugyung LeeVitaIncludes bibliographical references (pages 53-56)Thesis (M. S.)--School of Computing and Engineering. University of Missouri--Kansas City, 2013Due to rapid exponential growth in data, a couple of challenges we face today are how to handle big data and analyze large data sets. An IBM study showed the amount of data created in the last two years alone is 90% of the data in the world today. We have especially seen the exponential growth of images on the Web, e.g., more than 6 billion in Flickr, 1.5 billion in Google image engine, and more than 1 billon images in Instagram [1]. Since big data are not only a matter of a size, but are also heterogeneous types and sources of data, image searching with big data may not be scalable in practical settings. We envision Cloud computing as a new way to transform the big data challenge into a great opportunity. In this thesis, we intend to perform an efficient and accurate classification of a large collection of images using Cloud computing, which in turn supports semantic image searching. A novel approach with enhanced accuracy has been proposed to utilize semantic technology to classify images by analyzing both metadata and image data types. A two-level classification model was designed (i) semantic classification was performed on a metadata of images using TF-IDF, and (ii) image classification was performed using a hybrid image processing model combined with Euclidean distance and SURF FLANN measurements. A Cloud-based Semantic Image Search Engine (CSISE) is also developed to search an image using the proposed semantic model with the dynamic image repository by connecting online image search engines that include Google Image Search, Flickr, and Picasa. A series of experiments have been performed in a large-scale Hadoop environment using IBM's cloud on over half a million logo images of 76 types. The experimental results show that the performance of the CSISE engine (based on the proposed method) is comparable to the popular online image search engines as well as accurate with a higher rate (average precision of 71%) than existing approachesAbstract -- Contents -- Illustrations -- Tables -- Acknowledgements - Introduction -- Related work -- Cloud-based semantic image search engine model -- Cloud-based semantic image search engine (CSISE) implementation -- Experimental results and evaluation -- Conclusion and future work - Reference

    Real-time near replica detection over massive streams of shared photos

    Get PDF
    Aquest treball es basa en la detecció en temps real de repliques d'imatges en entorns distribuïts a partir de la indexació de vectors de característiques locals

    Security Camera Implementation System on Smart Home Using Web-Based Frame Difference Methods

    Get PDF
    The development of technology in the field of security has a big impact on people's lives throughout the world. The security is threatened when there is a burglary in the neighborhood. One of the applications of image processing is the frame difference method, which used comparison of the captured image frame circuit according to time order. Frame difference algorithm applied to grayscale image computing using ESP32-CAM cameras for security cameras. This research was developed with a system that can monitor objects detected in real-time by streaming using a website. The ESP32-CAM camera and the PC have a packet loss value of 4.29% and an average delay of 0.3 seconds so that the images captured by the camera can be well received by the PC. The frame difference method with moving objects is used and detected well during the day due to solar lighting, but at night only depends on luminous objects. This method has the disadvantage that when moving objects inhabit the same field of view, the system will define it as movement so that the frame difference method is not adaptive from changes in the captured area

    Interactive detection of incrementally learned concepts in images with ranking and semantic query interpretation

    Get PDF
    This research was performed in the GOOSE project, which is jointly funded by the MIST research program of the Dutch Ministry of Defense and the AMSN enabling technology program.The number of networked cameras is growing exponentially. Multiple applications in different domains result in an increasing need to search semantically over video sensor data. In this paper, we present the GOOSE demonstrator, which is a real-time general-purpose search engine that allows users to pose natural language queries to retrieve corresponding images. Top-down, this demonstrator interprets queries, which are presented as an intuitive graph to collect user feedback. Bottomup, the system automatically recognizes and localizes concepts in images and it can incrementally learn novel concepts. A smart ranking combines both and allows effective retrieval of relevant images.peer-reviewe
    corecore