87 research outputs found

    Using video objects and relevance feedback in video retrieval

    Get PDF
    Video retrieval is mostly based on using text from dialogue and this remains the most signiยฏcant component, despite progress in other aspects. One problem with this is when a searcher wants to locate video based on what is appearing in the video rather than what is being spoken about. Alternatives such as automatically-detected features and image-based keyframe matching can be used, though these still need further improvement in quality. One other modality for video retrieval is based on segmenting objects from video and allowing end users to use these as part of querying. This uses similarity between query objects and objects from video, and in theory allows retrieval based on what is actually appearing on-screen. The main hurdles to greater use of this are the overhead of object segmentation on large amounts of video and the issue of whether we can actually achieve effective object-based retrieval. We describe a system to support object-based video retrieval where a user selects example video objects as part of the query. During a search a user builds up a set of these which are matched against objects previously segmented from a video library. This match is based on MPEG-7 Dominant Colour, Shape Compaction and Texture Browsing descriptors. We use a user-driven semi-automated segmentation process to segment the video archive which is very accurate and is faster than conventional video annotation

    Content-Based Image Retrieval Using Associative Memories

    Get PDF
    The rapid growth in the number of large-scale repositories has brought the need for efficient and effective content-based image retrieval (CBIR) systems. The state of the art in the CBIR systems is to search images in database that are โ€œcloseโ€ to the query image using some similarity measure. The current CBIR systems capture image features that represent properties such as color, texture, and/or shape of the objects in the query image and try to retrieve images from the database with similar features. In this paper, we propose a new architecture for a CBIR system. We try to mimic the human memory. We use generalized bi-directional associative memory (BAMg) to store and retrieve images from the database. We store and retrieve images based on association. We present three topologies of the generalized bi-directional associative memory that are similar to the local area network topologies: the bus, ring, and tree. We have developed software to implement the CBIR system. As an illustration, we have considered three sets of images. The results of our simulation are presented in the paper

    Shape-Based Tumor Retrieval in Mammograms Using Relevance-Feedback Techniques

    Get PDF
    Abstract. This paper presents an experimental "morphological analysis" retrieval system for mammograms, using Relevance-Feedback techniques. The features adopted are first-order statistics of the Normalized Radial Distance, extracted from the annotated mass boundary. The system is evaluated on an extensive dataset of 2274 masses of the DDSM database, which involves 7 distinct classes. The experiments verify that the involvement of the radiologist as part of the retrieval process improves the results, even for such a hard classification task, reaching the precision rate of almost 90%. Therefore, Relevance-Feedback can be employed as a very useful complementary tool to a Computer Aided Diagnosis system

    SPNet: Deep 3D Object Classification and Retrieval using Stereographic Projection

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(์„์‚ฌ)--์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› :๊ณต๊ณผ๋Œ€ํ•™ ์ „๊ธฐยท์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€,2019. 8. ์ด๊ฒฝ๋ฌด.๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” 3D ๋ฌผ์ฒด๋ถ„๋ฅ˜ ๋ฌธ์ œ๋ฅผ ํšจ์œจ์ ์œผ๋กœ ํ•ด๊ฒฐํ•˜๊ธฐ์œ„ํ•˜์—ฌ ์ž…์ฒดํ™”๋ฒ•์˜ ํˆฌ์‚ฌ๋ฅผ ํ™œ์šฉํ•œ ๋ชจ๋ธ์„ ์ œ์•ˆํ•œ๋‹ค. ๋จผ์ € ์ž…์ฒดํ™”๋ฒ•์˜ ํˆฌ์‚ฌ๋ฅผ ์‚ฌ์šฉํ•˜์—ฌ 3D ์ž…๋ ฅ ์˜์ƒ์„ 2D ํ‰๋ฉด ์ด๋ฏธ์ง€๋กœ ๋ณ€ํ™˜ํ•œ๋‹ค. ๋˜ํ•œ, ๊ฐ์ฒด์˜ ์นดํ…Œ๊ณ ๋ฆฌ๋ฅผ ์ถ”์ •ํ•˜๊ธฐ ์œ„ํ•˜์—ฌ ์–•์€ 2Dํ•ฉ์„ฑ๊ณฑ์‹ ์…ฉ๋ง(CNN)์„ ์ œ์‹œํ•˜๊ณ , ๋‹ค์ค‘์‹œ์ ์œผ๋กœ๋ถ€ํ„ฐ ์–ป์€ ๊ฐ์ฒด ์นดํ…Œ๊ณ ๋ฆฌ์˜ ์ถ”์ •๊ฐ’๋“ค์„ ๊ฒฐํ•ฉํ•˜์—ฌ ์„ฑ๋Šฅ์„ ๋”์šฑ ํ–ฅ์ƒ์‹œํ‚ค๋Š” ์•™์ƒ๋ธ” ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•œ๋‹ค. ์ด๋ฅผ์œ„ํ•ด (1) ์ž…์ฒดํ™”๋ฒ•ํˆฌ์‚ฌ๋ฅผ ํ™œ์šฉํ•˜์—ฌ 3D ๊ฐ์ฒด๋ฅผ 2D ํ‰๋ฉด ์ด๋ฏธ์ง€๋กœ ๋ณ€ํ™˜ํ•˜๊ณ  (2) ๋‹ค์ค‘์‹œ์  ์˜์ƒ๋“ค์˜ ํŠน์ง•์ ์„ ํ•™์Šต (3) ํšจ๊ณผ์ ์ด๊ณ  ๊ฐ•์ธํ•œ ์‹œ์ ์˜ ํŠน์ง•์ ์„ ์„ ๋ณ„ํ•œ ํ›„ (4) ๋‹ค์ค‘์‹œ์  ์•™์ƒ๋ธ”์„ ํ†ตํ•œ ์„ฑ๋Šฅ์„ ํ–ฅ์ƒ์‹œํ‚ค๋Š” 4๋‹จ๊ณ„๋กœ ๊ตฌ์„ฑ๋œ ํ•™์Šต๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•œ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์‹คํ—˜๊ฒฐ๊ณผ๋ฅผ ํ†ตํ•ด ์ œ์•ˆํ•˜๋Š” ๋ฐฉ๋ฒ•์ด ๋งค์šฐ ์ ์€ ๋ชจ๋ธ์˜ ํ•™์Šต ๋ณ€์ˆ˜์™€ GPU ๋ฉ”๋ชจ๋ฆฌ๋ฅผ ์‚ฌ์šฉํ•˜๋Š”๊ณผ ๋™์‹œ์— ๊ฐ์ฒด ๋ถ„๋ฅ˜ ๋ฐ ๊ฒ€์ƒ‰์—์„œ์˜ ์šฐ์ˆ˜ํ•œ ์„ฑ๋Šฅ์„ ๋ณด์ด๊ณ ์žˆ์Œ์„ ์ฆ๋ช…ํ•˜์˜€๋‹ค.We propose an efficient Stereographic Projection Neural Network (SPNet) for learning representations of 3D objects. We first transform a 3D input volume into a 2D planar image using stereographic projection. We then present a shallow 2D convolutional neural network (CNN) to estimate the object category followed by view ensemble, which combines the responses from multiple views of the object to further enhance the predictions. Specifically, the proposed approach consists of four stages: (1) Stereographic projection of a 3D object, (2) view-specific feature learning, (3) view selection and (4) view ensemble. The proposed approach performs comparably to the state-of-the-art methods while having substantially lower GPU memory as well as network parameters. Despite its lightness, the experiments on 3D object classification and shape retrievals demonstrate the high performance of the proposed method.1 INTRODUCTION 2 Related Work 2.1 Point cloud-based methods 2.2 3D model-based methods 2.3 2D/2.5D image-based methods 3 Proposed Stereographic Projection Network 3.1 Stereographic Representation 3.2 Network Architecture 3.3 View Selection 3.4 View Ensemble 4 Experimental Evaluation 4.1 Datasets 4.2 Training 4.3 Choice of Stereographic Projection 4.4 Test on View Selection Schemes 4.5 3D Object Classification 4.6 Shape Retrieval 4.7 Implementation 5 ConclusionsMaste

    Novel hybrid generative adversarial network for synthesizing image from sketch

    Get PDF
    In the area of sketch-based image retrieval process, there is a potential difference between retrieving the match images from defined dataset and constructing the synthesized image. The former process is quite easier while the latter process requires more faster, accurate, and intellectual decision making by the processor. After reviewing open-end research problems from existing approaches, the proposed scheme introduces a computational framework of hybrid generative adversarial network (GAN) as a solution to address the identified research problem. The model takes the input of query image which is processed by generator module running 3 different deep learning modes of ResNet, MobileNet, and U-Net. The discriminator module processes the input of real images as well as output from generator. With a novel interactive communication between generator and discriminator, the proposed model offers optimal retrieval performance along with an inclusion of optimizer. The study outcome shows significant performance improvement
    • โ€ฆ
    corecore