9 research outputs found

    Real-time shot detection based on motion analysis and multiple low-level techniques

    Full text link
    To index, search, browse and retrieve relevant material, indexes describing the video content are required. Here, a new and fast strategy which allows detecting abrupt and gradual transitions is proposed. A pixel-based analysis is applied to detect abrupt transitions and, in parallel, an edge-based analysis is used to detect gradual transitions. Both analysis are reinforced with a motion analysis in a second step, which significantly simplifies the threshold selection problem while preserving the computational requirements. The main advantage of the proposed system is its ability to work in real time and the experimental results show high recall and precision values

    Navigation framework using visual landmarks and a GIS

    Get PDF
    In an unfamiliar environment we spot and explore all available information which might guide us to a desired location. This largely unconscious processing is done by our trained sensory and cognitive systems. These recognise and memorise sets of landmarks which allow us to create a mental map of the environment, and this map enables us to navigate by exploiting very few but the most important landmarks stored in our memory. In this paper we present a route planning, localisation and navigation system which works in real time. It integrates a geographic information system of a building with visual landmarks for localising the user and for validating the navigation route. Although designed for visually impaired persons, the system can also be employed to assist or transport persons with reduced mobility in way finding in a complex building. © 2013 The Authors. Published by Elsevier B.V

    Temporal segmentation tool for high-quality real time video editing software

    Get PDF
    The increasing use of video editing software requires faster and more efficient editing tools. As a first step, these tools perform a temporal segmentation in shots that allows a later building of indexes describing the video content. Here, we propose a novel real-time high-quality shot detection strategy, suitable for the last generation of video editing software requiring both low computational cost and high quality results. While abrupt transitions are detected through a very fast pixel-based analysis, gradual transitions are obtained from an efficient edge-based analysis. Both analyses are reinforced with a motion analysis that helps to detect and discard false detections. This motion analysis is carried out exclusively over a reduced set of candidate transitions, thus maintaining the computational requirements demanded by new applications to fulfill user needs

    Live Video and Image Recolouring for Colour Vision Deficient Patients

    Get PDF
    Colour Vision Deficiency (CVD) is an important issue for a significant population across the globe. There are several types of CVD\u27s, such as monochromacy, dichromacy, trichromacy, and anomalous trichromacy. Each of these categories contain specific other subtypes. The aim of this research is to device a scheme to address CVD by using variations in pixel plotting of colours to capture colour disparities and perform colour compensation. The proposed scheme recolours the video and images by colour contrast variation of each colour for CVD patients, and depending on the type of deficiency, it is able to provide live results. Different types of CVD’s can be identified and cured by changing the particular colour related to it and based upon the type of diseases, it performs RGB (Red, Green, and Blue) to LMS (Long, Medium, and Short) transformation. This helps in colour identification and also adjustments of colour contrasts. The processing and rendering of recoloured video and images, allows the affected patients with CVD to see perfect shades in the recoloured frames of video or images and other modes of files. In this thesis, we propose an efficient recolouring algorithm with a strong focus on real-time applications that is capable of providing different recoloured outputs based on specific types of CVD

    Sistema de visão para a orientação e mobilidade em edifícios públicos

    Get PDF
    Dissertação de mest., Engenharia Elétrica e Eletrónica (Tecnologias da Informação e Telecomunicações), Instituto Superior de Engenharia, Univ. do Algarve, 2013Os sistemas de navegação, como o GPS, têm proliferado de forma surpreendente na última década, assumindo hoje um papel fundamental no nosso quotidiano. No entanto, tais tecnologias são de uso exclusivo em campo aberto e inúteis na utilização em ambientes interiores. Esta dissertação foca o desenvolvimento de um sistema para orientação e mobilidade dentro de edifícios públicos utilizando visão artificial para o reconhecimento de objetos e textos que sirvam de pontos de referência para a localização do indivíduo, tais como sinalizadores, placares de informação textual, extintores de incêndio, portas, escadas, janelas, cacifos, retângulos entre outros e um Sistema de Informação Geográfica para efetuar o enquadramento espacial da informação extraída por visão. A utilização de o Sistema de Informação Geográfica permite ainda o planeamento de uma rota de navegação através da verificação das acessibilidades entre as divisões do edifício caraterizado em mapa. Torna-se assim possível, conhecendo a posição e rota de navegação, orientar o indivíduo no seu trajeto até ao seu destino final. O interface com o utilizador, visando como utilizador final pessoas com deficiência visual, é efetuado com recurso a um sintetizador de voz que comunica verbalmente as informações tanto de reconhecimento como de localização e orientação

    Large Scale Pattern Detection in Videos and Images from the Wild

    Get PDF
    PhDPattern detection is a well-studied area of computer vision, but still current methods are unstable in images of poor quality. This thesis describes improvements over contemporary methods in the fast detection of unseen patterns in a large corpus of videos that vary tremendously in colour and texture definition, captured “in the wild” by mobile devices and surveillance cameras. We focus on three key areas of this broad subject; First, we identify consistency weaknesses in existing techniques of processing an image and it’s horizontally reflected (mirror) image. This is important in police investigations where subjects change their appearance to try to avoid recognition, and we propose that invariance to horizontal reflection should be more widely considered in image description and recognition tasks too. We observe online Deep Learning system behaviours in this respect, and provide a comprehensive assessment of 10 popular low level feature detectors. Second, we develop simple and fast algorithms that combine to provide memory- and processing-efficient feature matching. These involve static scene elimination in the presence of noise and on-screen time indicators, a blur-sensitive feature detection that finds a greater number of corresponding features in images of varying sharpness, and a combinatorial texture and colour feature matching algorithm that matches features when either attribute may be poorly defined. A comprehensive evaluation is given, showing some improvements over existing feature correspondence methods. Finally, we study random decision forests for pattern detection. A new method of indexing patterns in video sequences is devised and evaluated. We automatically label positive and negative image training data, reducing a task of unsupervised learning to one of supervised learning, and devise a node split function that is invariant to mirror reflection and rotation through 90 degree angles. A high dimensional vote accumulator encodes the hypothesis support, yielding implicit back-projection for pattern detection.European Union’s Seventh Framework Programme, specific topic “framework and tools for (semi-) automated exploitation of massive amounts of digital data for forensic purposes”, under grant agreement number 607480 (LASIE IP project)

    Object Duplicate Detection

    Get PDF
    With the technological evolution of digital acquisition and storage technologies, millions of images and video sequences are captured every day and shared in online services. One way of exploring this huge volume of images and videos is through searching a particular object depicted in images or videos by making use of object duplicate detection. Therefore, need of research on object duplicate detection is validated by several image and video retrieval applications, such as tag propagation, augmented reality, surveillance, mobile visual search, and television statistic measurement. Object duplicate detection is detecting visually same or very similar object to a query. Input is not restricted to an image, it can be several images from an object or even it can be a video. This dissertation describes the author's contribution to solve problems on object duplicate detection in computer vision. A novel graph-based approach is introduced for 2D and 3D object duplicate detection in still images. Graph model is used to represent the 3D spatial information of the object based on the local features extracted from training images so that an explicit and complex 3D object modeling is avoided. Therefore, improved performance can be achieved in comparison to existing methods in terms of both robustness and computational complexity. Our method is shown to be robust in detecting the same objects even when images containing the objects are taken from very different viewpoints or distances. Furthermore, we apply our object duplicate detection method to video, where the training images are added iteratively to the video sequence in order to compensate for 3D view variations, illumination changes and partial occlusions. Finally, we show several mobile applications for object duplicate detection, such as object recognition based museum guide, money recognition or flower recognition. General object duplicate detection may fail to detection chess figures, however considering context, like chess board position and height of the chess figure, detection can be more accurate. We show that user interaction further improves image retrieval compared to pure content-based methods through a game, called Epitome
    corecore