6 research outputs found

    Video Data Visualization System: Semantic Classification And Personalization

    Full text link
    We present in this paper an intelligent video data visualization tool, based on semantic classification, for retrieving and exploring a large scale corpus of videos. Our work is based on semantic classification resulting from semantic analysis of video. The obtained classes will be projected in the visualization space. The graph is represented by nodes and edges, the nodes are the keyframes of video documents and the edges are the relation between documents and the classes of documents. Finally, we construct the user's profile, based on the interaction with the system, to render the system more adequate to its references.Comment: graphic

    Adaptive image segmentation by combining photometric invariant region and edge information

    No full text

    Two and three dimensional segmentation of multimodal imagery

    Get PDF
    The role of segmentation in the realms of image understanding/analysis, computer vision, pattern recognition, remote sensing and medical imaging in recent years has been significantly augmented due to accelerated scientific advances made in the acquisition of image data. This low-level analysis protocol is critical to numerous applications, with the primary goal of expediting and improving the effectiveness of subsequent high-level operations by providing a condensed and pertinent representation of image information. In this research, we propose a novel unsupervised segmentation framework for facilitating meaningful segregation of 2-D/3-D image data across multiple modalities (color, remote-sensing and biomedical imaging) into non-overlapping partitions using several spatial-spectral attributes. Initially, our framework exploits the information obtained from detecting edges inherent in the data. To this effect, by using a vector gradient detection technique, pixels without edges are grouped and individually labeled to partition some initial portion of the input image content. Pixels that contain higher gradient densities are included by the dynamic generation of segments as the algorithm progresses to generate an initial region map. Subsequently, texture modeling is performed and the obtained gradient, texture and intensity information along with the aforementioned initial partition map are used to perform a multivariate refinement procedure, to fuse groups with similar characteristics yielding the final output segmentation. Experimental results obtained in comparison to published/state-of the-art segmentation techniques for color as well as multi/hyperspectral imagery, demonstrate the advantages of the proposed method. Furthermore, for the purpose of achieving improved computational efficiency we propose an extension of the aforestated methodology in a multi-resolution framework, demonstrated on color images. Finally, this research also encompasses a 3-D extension of the aforementioned algorithm demonstrated on medical (Magnetic Resonance Imaging / Computed Tomography) volumes

    Shadow segmentation and tracking in real-world conditions

    Get PDF
    Visual information, in the form of images and video, comes from the interaction of light with objects. Illumination is a fundamental element of visual information. Detecting and interpreting illumination effects is part of our everyday life visual experience. Shading for instance allows us to perceive the three-dimensional nature of objects. Shadows are particularly salient cues for inferring depth information. However, we do not make any conscious or unconscious effort to avoid them as if they were an obstacle when we walk around. Moreover, when humans are asked to describe a picture, they generally omit the presence of illumination effects, such as shadows, shading, and highlights, to give a list of objects and their relative position in the scene. Processing visual information in a way that is close to what the human visual system does, thus being aware of illumination effects, represents a challenging task for computer vision systems. Illumination phenomena interfere in fact with fundamental tasks in image analysis and interpretation applications, such as object extraction and description. On the other hand, illumination conditions are an important element to be considered when creating new and richer visual content that combines objects from different sources, both natural and synthetic. When taken into account, illumination effects can play an important role in achieving realism. Among illumination effects, shadows are often integral part of natural scenes and one of the elements contributing to naturalness of synthetic scenes. In this thesis, the problem of extracting shadows from digital images is discussed. A new analysis method for the segmentation of cast shadows in still and moving images without the need of human supervision is proposed. The problem of separating moving cast shadows from moving objects in image sequences is particularly relevant for an always wider range of applications, ranging from video analysis to video coding, and from video manipulation to interactive environments. Therefore, particular attention has been dedicated to the segmentation of shadows in video. The validity of the proposed approach is however also demonstrated through its application to the detection of cast shadows in still color images. Shadows are a difficult phenomenon to model. Their appearance changes with changes in the appearance of the surface they are cast upon. It is therefore important to exploit multiple constraints derived from the analysis of the spectral, geometric and temporal properties of shadows to develop effective techniques for their extraction. The proposed method combines an analysis of color information and of photometric invariant features to a spatio-temporal verification process. With regards to the use of color information for shadow analysis, a complete picture of the existing solutions is provided, which points out the fundamental assumptions, the adopted color models and the link with research problems such as computational color constancy and color invariance. The proposed spatial verification does not make any assumption about scene geometry nor about object shape. The temporal analysis is based on a novel shadow tracking technique. On the basis of the tracking results, a temporal reliability estimation of shadows is proposed which allows to discard shadows which do not present time coherence. The proposed approach is general and can be applied to a wide class of applications and input data. The proposed cast shadow segmentation method has been evaluated on a number of different video data representing indoor and outdoor real-world environments. The obtained results have confirmed the validity of the approach, in particular its ability to deal with different types of content and its robustness to different physically important independent variables, and have demonstrated the improvement with respect to the state of the art. Examples of application of the proposed shadow segmentation tool to the enhancement of video object segmentation, tracking and description operations, and to video composition, have demonstrated the advantages of a shadow-aware video processing

    Navigation visuelle d'un robot mobile dans un environnement d'extérieur semi-structuré

    Get PDF
    Cette thèse porte sur le traitement automatique d'images couleur, et son application à la robotique dans des environnements semi-structurés d'extérieur. Nous proposons une méthode de navigation visuelle pour des robots mobiles en utilisant une caméra couleur. Les domaines d'application de ce travail se situent dans l'automatisation de machines agricoles, en vue de la navigation automatique dans un réseau de chemins (pour aller d'une ferme à un champ par exemple). Nous présentons tout d'abord une analyse des principaux travaux de recherche dans la littérature sur la navigation visuelle. Une chaîne de pré-traitement pour le rendu couleur d'images numériques mono-capteur dotées d'un filtre Bayer est présentée ; elle se base sur une étude des techniques de démosaïquage, le calibrage chromatique d'images (balance de blancs) et la correction gamma. Une méthode d'interprétation monoculaire de la scène courante permet d'extraire les régions navigables et un modèle 2D de la scène. Nous traitons de la segmentation d'une image couleur en régions, puis de la caractérisation de ces régions par des attributs de texture et de couleur, et enfin, de l'identification des diverses entités de la scène courante (chemin, herbe, arbre, ciel, champ labouré,. . . ). Pour cela, nous exploitons deux méthodes de classification supervisée : la méthode de Support Vector Machine (SVM) et celle des k plus proches voisins (k-PPV). Une réduction d'information redondante par une analyse en composantes indépendantes (ACI) a permis d'améliorer le taux global de reconnaissance. Dans un réseau de chemins, le robot doit reconnaître les intersections de chemins lui permettant (a) dans une phase d'apprentissage, de construire un modèle topologique du réseau dans lequel il va devoir se déplacer et (b) dans une phase de navigation, de planifier et exécuter une trajectoire topologique définie dans ce réseau. Nous proposons donc une méthode de détection et classification du chemin : ligne droite, virage gauche, virage droite, carrefour en X, en T ou en Y. Une approche pour la représentation de la forme et de la catégorisation des contours (Shape Context) est utilisée à cet effet. Une validation a été effectuée sur une base d'images de routes ou chemins de campagne. En exploitant cette méthode pour détecter et classifier les noeuds du réseau de chemins, un modèle topologique sous forme d'un graphe est construit ; la méthode est validée sur une séquence d'images de synthèse. Enfin, dans la dernière partie de la thèse, nous décrivons des résultats expérimentaux obtenus sur le démonstrateur Dala du groupe Robotique et IA du LAAS-CNRS. Le déplacement du robot est contrôlé et guidé par l'information fournie par le système de vision à travers des primitives de déplacement élémentaires (Suivi-Chemin, Suivi-Objet, Suivi-Bordure,. . . ). Le robot se place au milieu du chemin en construisant une trajectoire à partir du contour de cette région navigable. étant donné que le modèle sémantique de la scène est produit à basse fréquence (de 0,5 à 1 Hz) par le module de vision couleur, nous avons intégré avec celui-ci, un module de suivi temporel des bords du chemin (par Snakes), pour augmenter la fréquence d'envoi des consignes (de 5 à 10 Hz) au module de locomotion. Modules de vision couleur et de suivi temporel doivent être synchronisés de sorte que le suivi puisse être réinitialisé en cas de dérive. Après chaque détection du chemin, une trajectoire sur le sol est planifiée et exécutée ; les anciennes consignes qui ne sont pas encore exécutées sont fusionnées et filtrées avec les nouvelles, donnant de la stabilité au système. ABSTRACT : This thesis deals with the automatic processing of color images, and its application to robotics in outdoor semi-structured environments. We propose a visual-based navigation method for mobile robots by using an onboard color camera. The objective is the robotization of agricultural machines, in order to navigate automatically on a network of roads (to go from a farm to a given field). Firstly, we present an analysis of the main research works about visual-based navigation literature. A preprocessing chain for color rendering on mono-sensor digital images equipped with a Bayer filter, is presented ; it is based on the analysis of the demosaicking techniques, the chromatic calibration of images (white point balance) and the correction gamma. Our monocular scene interpretation method makes possible to extract the navigable regions and a basic 2D scene modeling. We propose functions for the segmentation of the color images, then for the characterization of the extracted regions by texture and color attributes, and at last, for their classification in order to recognize the road and other entities of the current scene (grass, trees, clouds, hedges, fields,. . . ). Thus, we use two supervised classification methods : Support Vector Machines (SVM) and k nearest neighbors (k-NN). A redundancy reduction by using independent components analysis (ICA) was performed in order to improve the overall recognition rate. In a road network, the robot needs to recognize the roads intersections in order to navigate and to build a topological model from its trajectory. An approach for the road classification is proposed to recognize : straight ahead, turn-left, turn-right, road intersections and road bifurcations. An approach based on the road shape representation and categorization (shape context) is used for this purpose. A validation was carried out on an image dataset of roads or country lanes. By exploiting this method to detect and classify the nodes of a road network, a topological model based on a graph is built ; the method is validated on a sequence of synthetic images. Finally, Robot displacement is controlled and guided by the information provided by the vision system through elementary displacement primitives (Road-Follow, Follow-Object, Follow-Border,. . . ). Robot Dala is placed in the middle of the road by computing a trajectory obtained from the navigable region contours. As retrieving semantic information from vision is computationally demanding (low frequency 0.5 ¼ 1 Hz), a Snakes tracking process was implemented to speed up the transfer of instructions (5 ¼ 10 Hz) to the locomotion module. Both tasks must be synchronized, so the tracking can be re-initialized if a failure is detected. Locomotion tasks are planned and carried out while waiting for new data from the vision module ; the instructions which are not yet carried out, are merged and filtered with the new ones, which provides stability to the system
    corecore