172 research outputs found

    View recommendation for multi-camera demonstration-based training

    Get PDF
    While humans can effortlessly pick a view from multiple streams, automatically choosing the best view is a challenge. Choosing the best view from multi-camera streams poses a problem regarding which objective metrics should be considered. Existing works on view selection lack consensus about which metrics should be considered to select the best view. The literature on view selection describes diverse possible metrics. And strategies such as information-theoretic, instructional design, or aesthetics-motivated fail to incorporate all approaches. In this work, we postulate a strategy incorporating information-theoretic and instructional design-based objective metrics to select the best view from a set of views. Traditionally, information-theoretic measures have been used to find the goodness of a view, such as in 3D rendering. We adapted a similar measure known as the viewpoint entropy for real-world 2D images. Additionally, we incorporated similarity penalization to get a more accurate measure of the entropy of a view, which is one of the metrics for the best view selection. Since the choice of the best view is domain-dependent, we chose demonstration-based training scenarios as our use case. The limitation of our chosen scenarios is that they do not include collaborative training and solely feature a single trainer. To incorporate instructional design considerations, we included the trainer’s body pose, face, face when instructing, and hands visibility as metrics. To incorporate domain knowledge we included predetermined regions’ visibility as another metric. All of those metrics are taken into account to produce a parameterized view recommendation approach for demonstration-based training. An online study using recorded multi-camera video streams from a simulation environment was used to validate those metrics. Furthermore, the responses from the online study were used to optimize the view recommendation performance with a normalized discounted cumulative gain (NDCG) value of 0.912, which shows good performance with respect to matching user choices

    A comparative study of algorithms for automatic segmentation of dermoscopic images

    Get PDF
    Melanoma is the most common as well as the most dangerous type of skin cancer. Nevertheless, it can be effectively treated if detected early. Dermoscopy is one of the major non-invasive imaging techniques for the diagnosis of skin lesions. The computer-aided diagnosis based on the processing of dermoscopic images aims to reduce the subjectivity and time-consuming analysis related to traditional diagnosis. The first step of automatic diagnosis is image segmentation. In this project, the implementation and evaluation of several methods were proposed for the automatic segmentation of lesion regions in dermoscopic images, along with the corresponding implemented phases for image preprocessing and postprocessing. The developed algorithms include methods based on different state of the art techniques. The main groups of techniques which have been selected to be studied and implemented are thresholding-based methods, region-based methods, segmentation based on deformable models, as well as a new proposed approach based on the bag-of-words model. The implemented methods incorporate modifications for a better adaptation to features associated with dermoscopic images. Each implemented method was applied to a database constituted by 724 dermoscopic images. The output of the automatic segmentation procedure for each image was compared with the corresponding manual segmentation in order to evaluate the performance. The comparison between algorithms was carried out regarding the obtained evaluation metrics. The best results were achieved by the combination of region-based segmentation based on the multi-region adaptation of the k-means algorithm and the subIngeniería de Sistemas Audiovisuale

    Deep learning for digitized histology image analysis

    Get PDF
    “Cervical cancer is the fourth most frequent cancer that affects women worldwide. Assessment of cervical intraepithelial neoplasia (CIN) through histopathology remains as the standard for absolute determination of cancer. The examination of tissue samples under a microscope requires considerable time and effort from expert pathologists. There is a need to design an automated tool to assist pathologists for digitized histology slide analysis. Pre-cervical cancer is generally determined by examining the CIN which is the growth of atypical cells from the basement membrane (bottom) to the top of the epithelium. It has four grades, including: Normal, CIN1, CIN2, and CIN3. In this research, different facets of an automated digitized histology epithelium assessment pipeline have been explored to mimic the pathologist diagnostic approach. The entire pipeline from slide to epithelium CIN grade has been designed and developed using deep learning models and imaging techniques to analyze the whole slide image (WSI). The process is as follows: 1) identification of epithelium by filtering the regions extracted from a low-resolution image with a binary classifier network; 2) epithelium segmentation; 3) deep regression for pixel-wise segmentation of epithelium by patch-based image analysis; 4) attention-based CIN classification with localized sequential feature modeling. Deep learning-based nuclei detection by superpixels was performed as an extension of our research. Results from this research indicate an improved performance of CIN assessment over state-of-the-art methods for nuclei segmentation, epithelium segmentation, and CIN classification, as well as the development of a prototype WSI-level tool”--Abstract, page iv

    Biometric Systems

    Get PDF
    Because of the accelerating progress in biometrics research and the latest nation-state threats to security, this book's publication is not only timely but also much needed. This volume contains seventeen peer-reviewed chapters reporting the state of the art in biometrics research: security issues, signature verification, fingerprint identification, wrist vascular biometrics, ear detection, face detection and identification (including a new survey of face recognition), person re-identification, electrocardiogram (ECT) recognition, and several multi-modal systems. This book will be a valuable resource for graduate students, engineers, and researchers interested in understanding and investigating this important field of study

    GEOBIA 2016 : Solutions and Synergies., 14-16 September 2016, University of Twente Faculty of Geo-Information and Earth Observation (ITC): open access e-book

    Get PDF

    Aprendizado ativo baseado em atributos contextuais de superpixel para classificação de imagem de sensoriamento remoto

    Get PDF
    Orientadores: Alexandre Xavier Falcão, Jefersson Alex dos SantosDissertação (mestrado) - Universidade Estadual de Campinas, Instituto de ComputaçãoResumo: Recentemente, técnicas de aprendizado de máquina têm sido propostas para criar mapas temáticos a partir de imagens de sensoriamento remoto. Estas técnicas podem ser divididas em métodos de classificação baseados em pixels ou regiões. Este trabalho concentra-se na segunda abordagem, uma vez que estamos interessados em imagens com milhões de pixels e a segmentação da imagem em regiões (superpixels) pode reduzir consideravelmente o número de amostras a serem classificadas. Porém, mesmo utilizando superpixels, o número de amostras ainda é grande para anotá-las manualmente e treinar o classificador. As técnicas de aprendizado ativo propostas resolvem este problema começando pela seleção de um conjunto pequeno de amostras selecionadas aleatoriamente. Tais amostras são anotadas manualmente e utilizadas para treinar a primeira instância do classificador. Em cada iteração do ciclo de aprendizagem, o classificador atribui rótulos e seleciona as amostras mais informativas para a correção/confirmação pelo usuário, aumentando o tamanho do conjunto de treinamento. A instância do classificador é melhorada no final de cada iteração pelo seu treinamento e utilizada na iteração seguinte até que o usuário esteja satisfeito com o classificador. Observamos que a maior parte dos métodos reclassificam o conjunto inteiro de dados em cada iteração do ciclo de aprendizagem, tornando este processo inviável para interação com o usuário. Portanto, enderaçamos dois problemas importantes em classificação baseada em regiões de imagens de sensoriamento remoto: (a) a descrição efetiva de superpixels e (b) a redução do tempo requerido para seleção de amostras em aprendizado ativo. Primeiro, propusemos um descritor contextual de superpixels baseado na técnica de sacola de palavras, que melhora o resultado de descritores de cor e textura amplamente utilizados. Posteriormente, propusemos um método supervisionado de redução do conjunto de dados que é baseado em um método do estado da arte em aprendizado ativo chamado Multi-Class Level Uncertainty (MCLU). Nosso método mostrou-se tão eficaz quanto o MCLU e ao mesmo tempo consideravelmente mais eficiente. Adicionalmente, melhoramos seu desempenho por meio da aplicação de um processo de relaxação no mapa de classificação, utilizando Campos Aleatórios de MarkovAbstract: In recent years, machine learning techniques have been proposed to create classification maps from remote sensing images. These techniques can be divided into pixel- and region-based image classification methods. This work concentrates on the second approach, since we are interested in images with millions of pixels and the segmentation of the image into regions (superpixels) can considerably reduce the number of samples for classification. However, even using superpixels the number of samples is still large for manual annotation of samples to train the classifier. Active learning techniques have been proposed to address the problem by starting from a small set of randomly selected samples, which are manually labeled and used to train a first instance of the classifier. At each learning iteration, the classifier assigns labels and selects the most informative samples for user correction/confirmation, increasing the size of the training set. An improved instance of the classifier is created by training, after each iteration, and used in the next iteration until the user is satisfied with the classifier. We observed that most methods reclassify the entire pool of unlabeled samples at every learning iteration, making the process unfeasible for user interaction. Therefore, we address two important problems in region-based classification of remote sensing images: (a) the effective superpixel description and (b) the reduction of the time required for sample selection in active learning. First, we propose a contextual superpixel descriptor, based on bag of visual words, that outperforms widely used color and texture descriptors. Second, we propose a supervised method for dataset reduction that is based on a state-of-art active learning technique, called Multi-Class Level Uncertainty (MCLU). Our method has shown to be as effective as MCLU, while being considerably more efficient. Additionally, we further improve its performance by applying a relaxation process on the classification map by using Markov Random FieldsMestradoCiência da ComputaçãoMestre em Ciência da Computaçã

    Supervised evaluation of image segmentation and object proposal techniques

    Get PDF
    This paper tackles the supervised evaluation of image segmentation and object proposal algorithms. It surveys, structures, and deduplicates the measures used to compare both segmentation results and object proposals with a ground truth database; and proposes a new measure: the precision-recall for objects and parts. To compare the quality of these measures, eight state-of-the-art object proposal techniques are analyzed and two quantitative meta-measures involving nine state of the art segmentation methods are presented. The meta-measures consist in assuming some plausible hypotheses about the results and assessing how well each measure reflects these hypotheses. As a conclusion of the performed experiments, this paper proposes the tandem of precision-recall curves for boundaries and for objects-and-parts as the tool of choice for the supervised evaluation of image segmentation. We make the datasets and code of all the measures publicly available.Peer ReviewedPostprint (author's final draft

    Automatic extraction of retinal features to assist diagnosis of glaucoma disease

    Get PDF
    Glaucoma is a group of eye diseases that have common traits such as high eye pressure, damage to the Optic Nerve Head (ONH) and gradual vision loss. It affects the peripheral vision and eventually leads to blindness if left untreated. The current common methods of diagnosis of glaucoma are performed manually by the clinicians. Clinicians perform manual image operations such as change of contrast, zooming in zooming out etc to observe glaucoma related clinical indications. This type of diagnostic process is time consuming and subjective. With the advancement of image and vision computing, by automating steps in the diagnostic process, more patients can be screened and early treatment can be provided to prevent any or further loss of vision. The aim of this work is to develop a system called Glaucoma Detection Framework (GDF), which can automatically determine changes in retinal structures and imagebased pattern associated with glaucoma so as to assist the eye clinicians for glaucoma diagnosis in a timely and effective manner. In this work, several major contributions have been made towards the development of the automatic GDF consisting of the stages of preprocessing, optic disc and cup segmentation and regional image feature methods for classification between glaucoma and normal images. Firstly, in the preprocessing step, a retinal area detector based on superpixel classification model has been developed in order to automatically determine true retinal area from a Scanning Laser Ophthalmoscope (SLO) image. The retinal area detector can automatically extract artefacts out from the SLO image while preserving the computational effciency and avoiding over-segmentation of the artefacts. Localization of the ONH is one of the important steps towards the glaucoma analysis. A new weighted feature map approach has been proposed, which can enhance the region of ONH for accurate localization. For determining vasculature shift, which is one of glaucoma indications, we proposed the ONH cropped image based vasculature classification model to segment out the vasculature from the ONH cropped image. The ONH cropped image based vasculature classification model is developed in order to avoid misidentification of optic disc boundary and Peripapillary Atrophy (PPA) around the ONH of being a part of the vasculature area. Secondly, for automatic determination of optic disc and optic cup boundaries, a Point Edge Model (PEM), a Weighted Point Edge Model (WPEM) and a Region Classification Model (RCM) have been proposed. The RCM initially determines the optic disc region using the set of feature maps most suitable for the region classification whereas the PEM updates the contour using the force field of the feature maps with strong edge profile. The combination of PEM and RCM entitled Point Edge and Region Classification Model (PERCM) has significantly increased the accuracy of optic disc segmentation with respect to clinical annotations around optic disc. On the other hand, the WPEM determines the force field using the weighted feature maps calculated by the RCM for optic cup in order to enhance the optic cup region compared to rim area in the ONH. The combination of WPEM and RCM entitled Weighted Point Edge and Region Classification Model (WPERCM) can significantly enhance the accuracy of optic cup segmentation. Thirdly, this work proposes a Regional Image Features Model (RIFM) which can automatically perform classification between normal and glaucoma images on the basis of regional information. Different from the existing methods focusing on global features information only, our approach after optic disc localization and segmentation can automatically divide an image into five regions (i.e. optic disc or Optic Nerve Head (ONH) area, inferior (I), superior(S), nasal(N) and temporal(T)). These regions are usually used for diagnosis of glaucoma by clinicians through visual observation only. It then extracts image-based information such as textural, spatial and frequency based information so as to distinguish between normal and glaucoma images. The method provides a new way to identify glaucoma symptoms without determining any geometrical measurement associated with clinical indications glaucoma. Finally, we have accommodated clinical indications of glaucoma including the CDR, vasculature shift and neuroretinal rim loss with the RIFM classification and performed automatic classification between normal and glaucoma images. Since based on the clinical literature, no geometrical measurement is the guaranteed sign of glaucoma, the accommodation of the RIFM classification results with clinical indications of glaucoma can lead to more accurate classification between normal and glaucoma images. The proposed methods in this work have been tested against retinal image databases of 208 fundus images and 102 Scanning Laser Ophthalmoscope (SLO) images. These databases have been annotated by the clinicians around different anatomical structures associated with glaucoma as well as annotated with healthy or glaucomatous images. In fundus images, ONH cropped images have resolution varying from 300 to 900 whereas in SLO images, the resolution is 341 x 341. The accuracy of classification between normal and glaucoma images on fundus images and the SLO images is 94.93% and 98.03% respectively
    corecore