404 research outputs found

    3D-BEVIS: Bird's-Eye-View Instance Segmentation

    Full text link
    Recent deep learning models achieve impressive results on 3D scene analysis tasks by operating directly on unstructured point clouds. A lot of progress was made in the field of object classification and semantic segmentation. However, the task of instance segmentation is less explored. In this work, we present 3D-BEVIS, a deep learning framework for 3D semantic instance segmentation on point clouds. Following the idea of previous proposal-free instance segmentation approaches, our model learns a feature embedding and groups the obtained feature space into semantic instances. Current point-based methods scale linearly with the number of points by processing local sub-parts of a scene individually. However, to perform instance segmentation by clustering, globally consistent features are required. Therefore, we propose to combine local point geometry with global context information from an intermediate bird's-eye view representation.Comment: camera-ready version for GCPR '1

    Efficient Active Learning for Image Classification and Segmentation using a Sample Selection and Conditional Generative Adversarial Network

    Get PDF
    Training robust deep learning (DL) systems for medical image classification or segmentation is challenging due to limited images covering different disease types and severity. We propose an active learning (AL) framework to select most informative samples and add to the training data. We use conditional generative adversarial networks (cGANs) to generate realistic chest xray images with different disease characteristics by conditioning its generation on a real image sample. Informative samples to add to the training set are identified using a Bayesian neural network. Experiments show our proposed AL framework is able to achieve state of the art performance by using about 35% of the full dataset, thus saving significant time and effort over conventional methods

    Semi-Supervised Deep Learning for Fully Convolutional Networks

    Full text link
    Deep learning usually requires large amounts of labeled training data, but annotating data is costly and tedious. The framework of semi-supervised learning provides the means to use both labeled data and arbitrary amounts of unlabeled data for training. Recently, semi-supervised deep learning has been intensively studied for standard CNN architectures. However, Fully Convolutional Networks (FCNs) set the state-of-the-art for many image segmentation tasks. To the best of our knowledge, there is no existing semi-supervised learning method for such FCNs yet. We lift the concept of auxiliary manifold embedding for semi-supervised learning to FCNs with the help of Random Feature Embedding. In our experiments on the challenging task of MS Lesion Segmentation, we leverage the proposed framework for the purpose of domain adaptation and report substantial improvements over the baseline model.Comment: 9 pages, 6 figure

    Hydrodynamic instability and sound amplification over a perforated plate backed by a cavity

    Get PDF
    International audienceThe long-wavelength hydrodynamic behaviour over a cavity-backed perforated plate, in a duct with a mean shear flow, is studied numerically using the multimodal method, where the acoustic and hydrodynamic disturbances are calculated from the linearized Euler equations. The flow-acoustic coupling near the perforated plate is first solved hole by hole, and results indicate a well-defined large-scale hydrodynamic wave over the plate, with a wavelength close to the plate length at the peak sound amplification frequency when a plane acoustic wave is introduced from the upstream duct. Since the hydrodynamic wavelength is one order larger than the period of the perforation, the effect of the perforated plate is then described by a homogeneous plate impedance. It is shown that the homogenized approach approximately represents the discrete approach in this problem

    Improving Whole Slide Segmentation Through Visual Context - A Systematic Study

    Full text link
    While challenging, the dense segmentation of histology images is a necessary first step to assess changes in tissue architecture and cellular morphology. Although specific convolutional neural network architectures have been applied with great success to the problem, few effectively incorporate visual context information from multiple scales. With this paper, we present a systematic comparison of different architectures to assess how including multi-scale information affects segmentation performance. A publicly available breast cancer and a locally collected prostate cancer datasets are being utilised for this study. The results support our hypothesis that visual context and scale play a crucial role in histology image classification problems

    Experimental observation of a hydrodynamic mode in a flow duct with a porous material

    Get PDF
    This paper experimentally investigates the acoustic behavior of a homogeneous porous material with a rigid frame (metallic foam) under grazing flow. The transmission coefficient shows an unusual oscillation over a particular range of frequencies which reports the presence of an unstable hydrodynamic wave that can exchange energy with the acoustic waves. This coupling of acoustic and hydrodynamic waves becomes larger when the Mach number increases. A rise of the static pressure drop in the lined region is induced by an acoustic excitation when the hydrodynamic wave is present

    Dilated Convolutional Neural Networks for Cardiovascular MR Segmentation in Congenital Heart Disease

    Full text link
    We propose an automatic method using dilated convolutional neural networks (CNNs) for segmentation of the myocardium and blood pool in cardiovascular MR (CMR) of patients with congenital heart disease (CHD). Ten training and ten test CMR scans cropped to an ROI around the heart were provided in the MICCAI 2016 HVSMR challenge. A dilated CNN with a receptive field of 131x131 voxels was trained for myocardium and blood pool segmentation in axial, sagittal and coronal image slices. Performance was evaluated within the HVSMR challenge. Automatic segmentation of the test scans resulted in Dice indices of 0.80±\pm0.06 and 0.93±\pm0.02, average distances to boundaries of 0.96±\pm0.31 and 0.89±\pm0.24 mm, and Hausdorff distances of 6.13±\pm3.76 and 7.07±\pm3.01 mm for the myocardium and blood pool, respectively. Segmentation took 41.5±\pm14.7 s per scan. In conclusion, dilated CNNs trained on a small set of CMR images of CHD patients showing large anatomical variability provide accurate myocardium and blood pool segmentations

    Concurrent Segmentation and Localization for Tracking of Surgical Instruments

    Full text link
    Real-time instrument tracking is a crucial requirement for various computer-assisted interventions. In order to overcome problems such as specular reflections and motion blur, we propose a novel method that takes advantage of the interdependency between localization and segmentation of the surgical tool. In particular, we reformulate the 2D instrument pose estimation as heatmap regression and thereby enable a concurrent, robust and near real-time regression of both tasks via deep learning. As demonstrated by our experimental results, this modeling leads to a significantly improved performance than directly regressing the tool position and allows our method to outperform the state of the art on a Retinal Microsurgery benchmark and the MICCAI EndoVis Challenge 2015.Comment: I. Laina and N. Rieke contributed equally to this work. Accepted to MICCAI 201

    A deep level set method for image segmentation

    Full text link
    This paper proposes a novel image segmentation approachthat integrates fully convolutional networks (FCNs) with a level setmodel. Compared with a FCN, the integrated method can incorporatesmoothing and prior information to achieve an accurate segmentation.Furthermore, different than using the level set model as a post-processingtool, we integrate it into the training phase to fine-tune the FCN. Thisallows the use of unlabeled data during training in a semi-supervisedsetting. Using two types of medical imaging data (liver CT and left ven-tricle MRI data), we show that the integrated method achieves goodperformance even when little training data is available, outperformingthe FCN or the level set model alone

    Unsupervised Holistic Image Generation from Key Local Patches

    Full text link
    We introduce a new problem of generating an image based on a small number of key local patches without any geometric prior. In this work, key local patches are defined as informative regions of the target object or scene. This is a challenging problem since it requires generating realistic images and predicting locations of parts at the same time. We construct adversarial networks to tackle this problem. A generator network generates a fake image as well as a mask based on the encoder-decoder framework. On the other hand, a discriminator network aims to detect fake images. The network is trained with three losses to consider spatial, appearance, and adversarial information. The spatial loss determines whether the locations of predicted parts are correct. Input patches are restored in the output image without much modification due to the appearance loss. The adversarial loss ensures output images are realistic. The proposed network is trained without supervisory signals since no labels of key parts are required. Experimental results on six datasets demonstrate that the proposed algorithm performs favorably on challenging objects and scenes.Comment: 16 page
    • …
    corecore