14,800 research outputs found
DEEP FULLY RESIDUAL CONVOLUTIONAL NEURAL NETWORK FOR SEMANTIC IMAGE SEGMENTATION
Department of Computer Science and EngineeringThe goal of semantic image segmentation is to partition the pixels of an image into semantically meaningful parts and classifying those parts according to a predefined label set. Although object recognition
models achieved remarkable performance recently and they even surpass human???s ability to recognize
objects, but semantic segmentation models are still behind. One of the reason that makes semantic
segmentation relatively a hard problem is the image understanding at pixel level by considering global
context as oppose to object recognition. One other challenge is transferring the knowledge of an object
recognition model for the task of semantic segmentation. In this thesis, we are delineating some of the
main challenges we faced approaching semantic image segmentation with machine learning algorithms.
Our main focus was how we can use deep learning algorithms for this task since they require the
least amount of feature engineering and also it was shown that such models can be applied to large scale
datasets and exhibit remarkable performance. More precisely, we worked on a variation of convolutional
neural networks (CNN) suitable for the semantic segmentation task. We proposed a model called deep
fully residual convolutional networks (DFRCN) to tackle this problem. Utilizing residual learning makes
training of deep models feasible which ultimately leads to having a rich powerful visual representation.
Our model also benefits from skip-connections which ease the propagation of information from the
encoder module to the decoder module. This would enable our model to have less parameters in the
decoder module while it also achieves better performance. We also benchmarked the effective variation
of the proposed model on a semantic segmentation benchmark.
We first make a thorough review of current high-performance models and the problems one might
face when trying to replicate such models which mainly arose from the lack of sufficient provided
information. Then, we describe our own novel method which we called deep fully residual convolutional
network (DFRCN). We showed that our method exhibits state of the art performance on a challenging
benchmark for aerial image segmentation.clos
Real-time Semantic Segmentation of Crop and Weed for Precision Agriculture Robots Leveraging Background Knowledge in CNNs
Precision farming robots, which target to reduce the amount of herbicides
that need to be brought out in the fields, must have the ability to identify
crops and weeds in real time to trigger weeding actions. In this paper, we
address the problem of CNN-based semantic segmentation of crop fields
separating sugar beet plants, weeds, and background solely based on RGB data.
We propose a CNN that exploits existing vegetation indexes and provides a
classification in real time. Furthermore, it can be effectively re-trained to
so far unseen fields with a comparably small amount of training data. We
implemented and thoroughly evaluated our system on a real agricultural robot
operating in different fields in Germany and Switzerland. The results show that
our system generalizes well, can operate at around 20Hz, and is suitable for
online operation in the fields.Comment: Accepted for publication at IEEE International Conference on Robotics
and Automation 2018 (ICRA 2018
Deep Extreme Cut: From Extreme Points to Object Segmentation
This paper explores the use of extreme points in an object (left-most,
right-most, top, bottom pixels) as input to obtain precise object segmentation
for images and videos. We do so by adding an extra channel to the image in the
input of a convolutional neural network (CNN), which contains a Gaussian
centered in each of the extreme points. The CNN learns to transform this
information into a segmentation of an object that matches those extreme points.
We demonstrate the usefulness of this approach for guided segmentation
(grabcut-style), interactive segmentation, video object segmentation, and dense
segmentation annotation. We show that we obtain the most precise results to
date, also with less user input, in an extensive and varied selection of
benchmarks and datasets. All our models and code are publicly available on
http://www.vision.ee.ethz.ch/~cvlsegmentation/dextr/.Comment: CVPR 2018 camera ready. Project webpage and code:
http://www.vision.ee.ethz.ch/~cvlsegmentation/dextr
RAR-U-Net: a Residual Encoder to Attention Decoder by Residual Connections Framework for Spine Segmentation under Noisy Labels
Segmentation algorithms of medical image volumes are widely studied for many
clinical and research purposes. We propose a novel and efficient framework for
medical image segmentation. The framework functions under a deep learning
paradigm, incorporating four novel contributions. Firstly, a residual
interconnection is explored in different scale encoders. Secondly, four copy
and crop connections are replaced to residual-block-based concatenation to
alleviate the disparity between encoders and decoders, respectively. Thirdly,
convolutional attention modules for feature refinement are studied on all scale
decoders. Finally, an adaptive denoising learning strategy(ADL) based on the
training process from underfitting to overfitting is studied. Experimental
results are illustrated on a publicly available benchmark database of spine
CTs. Our segmentation framework achieves competitive performance with other
state-of-the-art methods over a variety of different evaluation measures
- …