354 research outputs found
State-of-the-art and gaps for deep learning on limited training data in remote sensing
Deep learning usually requires big data, with respect to both volume and
variety. However, most remote sensing applications only have limited training
data, of which a small subset is labeled. Herein, we review three
state-of-the-art approaches in deep learning to combat this challenge. The
first topic is transfer learning, in which some aspects of one domain, e.g.,
features, are transferred to another domain. The next is unsupervised learning,
e.g., autoencoders, which operate on unlabeled data. The last is generative
adversarial networks, which can generate realistic looking data that can fool
the likes of both a deep learning network and human. The aim of this article is
to raise awareness of this dilemma, to direct the reader to existing work and
to highlight current gaps that need solving.Comment: arXiv admin note: text overlap with arXiv:1709.0030
ARIGAN: Synthetic Arabidopsis Plants using Generative Adversarial Network
In recent years, there has been an increasing interest in image-based plant
phenotyping, applying state-of-the-art machine learning approaches to tackle
challenging problems, such as leaf segmentation (a multi-instance problem) and
counting. Most of these algorithms need labelled data to learn a model for the
task at hand. Despite the recent release of a few plant phenotyping datasets,
large annotated plant image datasets for the purpose of training deep learning
algorithms are lacking. One common approach to alleviate the lack of training
data is dataset augmentation. Herein, we propose an alternative solution to
dataset augmentation for plant phenotyping, creating artificial images of
plants using generative neural networks. We propose the Arabidopsis Rosette
Image Generator (through) Adversarial Network: a deep convolutional network
that is able to generate synthetic rosette-shaped plants, inspired by DCGAN (a
recent adversarial network model using convolutional layers). Specifically, we
trained the network using A1, A2, and A4 of the CVPPP 2017 LCC dataset,
containing Arabidopsis Thaliana plants. We show that our model is able to
generate realistic 128x128 colour images of plants. We train our network
conditioning on leaf count, such that it is possible to generate plants with a
given number of leaves suitable, among others, for training regression based
models. We propose a new Ax dataset of artificial plants images, obtained by
our ARIGAN. We evaluate this new dataset using a state-of-the-art leaf counting
algorithm, showing that the testing error is reduced when Ax is used as part of
the training data.Comment: 8 pages, 6 figures, 1 table, ICCV CVPPP Workshop 201
Convolutional Neural Net-Based Cassava Storage Root Counting Using Real and Synthetic Images
© Copyright © 2019 Atanbori, Montoya-P, Selvaraj, French and Pridmore. Cassava roots are complex structures comprising several distinct types of root. The number and size of the storage roots are two potential phenotypic traits reflecting crop yield and quality. Counting and measuring the size of cassava storage roots are usually done manually, or semi-automatically by first segmenting cassava root images. However, occlusion of both storage and fibrous roots makes the process both time-consuming and error-prone. While Convolutional Neural Nets have shown performance above the state-of-the-art in many image processing and analysis tasks, there are currently a limited number of Convolutional Neural Net-based methods for counting plant features. This is due to the limited availability of data, annotated by expert plant biologists, which represents all possible measurement outcomes. Existing works in this area either learn a direct image-to-count regressor model by regressing to a count value, or perform a count after segmenting the image. We, however, address the problem using a direct image-to-count prediction model. This is made possible by generating synthetic images, using a conditional Generative Adversarial Network (GAN), to provide training data for missing classes. We automatically form cassava storage root masks for any missing classes using existing ground-truth masks, and input them as a condition to our GAN model to generate synthetic root images. We combine the resulting synthetic images with real images to learn a direct image-to-count prediction model capable of counting the number of storage roots in real cassava images taken from a low cost aeroponic growth system. These models are used to develop a system that counts cassava storage roots in real images. Our system first predicts age group ('young' and 'old' roots; pertinent to our image capture regime) in a given image, and then, based on this prediction, selects an appropriate model to predict the number of storage roots. We achieve 91% accuracy on predicting ages of storage roots, and 86% and 71% overall percentage agreement on counting 'old' and 'young' storage roots respectively. Thus we are able to demonstrate that synthetically generated cassava root images can be used to supplement missing root classes, turning the counting problem into a direct image-to-count prediction task
A Comprehensive Survey of Image Augmentation Techniques for Deep Learning
Deep learning has been achieving decent performance in computer vision
requiring a large volume of images, however, collecting images is expensive and
difficult in many scenarios. To alleviate this issue, many image augmentation
algorithms have been proposed as effective and efficient strategies.
Understanding current algorithms is essential to find suitable methods or
develop novel techniques for given tasks. In this paper, we perform a
comprehensive survey on image augmentation for deep learning with a novel
informative taxonomy. To get the basic idea why we need image augmentation, we
introduce the challenges in computer vision tasks and vicinity distribution.
Then, the algorithms are split into three categories; model-free, model-based,
and optimizing policy-based. The model-free category employs image processing
methods while the model-based method leverages trainable image generation
models. In contrast, the optimizing policy-based approach aims to find the
optimal operations or their combinations. Furthermore, we discuss the current
trend of common applications with two more active topics, leveraging different
ways to understand image augmentation, such as group and kernel theory, and
deploying image augmentation for unsupervised learning. Based on the analysis,
we believe that our survey gives a better understanding helpful to choose
suitable methods or design novel algorithms for practical applications.Comment: Revisio
Multi-Spectral Image Synthesis for Crop/Weed Segmentation in Precision Farming
An effective perception system is a fundamental component for farming robots,
as it enables them to properly perceive the surrounding environment and to
carry out targeted operations. The most recent approaches make use of
state-of-the-art machine learning techniques to learn an effective model for
the target task. However, those methods need a large amount of labelled data
for training. A recent approach to deal with this issue is data augmentation
through Generative Adversarial Networks (GANs), where entire synthetic scenes
are added to the training data, thus enlarging and diversifying their
informative content. In this work, we propose an alternative solution with
respect to the common data augmentation techniques, applying it to the
fundamental problem of crop/weed segmentation in precision farming. Starting
from real images, we create semi-artificial samples by replacing the most
relevant object classes (i.e., crop and weeds) with their synthesized
counterparts. To do that, we employ a conditional GAN (cGAN), where the
generative model is trained by conditioning the shape of the generated object.
Moreover, in addition to RGB data, we take into account also near-infrared
(NIR) information, generating four channel multi-spectral synthetic images.
Quantitative experiments, carried out on three publicly available datasets,
show that (i) our model is capable of generating realistic multi-spectral
images of plants and (ii) the usage of such synthetic images in the training
process improves the segmentation performance of state-of-the-art semantic
segmentation Convolutional Networks.Comment: Submitted to Robotics and Autonomous System
- …