108 research outputs found
Evaluation of Deep Learning based Pose Estimation for Sign Language Recognition
Human body pose estimation and hand detection are two important tasks for
systems that perform computer vision-based sign language recognition(SLR).
However, both tasks are challenging, especially when the input is color videos,
with no depth information. Many algorithms have been proposed in the literature
for these tasks, and some of the most successful recent algorithms are based on
deep learning. In this paper, we introduce a dataset for human pose estimation
for SLR domain. We evaluate the performance of two deep learning based pose
estimation methods, by performing user-independent experiments on our dataset.
We also perform transfer learning, and we obtain results that demonstrate that
transfer learning can improve pose estimation accuracy. The dataset and results
from these methods can create a useful baseline for future works
Deep Bottleneck Feature for Image Classification
Effective image representation plays an important role for image classification and retrieval. Bag-of-Features (BoF) is well known as an effective and robust visual representation. However, on large datasets, convolutional neural networks (CNN) tend to perform much better, aided by the availability of large amounts of training data. In this paper, we propose a bag of Deep Bottleneck Features (DBF) for image classification, effectively combining the strengths of a CNN within a BoF framework. The DBF features, obtained from a previously well-trained CNN, form a compact and low-dimensional representation of the original inputs, effective for even small datasets. We will demonstrate that the resulting BoDBF method has a very powerful and discriminative capability that is generalisable to other image classification tasks
Wildlife surveillance using deep learning methods
Wildlife conservation and the management of humanâwildlife conflicts require costâeffective methods of monitoring wild animal behavior. Still and video camera surveillance can generate enormous quantities of data, which is laborious and expensive to screen for the species of interest. In the present study, we describe a stateâofâtheâart, deep learning approach for automatically identifying and isolating speciesâspecific activity from still images and video data.
We used a dataset consisting of 8,368 images of wild and domestic animals in farm buildings, and we developed an approach firstly to distinguish badgers from other species (binary classification) and secondly to distinguish each of six animal species (multiclassification). We focused on binary classification of badgers first because such a tool would be relevant to efforts to manage Mycobacterium bovis (the cause of bovine tuberculosis) transmission between badgers and cattle.
We used two deep learning frameworks for automatic image recognition. They achieved high accuracies, in the order of 98.05% for binary classification and 90.32% for multiclassification. Based on the deep learning framework, a detection process was also developed for identifying animals of interest in video footage, which to our knowledge is the first application for this purpose.
The algorithms developed here have wide applications in wildlife monitoring where large quantities of visual data require screening for certain species
Learning to Selectively Transfer: Reinforced Transfer Learning for Deep Text Matching
Deep text matching approaches have been widely studied for many applications
including question answering and information retrieval systems. To deal with a
domain that has insufficient labeled data, these approaches can be used in a
Transfer Learning (TL) setting to leverage labeled data from a resource-rich
source domain. To achieve better performance, source domain data selection is
essential in this process to prevent the "negative transfer" problem. However,
the emerging deep transfer models do not fit well with most existing data
selection methods, because the data selection policy and the transfer learning
model are not jointly trained, leading to sub-optimal training efficiency.
In this paper, we propose a novel reinforced data selector to select
high-quality source domain data to help the TL model. Specifically, the data
selector "acts" on the source domain data to find a subset for optimization of
the TL model, and the performance of the TL model can provide "rewards" in turn
to update the selector. We build the reinforced data selector based on the
actor-critic framework and integrate it to a DNN based transfer learning model,
resulting in a Reinforced Transfer Learning (RTL) method. We perform a thorough
experimental evaluation on two major tasks for text matching, namely,
paraphrase identification and natural language inference. Experimental results
show the proposed RTL can significantly improve the performance of the TL
model. We further investigate different settings of states, rewards, and policy
optimization methods to examine the robustness of our method. Last, we conduct
a case study on the selected data and find our method is able to select source
domain data whose Wasserstein distance is close to the target domain data. This
is reasonable and intuitive as such source domain data can provide more
transferability power to the model.Comment: Accepted to WSDM 201
Recommended from our members
Deep Learning for Single-Molecule Science
Exploring and making predictions based on single-molecule data can be challenging, not only due to the sheer size of the datasets, but also because a priori knowledge about the signal characteristics is typically limited and poor signal-to-noise ratio. For example, hypothesis-driven data exploration, informed by an expectation of the signal characteristics, can lead to interpretation bias or loss of information. Equally, even when the different data categories are known, e.g., the four bases in DNA sequencing, it is often difficult to know how to make best use of the available information content. The latest developments in Machine Learning (ML), so-called Deep Learning (DL) offers an interesting, new avenues to address such challenges. In some applications, such as speech and image recognition, DL has been able to outperform conventional Machine Learning strategies and even human performance. However, to date DL has not been applied much in single-molecule science, presumably in part because relatively little is known about the 'internal workings' of such DL tools within single-molecule science as a field. In this Tutorial, we make an attempt to illustrate in a step-by-step guide how one of those, a Convolutional Neural Network, may be used for base calling in DNA sequencing applications. We compare it with a Support Vector Machine as a more conventional ML method, and and discuss some of the strengths and weaknesses of the approach. In particular, a 'deep' neural network has many features of a 'black box', which has important implications on how we look at and interpret data
Learning Fast Quadruped Robot Gaits with the RL PoWER Spline Parameterization
Legged robots are uniquely privileged over their wheeled counterparts in their potential to access rugged terrain. However, designing walking gaits by hand for legged robots is a difficult and time-consuming process, so we seek algorithms for learning such gaits to automatically using real world experimentation. Numerous previous studies have examined a variety of algorithms for learning gaits, using an assortment of different robots. It is often difficult to compare the algorithmic results from one study to the next, because the conditions and robots used vary. With this in mind, we have used an open-source, 3D printed quadruped robot called QuadraTot, so the results may be verified, and hopefully improved upon, by any group so desiring. Because many robots do not have accurate simulators, we test gait-learning algorithms entirely on the physical robot. Previous studies using the QuadraTot have compared parameterized splines, the HyperNEAT generative encoding and genetic algorithm. Among these, the research on the genetic algorithm was conducted by (G l e t t e et al., 2012) in a simulator and tested on a real robot. Here we compare these results to an algorithm called Policy learning by Weighting Exploration with the Returns, or RL PoWER. We report that this algorithm has learned the fastest gait through only physical experiments yet reported in the literature, 16.3% faster than reported for HyperNEAT. In addition, the learned gaits are less taxing on the robot and more repeatable than previous record-breaking gaits
- âŠ