89 research outputs found
Efficient On-the-fly Category Retrieval using ConvNets and GPUs
We investigate the gains in precision and speed, that can be obtained by
using Convolutional Networks (ConvNets) for on-the-fly retrieval - where
classifiers are learnt at run time for a textual query from downloaded images,
and used to rank large image or video datasets.
We make three contributions: (i) we present an evaluation of state-of-the-art
image representations for object category retrieval over standard benchmark
datasets containing 1M+ images; (ii) we show that ConvNets can be used to
obtain features which are incredibly performant, and yet much lower dimensional
than previous state-of-the-art image representations, and that their
dimensionality can be reduced further without loss in performance by
compression using product quantization or binarization. Consequently, features
with the state-of-the-art performance on large-scale datasets of millions of
images can fit in the memory of even a commodity GPU card; (iii) we show that
an SVM classifier can be learnt within a ConvNet framework on a GPU in parallel
with downloading the new training images, allowing for a continuous refinement
of the model as more images become available, and simultaneous training and
ranking. The outcome is an on-the-fly system that significantly outperforms its
predecessors in terms of: precision of retrieval, memory requirements, and
speed, facilitating accurate on-the-fly learning and ranking in under a second
on a single GPU.Comment: Published in proceedings of ACCV 201
OnionNet: Sharing Features in Cascaded Deep Classifiers
The focus of our work is speeding up evaluation of deep neural networks in
retrieval scenarios, where conventional architectures may spend too much time
on negative examples. We propose to replace a monolithic network with our novel
cascade of feature-sharing deep classifiers, called OnionNet, where subsequent
stages may add both new layers as well as new feature channels to the previous
ones. Importantly, intermediate feature maps are shared among classifiers,
preventing them from the necessity of being recomputed. To accomplish this, the
model is trained end-to-end in a principled way under a joint loss. We validate
our approach in theory and on a synthetic benchmark. As a result demonstrated
in three applications (patch matching, object detection, and image retrieval),
our cascade can operate significantly faster than both monolithic networks and
traditional cascades without sharing at the cost of marginal decrease in
precision.Comment: Accepted to BMVC 201
On Video Analysis of Omnidirectional Bee Traffic: Counting Bee Motions with Motion Detection and Image Classification
Omnidirectional bee traffic is the number of bees moving in arbitrary directions in close proximity to the landing pad of a given hive over a given period of time. Video bee traffic analysis has the potential to automate the assessment of omnidirectional bee traffic levels, which, in turn, may lead to a complete or partial automation of honeybee colony health assessment. In this investigation, we proposed, implemented, and partially evaluated a two-tier method for counting bee motions to estimate levels of omnidirectional bee traffic in bee traffic videos. Our method couples motion detection with image classification so that motion detection acts as a class-agnostic object location method that generates a set of regions with possible objects and each such region is classified by a class-specific classifier such as a convolutional neural network or a support vector machine or an ensemble of classifiers such as a random forest. The method has been, and is being iteratively field tested in BeePi monitors, multi-sensor electronic beehive monitoring systems, installed on live Langstroth beehives in real apiaries. Deployment of a BeePi monitor on top of a beehive does not require any structural modification of the beehive’s woodenware, and is not disruptive to natural beehive cycles. To ensure the replicability of the reported findings and to provide a performance benchmark for interested research communities and citizen scientists, we have made public our curated and labeled image datasets of 167,261 honeybee images and our omnidirectional bee traffic videos used in this investigation
Iterative Soft Shrinkage Learning for Efficient Image Super-Resolution
Image super-resolution (SR) has witnessed extensive neural network designs
from CNN to transformer architectures. However, prevailing SR models suffer
from prohibitive memory footprint and intensive computations, which limits
further deployment on edge devices. This work investigates the potential of
network pruning for super-resolution to take advantage of off-the-shelf network
designs and reduce the underlying computational overhead. Two main challenges
remain in applying pruning methods for SR. First, the widely-used filter
pruning technique reflects limited granularity and restricted adaptability to
diverse network structures. Second, existing pruning methods generally operate
upon a pre-trained network for the sparse structure determination, hard to get
rid of dense model training in the traditional SR paradigm. To address these
challenges, we adopt unstructured pruning with sparse models directly trained
from scratch. Specifically, we propose a novel Iterative Soft
Shrinkage-Percentage (ISS-P) method by optimizing the sparse structure of a
randomly initialized network at each iteration and tweaking unimportant weights
with a small amount proportional to the magnitude scale on-the-fly. We observe
that the proposed ISS-P can dynamically learn sparse structures adapting to the
optimization process and preserve the sparse model's trainability by yielding a
more regularized gradient throughput. Experiments on benchmark datasets
demonstrate the effectiveness of the proposed ISS-P over diverse network
architectures. Code is available at
https://github.com/Jiamian-Wang/Iterative-Soft-Shrinkage-SRComment: Accepted by ICCV 2023, code released at
https://github.com/Jiamian-Wang/Iterative-Soft-Shrinkage-S
A Survey From Distributed Machine Learning to Distributed Deep Learning
Artificial intelligence has achieved significant success in handling complex
tasks in recent years. This success is due to advances in machine learning
algorithms and hardware acceleration. In order to obtain more accurate results
and solve more complex problems, algorithms must be trained with more data.
This huge amount of data could be time-consuming to process and require a great
deal of computation. This solution could be achieved by distributing the data
and algorithm across several machines, which is known as distributed machine
learning. There has been considerable effort put into distributed machine
learning algorithms, and different methods have been proposed so far. In this
article, we present a comprehensive summary of the current state-of-the-art in
the field through the review of these algorithms. We divide this algorithms in
classification and clustering (traditional machine learning), deep learning and
deep reinforcement learning groups. Distributed deep learning has gained more
attention in recent years and most of studies worked on this algorithms. As a
result, most of the articles we discussed here belong to this category. Based
on our investigation of algorithms, we highlight limitations that should be
addressed in future research
- …