13,521 research outputs found
Incremental Few-Shot Object Detection via Simple Fine-Tuning Approach
In this paper, we explore incremental few-shot object detection (iFSD), which
incrementally learns novel classes using only a few examples without revisiting
base classes. Previous iFSD works achieved the desired results by applying
meta-learning. However, meta-learning approaches show insufficient performance
that is difficult to apply to practical problems. In this light, we propose a
simple fine-tuning-based approach, the Incremental Two-stage Fine-tuning
Approach (iTFA) for iFSD, which contains three steps: 1) base training using
abundant base classes with the class-agnostic box regressor, 2) separation of
the RoI feature extractor and classifier into the base and novel class branches
for preserving base knowledge, and 3) fine-tuning the novel branch using only a
few novel class examples. We evaluate our iTFA on the real-world datasets
PASCAL VOC, COCO, and LVIS. iTFA achieves competitive performance in COCO and
shows a 30% higher AP accuracy than meta-learning methods in the LVIS dataset.
Experimental results show the effectiveness and applicability of our proposed
method.Comment: Accepted to ICRA 202
Prototypical quadruplet for few-shot class incremental learning
Many modern computer vision algorithms suffer from two major bottlenecks:
scarcity of data and learning new tasks incrementally. While training the model
with new batches of data the model looses it's ability to classify the previous
data judiciously which is termed as catastrophic forgetting. Conventional
methods have tried to mitigate catastrophic forgetting of the previously
learned data while the training at the current session has been compromised.
The state-of-the-art generative replay based approaches use complicated
structures such as generative adversarial network (GAN) to deal with
catastrophic forgetting. Additionally, training a GAN with few samples may lead
to instability. In this work, we present a novel method to deal with these two
major hurdles. Our method identifies a better embedding space with an improved
contrasting loss to make classification more robust. Moreover, our approach is
able to retain previously acquired knowledge in the embedding space even when
trained with new classes. We update previous session class prototypes while
training in such a way that it is able to represent the true class mean. This
is of prime importance as our classification rule is based on the nearest class
mean classification strategy. We have demonstrated our results by showing that
the embedding space remains intact after training the model with new classes.
We showed that our method preformed better than the existing state-of-the-art
algorithms in terms of accuracy across different sessions
A Survey on Few-Shot Class-Incremental Learning
Large deep learning models are impressive, but they struggle when real-time data is not available. Few-shot class-incremental learning (FSCIL) poses a significant challenge for deep neural networks to learn new tasks from just a few labeled samples without forgetting the previously learned ones. This setup can easily leads to catastrophic forgetting and overfitting problems, severely affecting model performance. Studying FSCIL helps overcome deep learning model limitations on data volume and acquisition time, while improving practicality and adaptability of machine learning models. This paper provides a comprehensive survey on FSCIL. Unlike previous surveys, we aim to synthesize few-shot learning and incremental learning, focusing on introducing FSCIL from two perspectives, while reviewing over 30 theoretical research studies and more than 20 applied research studies. From the theoretical perspective, we provide a novel categorization approach that divides the field into five subcategories, including traditional machine learning methods, meta learning-based methods, feature and feature space-based methods, replay-based methods, and dynamic network structure-based methods. We also evaluate the performance of recent theoretical research on benchmark datasets of FSCIL. From the application perspective, FSCIL has achieved impressive achievements in various fields of computer vision such as image classification, object detection, and image segmentation, as well as in natural language processing and graph. We summarize the important applications. Finally, we point out potential future research directions, including applications, problem setups, and theory development. Overall, this paper offers a comprehensive analysis of the latest advances in FSCIL from a methodological, performance, and application perspective
Learning Multimodal Latent Attributes
Abstract—The rapid development of social media sharing has created a huge demand for automatic media classification and annotation techniques. Attribute learning has emerged as a promising paradigm for bridging the semantic gap and addressing data sparsity via transferring attribute knowledge in object recognition and relatively simple action classification. In this paper, we address the task of attribute learning for understanding multimedia data with sparse and incomplete labels. In particular we focus on videos of social group activities, which are particularly challenging and topical examples of this task because of their multi-modal content and complex and unstructured nature relative to the density of annotations. To solve this problem, we (1) introduce a concept of semi-latent attribute space, expressing user-defined and latent attributes in a unified framework, and (2) propose a novel scalable probabilistic topic model for learning multi-modal semi-latent attributes, which dramatically reduces requirements for an exhaustive accurate attribute ontology and expensive annotation effort. We show that our framework is able to exploit latent attributes to outperform contemporary approaches for addressing a variety of realistic multimedia sparse data learning tasks including: multi-task learning, learning with label noise, N-shot transfer learning and importantly zero-shot learning
A Survey on Few-Shot Class-Incremental Learning
Large deep learning models are impressive, but they struggle when real-time
data is not available. Few-shot class-incremental learning (FSCIL) poses a
significant challenge for deep neural networks to learn new tasks from just a
few labeled samples without forgetting the previously learned ones. This setup
easily leads to catastrophic forgetting and overfitting problems, severely
affecting model performance. Studying FSCIL helps overcome deep learning model
limitations on data volume and acquisition time, while improving practicality
and adaptability of machine learning models. This paper provides a
comprehensive survey on FSCIL. Unlike previous surveys, we aim to synthesize
few-shot learning and incremental learning, focusing on introducing FSCIL from
two perspectives, while reviewing over 30 theoretical research studies and more
than 20 applied research studies. From the theoretical perspective, we provide
a novel categorization approach that divides the field into five subcategories,
including traditional machine learning methods, meta-learning based methods,
feature and feature space-based methods, replay-based methods, and dynamic
network structure-based methods. We also evaluate the performance of recent
theoretical research on benchmark datasets of FSCIL. From the application
perspective, FSCIL has achieved impressive achievements in various fields of
computer vision such as image classification, object detection, and image
segmentation, as well as in natural language processing and graph. We summarize
the important applications. Finally, we point out potential future research
directions, including applications, problem setups, and theory development.
Overall, this paper offers a comprehensive analysis of the latest advances in
FSCIL from a methodological, performance, and application perspective
Transductive Multi-View Zero-Shot Learning
(c) 2012. The copyright of this document resides with its authors.
It may be distributed unchanged freely in print or electronic forms
- …