1,333 research outputs found
Generalized Minimum Error with Fiducial Points Criterion for Robust Learning
The conventional Minimum Error Entropy criterion (MEE) has its limitations,
showing reduced sensitivity to error mean values and uncertainty regarding
error probability density function locations. To overcome this, a MEE with
fiducial points criterion (MEEF), was presented. However, the efficacy of the
MEEF is not consistent due to its reliance on a fixed Gaussian kernel. In this
paper, a generalized minimum error with fiducial points criterion (GMEEF) is
presented by adopting the Generalized Gaussian Density (GGD) function as
kernel. The GGD extends the Gaussian distribution by introducing a shape
parameter that provides more control over the tail behavior and peakedness. In
addition, due to the high computational complexity of GMEEF criterion, the
quantized idea is introduced to notably lower the computational load of the
GMEEF-type algorithm. Finally, the proposed criterions are introduced to the
domains of adaptive filter, kernel recursive algorithm, and multilayer
perceptron. Several numerical simulations, which contain system identification,
acoustic echo cancellation, times series prediction, and supervised
classification, indicate that the novel algorithms' performance performs
excellently.Comment: 12 pages, 9 figure
Few-Shot Classification with Contrastive Learning
A two-stage training paradigm consisting of sequential pre-training and
meta-training stages has been widely used in current few-shot learning (FSL)
research. Many of these methods use self-supervised learning and contrastive
learning to achieve new state-of-the-art results. However, the potential of
contrastive learning in both stages of FSL training paradigm is still not fully
exploited. In this paper, we propose a novel contrastive learning-based
framework that seamlessly integrates contrastive learning into both stages to
improve the performance of few-shot classification. In the pre-training stage,
we propose a self-supervised contrastive loss in the forms of feature vector
vs. feature map and feature map vs. feature map, which uses global and local
information to learn good initial representations. In the meta-training stage,
we propose a cross-view episodic training mechanism to perform the nearest
centroid classification on two different views of the same episode and adopt a
distance-scaled contrastive loss based on them. These two strategies force the
model to overcome the bias between views and promote the transferability of
representations. Extensive experiments on three benchmark datasets demonstrate
that our method achieves competitive results.Comment: To appear in ECCV 202
- …