23 research outputs found
Online Learning with Multiple Operator-valued Kernels
We consider the problem of learning a vector-valued function f in an online
learning setting. The function f is assumed to lie in a reproducing Hilbert
space of operator-valued kernels. We describe two online algorithms for
learning f while taking into account the output structure. A first contribution
is an algorithm, ONORMA, that extends the standard kernel-based online learning
algorithm NORMA from scalar-valued to operator-valued setting. We report a
cumulative error bound that holds both for classification and regression. We
then define a second algorithm, MONORMA, which addresses the limitation of
pre-defining the output structure in ONORMA by learning sequentially a linear
combination of operator-valued kernels. Our experiments show that the proposed
algorithms achieve good performance results with low computational cost
An Online Discriminative Approach to Background Subtraction
We present a simple, principled approach to detecting foreground objects in video sequences in real-time. Our method is based on an on-line discriminative learning technique that is able to cope with illumination changes due to discontinuous switching, or illumination drifts caused by slower processes such as varying time of the day. Starting from a discriminative learning principle, we derive a training algorithm that, for each pixel, computes a weighted linear combination of selected past observations with time-decay. We present experimental results that show the proposed approach outperforms existing methods on both synthetic sequences and real video data
Fast Bounded Online Gradient Descent Algorithms for Scalable Kernel-Based Online Learning
Kernel-based online learning has often shown state-of-the-art performance for
many online learning tasks. It, however, suffers from a major shortcoming, that
is, the unbounded number of support vectors, making it non-scalable and
unsuitable for applications with large-scale datasets. In this work, we study
the problem of bounded kernel-based online learning that aims to constrain the
number of support vectors by a predefined budget. Although several algorithms
have been proposed in literature, they are neither computationally efficient
due to their intensive budget maintenance strategy nor effective due to the use
of simple Perceptron algorithm. To overcome these limitations, we propose a
framework for bounded kernel-based online learning based on an online gradient
descent approach. We propose two efficient algorithms of bounded online
gradient descent (BOGD) for scalable kernel-based online learning: (i) BOGD by
maintaining support vectors using uniform sampling, and (ii) BOGD++ by
maintaining support vectors using non-uniform sampling. We present theoretical
analysis of regret bound for both algorithms, and found promising empirical
performance in terms of both efficacy and efficiency by comparing them to
several well-known algorithms for bounded kernel-based online learning on
large-scale datasets.Comment: ICML201
Online Learning Models for Vehicle Usage Prediction During COVID-19
Today, there is an ongoing transition to more sustainable transportation, for
which an essential part is the switch from combustion engine vehicles to
battery electric vehicles (BEVs). BEVs have many advantages from a
sustainability perspective, but issues such as limited driving range and long
recharge times slow down the transition from combustion engines. One way to
mitigate these issues is by performing battery thermal preconditioning, which
increases the energy efficiency of the battery. However, to optimally perform
battery thermal preconditioning, the vehicle usage pattern needs to be known,
i.e., how and when the vehicle will be used. This study attempts to predict the
departure time and distance of the first drive each day using online machine
learning models. The online machine learning models are trained and evaluated
on historical driving data collected from a fleet of BEVs during the COVID-19
pandemic. Additionally, the prediction models are extended to quantify the
uncertainty of their predictions, which can be used to decide whether the
prediction should be used or dismissed. Based on our results, the
best-performing prediction models yield an aggregated mean absolute error of
2.75 hours when predicting departure time and 13.37 km when predicting trip
distance.Comment: This article has been accepted for publication in IEEE Transactions
on Intelligent Transportation Systems. This is the author's version which has
not been fully edited and content may change prior to final publication.
Citation information: DOI 10.1109/TITS.2024.336167