38,673 research outputs found
An Adaptive Strategy for Active Learning with Smooth Decision Boundary
We present the first adaptive strategy for active learning in the setting of
classification with smooth decision boundary. The problem of adaptivity (to
unknown distributional parameters) has remained opened since the seminal work
of Castro and Nowak (2007), which first established (active learning) rates for
this setting. While some recent advances on this problem establish adaptive
rates in the case of univariate data, adaptivity in the more practical setting
of multivariate data has so far remained elusive. Combining insights from
various recent works, we show that, for the multivariate case, a careful
reduction to univariate-adaptive strategies yield near-optimal rates without
prior knowledge of distributional parameters
Adaptivity to Noise Parameters in Nonparametric Active Learning
This work addresses various open questions in the theory of active learning
for nonparametric classification. Our contributions are both statistical and
algorithmic: -We establish new minimax-rates for active learning under common
\textit{noise conditions}. These rates display interesting transitions -- due
to the interaction between noise \textit{smoothness and margin} -- not present
in the passive setting. Some such transitions were previously conjectured, but
remained unconfirmed. -We present a generic algorithmic strategy for adaptivity
to unknown noise smoothness and margin; our strategy achieves optimal rates in
many general situations; furthermore, unlike in previous work, we avoid the
need for \textit{adaptive confidence sets}, resulting in strictly milder
distributional requirements
Sequential Design for Ranking Response Surfaces
We propose and analyze sequential design methods for the problem of ranking
several response surfaces. Namely, given response surfaces over a
continuous input space , the aim is to efficiently find the index of
the minimal response across the entire . The response surfaces are not
known and have to be noisily sampled one-at-a-time. This setting is motivated
by stochastic control applications and requires joint experimental design both
in space and response-index dimensions. To generate sequential design
heuristics we investigate stepwise uncertainty reduction approaches, as well as
sampling based on posterior classification complexity. We also make connections
between our continuous-input formulation and the discrete framework of pure
regret in multi-armed bandits. To model the response surfaces we utilize
kriging surrogates. Several numerical examples using both synthetic data and an
epidemics control problem are provided to illustrate our approach and the
efficacy of respective adaptive designs.Comment: 26 pages, 7 figures (updated several sections and figures
An Incremental Construction of Deep Neuro Fuzzy System for Continual Learning of Non-stationary Data Streams
Existing FNNs are mostly developed under a shallow network configuration
having lower generalization power than those of deep structures. This paper
proposes a novel self-organizing deep FNN, namely DEVFNN. Fuzzy rules can be
automatically extracted from data streams or removed if they play limited role
during their lifespan. The structure of the network can be deepened on demand
by stacking additional layers using a drift detection method which not only
detects the covariate drift, variations of input space, but also accurately
identifies the real drift, dynamic changes of both feature space and target
space. DEVFNN is developed under the stacked generalization principle via the
feature augmentation concept where a recently developed algorithm, namely
gClass, drives the hidden layer. It is equipped by an automatic feature
selection method which controls activation and deactivation of input attributes
to induce varying subsets of input features. A deep network simplification
procedure is put forward using the concept of hidden layer merging to prevent
uncontrollable growth of dimensionality of input space due to the nature of
feature augmentation approach in building a deep network structure. DEVFNN
works in the sample-wise fashion and is compatible for data stream
applications. The efficacy of DEVFNN has been thoroughly evaluated using seven
datasets with non-stationary properties under the prequential test-then-train
protocol. It has been compared with four popular continual learning algorithms
and its shallow counterpart where DEVFNN demonstrates improvement of
classification accuracy. Moreover, it is also shown that the concept drift
detection method is an effective tool to control the depth of network structure
while the hidden layer merging scenario is capable of simplifying the network
complexity of a deep network with negligible compromise of generalization
performance.Comment: This paper has been published in IEEE Transactions on Fuzzy System
- …