58,804 research outputs found
Overlearning in marginal distribution-based ICA: analysis and solutions
The present paper is written as a word of caution, with users of
independent component analysis (ICA) in mind, to overlearning
phenomena that are often observed.\\
We consider two types of overlearning, typical to high-order
statistics based ICA. These algorithms can be seen to maximise the
negentropy of the source estimates. The first kind of overlearning
results in the generation of spike-like signals, if there are not
enough samples in the data or there is a considerable amount of
noise present. It is argued that, if the data has power spectrum
characterised by curve, we face a more severe problem, which
cannot be solved inside the strict ICA model. This overlearning is
better characterised by bumps instead of spikes. Both overlearning
types are demonstrated in the case of artificial signals as well as
magnetoencephalograms (MEG). Several methods are suggested to
circumvent both types, either by making the estimation of the ICA
model more robust or by including further modelling of the data
Reducing "Structure From Motion": a General Framework for Dynamic Vision - Part 2: Experimental Evaluation
A number of methods have been proposed in the literature for estimating scene-structure and ego-motion from a sequence of images using dynamical models. Although all methods may be derived from a "natural" dynamical model within a unified framework, from an engineering perspective there are a number of trade-offs that lead to different strategies depending upon the specific applications and the goals one is targeting.
Which one is the winning strategy? In this paper we analyze the properties of the dynamical models that originate from each strategy under a variety of experimental conditions. For each model we assess the accuracy of the estimates, their robustness to measurement noise, sensitivity to initial conditions and visual angle, effects of the bas-relief ambiguity and occlusions, dependence upon the number of image measurements and their sampling rate
Reducing “Structure from Motion”: a general framework for dynamic vision. 2. Implementation and experimental assessment
For pt.1 see ibid., p.933-42 (1998). A number of methods have been proposed in the literature for estimating scene-structure and ego-motion from a sequence of images using dynamical models. Despite the fact that all methods may be derived from a “natural” dynamical model within a unified framework, from an engineering perspective there are a number of trade-offs that lead to different strategies depending upon the applications and the goals one is targeting. We want to characterize and compare the properties of each model such that the engineer may choose the one best suited to the specific application. We analyze the properties of filters derived from each dynamical model under a variety of experimental conditions, assess the accuracy of the estimates, their robustness to measurement noise, sensitivity to initial conditions and visual angle, effects of the bas-relief ambiguity and occlusions, dependence upon the number of image measurements and their sampling rate
Fitting a function to time-dependent ensemble averaged data
Time-dependent ensemble averages, i.e., trajectory-based averages of some
observable, are of importance in many fields of science. A crucial objective
when interpreting such data is to fit these averages (for instance, squared
displacements) with a function and extract parameters (such as diffusion
constants). A commonly overlooked challenge in such function fitting procedures
is that fluctuations around mean values, by construction, exhibit temporal
correlations. We show that the only available general purpose function fitting
methods, correlated chi-square method and the weighted least squares method
(which neglects correlation), fail at either robust parameter estimation or
accurate error estimation. We remedy this by deriving a new closed-form error
estimation formula for weighted least square fitting. The new formula uses the
full covariance matrix, i.e., rigorously includes temporal correlations, but is
free of the robustness issues, inherent to the correlated chi-square method. We
demonstrate its accuracy in four examples of importance in many fields:
Brownian motion, damped harmonic oscillation, fractional Brownian motion and
continuous time random walks. We also successfully apply our method, weighted
least squares including correlation in error estimation (WLS-ICE), to particle
tracking data. The WLS-ICE method is applicable to arbitrary fit functions, and
we provide a publically available WLS-ICE software.Comment: 47 pages (main text: 15 pages, supplementary: 32 pages
Variance-Optimal Offline and Streaming Stratified Random Sampling
Stratified random sampling (SRS) is a fundamental sampling technique that
provides accurate estimates for aggregate queries using a small size sample,
and has been used widely for approximate query processing. A key question in
SRS is how to partition a target sample size among different strata. While
Neyman allocation provides a solution that minimizes the variance of an
estimate using this sample, it works under the assumption that each stratum is
abundant, i.e., has a large number of data points to choose from. This
assumption may not hold in general: one or more strata may be bounded, and may
not contain a large number of data points, even though the total data size may
be large.
We first present VOILA, an offline method for allocating sample sizes to
strata in a variance-optimal manner, even for the case when one or more strata
may be bounded. We next consider SRS on streaming data that are continuously
arriving. We show a lower bound, that any streaming algorithm for SRS must have
(in the worst case) a variance that is {\Omega}(r) factor away from the
optimal, where r is the number of strata. We present S-VOILA, a practical
streaming algorithm for SRS that is locally variance-optimal in its allocation
of sample sizes to different strata. Our result from experiments on real and
synthetic data show that VOILA can have significantly (1.4 to 50.0 times)
smaller variance than Neyman allocation. The streaming algorithm S-VOILA
results in a variance that is typically close to VOILA, which was given the
entire input beforehand
- …