333 research outputs found
A combinatorial approach to jumping particles II: general boundary conditions
International audienceWe consider a model of particles jumping on a row, the TASEP. From the point of view of combinatorics a remarkable feauture of this Markov chain is that Catalan numbers are involved in several entries of its stationary distribution. In a companion paper, we gave a combinatorial interpretaion and a simple proof of these observations in the simplest case where the particles enter, jump and exit at the same rate. In this paper we show how to deal with general rates
Simple Asymmetric Exclusion Model and Lattice Paths: Bijections and Involutions
We study the combinatorics of the change of basis of three representations of
the stationary state algebra of the two parameter simple asymmetric exclusion
process. Each of the representations considered correspond to a different set
of weighted lattice paths which, when summed over, give the stationary state
probability distribution. We show that all three sets of paths are
combinatorially related via sequences of bijections and sign reversing
involutions.Comment: 28 page
Scaling of the atmosphere of self-avoiding walks
The number of free sites next to the end of a self-avoiding walk is known as
the atmosphere. The average atmosphere can be related to the number of
configurations. Here we study the distribution of atmospheres as a function of
length and how the number of walks of fixed atmosphere scale. Certain bounds on
these numbers can be proved. We use Monte Carlo estimates to verify our
conjectures. Of particular interest are walks that have zero atmosphere, which
are known as trapped. We demonstrate that these walks scale in the same way as
the full set of self-avoiding walks, barring an overall constant factor
Sharing Social Network Data: Differentially Private Estimation of Exponential-Family Random Graph Models
Motivated by a real-life problem of sharing social network data that contain
sensitive personal information, we propose a novel approach to release and
analyze synthetic graphs in order to protect privacy of individual
relationships captured by the social network while maintaining the validity of
statistical results. A case study using a version of the Enron e-mail corpus
dataset demonstrates the application and usefulness of the proposed techniques
in solving the challenging problem of maintaining privacy \emph{and} supporting
open access to network data to ensure reproducibility of existing studies and
discovering new scientific insights that can be obtained by analyzing such
data. We use a simple yet effective randomized response mechanism to generate
synthetic networks under -edge differential privacy, and then use
likelihood based inference for missing data and Markov chain Monte Carlo
techniques to fit exponential-family random graph models to the generated
synthetic networks.Comment: Updated, 39 page
Deep Markov Random Field for Image Modeling
Markov Random Fields (MRFs), a formulation widely used in generative image
modeling, have long been plagued by the lack of expressive power. This issue is
primarily due to the fact that conventional MRFs formulations tend to use
simplistic factors to capture local patterns. In this paper, we move beyond
such limitations, and propose a novel MRF model that uses fully-connected
neurons to express the complex interactions among pixels. Through theoretical
analysis, we reveal an inherent connection between this model and recurrent
neural networks, and thereon derive an approximated feed-forward network that
couples multiple RNNs along opposite directions. This formulation combines the
expressive power of deep neural networks and the cyclic dependency structure of
MRF in a unified model, bringing the modeling capability to a new level. The
feed-forward approximation also allows it to be efficiently learned from data.
Experimental results on a variety of low-level vision tasks show notable
improvement over state-of-the-arts.Comment: Accepted at ECCV 201
BlinkML: Efficient Maximum Likelihood Estimation with Probabilistic Guarantees
The rising volume of datasets has made training machine learning (ML) models
a major computational cost in the enterprise. Given the iterative nature of
model and parameter tuning, many analysts use a small sample of their entire
data during their initial stage of analysis to make quick decisions (e.g., what
features or hyperparameters to use) and use the entire dataset only in later
stages (i.e., when they have converged to a specific model). This sampling,
however, is performed in an ad-hoc fashion. Most practitioners cannot precisely
capture the effect of sampling on the quality of their model, and eventually on
their decision-making process during the tuning phase. Moreover, without
systematic support for sampling operators, many optimizations and reuse
opportunities are lost.
In this paper, we introduce BlinkML, a system for fast, quality-guaranteed ML
training. BlinkML allows users to make error-computation tradeoffs: instead of
training a model on their full data (i.e., full model), BlinkML can quickly
train an approximate model with quality guarantees using a sample. The quality
guarantees ensure that, with high probability, the approximate model makes the
same predictions as the full model. BlinkML currently supports any ML model
that relies on maximum likelihood estimation (MLE), which includes Generalized
Linear Models (e.g., linear regression, logistic regression, max entropy
classifier, Poisson regression) as well as PPCA (Probabilistic Principal
Component Analysis). Our experiments show that BlinkML can speed up the
training of large-scale ML tasks by 6.26x-629x while guaranteeing the same
predictions, with 95% probability, as the full model.Comment: 22 pages, SIGMOD 201
Discretization of variational regularization in Banach spaces
Consider a nonlinear ill-posed operator equation where is
defined on a Banach space . In general, for solving this equation
numerically, a finite dimensional approximation of and an approximation of
are required. Moreover, in general the given data \yd of are noisy.
In this paper we analyze finite dimensional variational regularization, which
takes into account operator approximations and noisy data: We show
(semi-)convergence of the regularized solution of the finite dimensional
problems and establish convergence rates in terms of Bregman distances under
appropriate sourcewise representation of a solution of the equation. The more
involved case of regularization in nonseparable Banach spaces is discussed in
detail. In particular we consider the space of finite total variation
functions, the space of functions of finite bounded deformation, and the
--space
Aerosol mass and black carbon concentrations, a two year record at NCO-P (5079 m, Southern Himalayas)
Aerosol mass and the absorbing fraction are important variables, needed to constrain the role of atmospheric particles in the Earth radiation budget, both directly and indirectly through CCN activation. In particular, their monitoring in remote areas and mountain sites is essential for determining source regions, elucidating the mechanisms of long range transport of anthropogenic pollutants, and validating regional and global models. Since March 2006, aerosol mass and black carbon concentration have been monitored at the Nepal Climate Observatory-Pyramid, a permanent high-altitude research station located in the Khumbu valley at 5079 m a.s.l. below Mt. Everest. The first two-year averages of PM<sub>1</sub> and PM<sub>1−10</sub> mass were 1.94 μg m<sup>−3</sup> and 1.88 μg m<sup>−3</sup>, with standard deviations of 3.90 μg m<sup>−3</sup> and 4.45 μg m<sup>−3</sup>, respectively, while the black carbon concentration average is 160.5 ng m<sup>−3</sup>, with a standard deviation of 296.1 ng m<sup>−3</sup>. Both aerosol mass and black carbon show well defined annual cycles, with a maximum during the pre-monsoon season and a minimum during the monsoon. They also display a typical diurnal cycle during all the seasons, with the lowest particle concentration recorded during the night, and a considerable increase during the afternoon, revealing the major role played by thermal winds in influencing the behaviour of atmospheric compounds over the high Himalayas. The aerosol concentration is subject to high variability: in fact, as well as frequent "background conditions" (55% of the time) when BC concentrations are mainly below 100 ng m<sup>−3</sup>, concentrations up to 5 μg m<sup>−3</sup> are reached during some episodes (a few days every year) in the pre-monsoon seasons. The variability of PM and BC is the result of both short-term changes due to thermal wind development in the valley, and long-range transport/synoptic circulation. At NCO-P, higher concentrations of PM<sub>1</sub> and BC are mostly associated with regional circulation and westerly air masses from the Middle East, while the strongest contributions of mineral dust arrive from the Middle East and regional circulation, with a special contribution from North Africa and South-West Arabian Peninsula in post-monsoon and winter season
Total Least Squares In-Field Identification for MEMS-Based Inertial Measurement Units †
Inertial Measurement Units are widely used in various applications and, hardware-wise, they primarily consist of a tri-axial accelerometer and a tri-axial gyroscope. For low-end commercial employments, the low cost of the device is crucial: this makes MEMS-based sensors a popular choice in this context. However, MEMS-based transducers are prone to significant, non-uniform and environmental-condition-dependent systematic errors, that require frequent re-calibration to be eliminated. To this end, identification methods that can be performed in-field by non-expert users, without the need for high-precision or costly equipment, are of particular interest. In this paper, we propose an in-field identification procedure based on the Total Least Squares method for both tri-axial accelerometers and gyroscopes. The proposed identification model is linear and requires no prior knowledge of the parameters to be identified. It enables accelerometer calibration without the need for specific reference surface orientation relative to Earth’s gravity and allows gyroscope calibration to be performed independently of accelerometer data, without requiring the sensor’s sensitive axes to be aligned with the rotation axes during calibration. Experiments conducted on NXP sensors FXOS8700CQ and FXAS21002 demonstrated that using parameters identified by our method reduced cross-validation standard deviations by about two orders of magnitude compared to those obtained using manufacturer-provided parameters. This result indicates that our method enables the effective calibration of IMU sensor parameters, relying only on simple 3D-printed equipment and significantly improving IMU performance at minimal cost
Preliminary Estimation of Black Carbon Deposition from Nepal Climate Observatory-Pyramid Data and Its Possible Impact on Snow Albedo Changes Over Himalayan Glaciers During the Pre-Monsoon Season
The possible minimal range of reduction in snow surface albedo due to dry deposition of black carbon (BC) in the pre-monsoon period (March-May) was estimated as a lower bound together with the estimation of its accuracy, based on atmospheric observations at the Nepal Climate Observatory-Pyramid (NCO-P) sited at 5079 m a.s.l. in the Himalayan region. We estimated a total BC deposition rate of 2.89 g m-2 day-1 providing a total deposition of 266 micrograms/ square m for March-May at the site, based on a calculation with a minimal deposition velocity of 1.0 10(exp -4) m/s with atmospheric data of equivalent BC concentration. Main BC size at NCO-P site was determined as 103.1-669.8 nm by correlation analysis between equivalent BC concentration and particulate size distribution in the atmosphere. We also estimated BC deposition from the size distribution data and found that 8.7% of the estimated dry deposition corresponds to the estimated BC deposition from equivalent BC concentration data. If all the BC is deposited uniformly on the top 2-cm pure snow, the corresponding BC concentration is 26.0-68.2 microgram/kg assuming snow density variations of 195-512 kg/ cubic m of Yala Glacier close to NCO-P site. Such a concentration of BC in snow could result in 2.0-5.2% albedo reductions. From a simple numerical calculations and if assuming these albedo reductions continue throughout the year, this would lead to a runoff increases of 70-204 mm of water drainage equivalent of 11.6-33.9% of the annual discharge of a typical Tibetan glacier. Our estimates of BC concentration in snow surface for pre-monsoon season can be considered comparable to those at similar altitude in the Himalayan region, where glaciers and perpetual snow region starts in the vicinity of NCO-P. Our estimates from only BC are likely to represent a lower bound for snow albedo reductions, since a fixed slower deposition velocity was used and atmospheric wind and turbulence effects, snow aging, dust deposition, and snow albedo feedbacks were not considered. This study represents the first investigation about BC deposition on snow from atmospheric aerosol data in Himalayas and related albedo effect is especially the first track at the southern slope of Himalayas
- …
