1,783 research outputs found
Directed quantum communication
We raise the question whether there is a way to characterize the quantum
information transport properties of a medium or material. For this analysis the
special features of quantum information have to be taken into account. We find
that quantum communication over an isotropic medium, as opposed to classical
information transfer, requires the transmitter to direct the signal towards the
receiver. Furthermore, for large classes of media there is a threshold, in the
sense that `sufficiently much' of the signal has to be collected. Therefore,
the medium's capacity for quantum communication can be characterized in terms
of how the size of the transmitter and receiver has to scale with the
transmission distance to maintain quantum information transmission. To
demonstrate the applicability of this concept, an n-dimensional spin lattice is
considered, yielding a sufficient scaling of d^(n/3) with the distance d
Impossibility of Growing Quantum Bit Commitments
Quantum key distribution (QKD) is often, more correctly, called key growing.
Given a short key as a seed, QKD enables two parties, connected by an insecure
quantum channel, to generate a secret key of arbitrary length. Conversely, no
key agreement is possible without access to an initial key. Here, we consider
another fundamental cryptographic task, commitments. While, similar to key
agreement, commitments cannot be realized from scratch, we ask whether they may
be grown. That is, given the ability to commit to a fixed number of bits, is
there a way to augment this to commitments to strings of arbitrary length?
Using recently developed information-theoretic techniques, we answer this
question to the negative.Comment: 10 pages, minor change
Spatial prediction of species’ distributions from occurrence-only records: combining point pattern analysis, ENFA and regression-kriging
A computational framework to map species’ distributions (realized density) using occurrence-only data and environmental predictors is presented and illustrated using a textbook example and two case studies: distribution of root vole (Microtes oeconomus) in the Netherlands, and distribution of white-tailed eagle nests (Haliaeetus albicilla) in Croatia. The framework combines strengths of point pattern analysis (kernel smoothing), Ecological Niche Factor Analysis (ENFA) and geostatistics (logistic regression-kriging), as implemented in the spatstat, adehabitat and gstat packages of the R environment for statistical computing. A procedure to generate pseudo-absences is proposed. It uses Habitat Suitability Index (HSI, derived through ENFA) and distance from observations as weight maps to allocate pseudo-absence points. This design ensures that the simulated pseudo-absences fall further away from the occurrence points in both feature and geographical spaces. The simulated pseudo-absences can then be combined with occurrence locations and used to build regression-kriging prediction models. The output of prediction are either probabilitiesy of species’ occurrence or density measures. Addition of the pseudo-absence locations has proven effective — the adjusted R-square increased from 0.71 to 0.80 for root vole (562 records), and from 0.69 to 0.83 for white-tailed eagle (135 records) respectively; pseudo-absences improve spreading of the points in feature space and ensure consistent mapping over the whole area of interest. Results of cross validation (leave-one-out method) for these two species showed that the model explains 98% of the total variability in the density values for the root vole, and 94% of the total variability for the white-tailed eagle. The framework could be further extended to Generalized multivariate Linear Geostatistical Models and spatial prediction of multiple species. A copy of the R script and step-by-step instructions to run such analysis are available via contact author’s website
On the uncertainty of stream networks derived from elevation data: the error propagation approach
DEM error propagation methodology is extended to the derivation of vector-based objects (stream networks) using geostatistical simulations. First, point sampled elevations are used to fit a variogram model. Next 100 DEM realizations are generated using conditional sequential Gaussian simulation; the stream network map is extracted for each of these realizations, and the collection of stream networks is analyzed to quantify the error propagation. At each grid cell, the probability of the occurrence of a stream and the propagated error are estimated. The method is illustrated using two small data sets: Baranja hill (30 m grid cell size; 16 512 pixels; 6367 sampled elevations), and Zlatibor (30 m grid cell size; 15 000 pixels; 2051 sampled elevations). All computations are run in the open source software for statistical computing R: package geoR is used to fit variogram; package gstat is used to run sequential Gaussian simulation; streams are extracted using the open source GIS SAGA via the RSAGA library. The resulting stream error map (Information entropy of a Bernoulli trial) clearly depicts areas where the extracted stream network is least precise – usually areas of low local relief and slightly convex (0–10 difference from the mean value). In both cases, significant parts of the study area (17.3% for Baranja Hill; 6.2% for Zlatibor) show high error (H>0.5) of locating streams. By correlating the propagated uncertainty of the derived stream network with various land surface parameters sampling of height measurements can be optimized so that delineated streams satisfy the required accuracy level. Such error propagation tool should become a standard functionality in any modern GIS. Remaining issue to be tackled is the computational burden of geostatistical simulations: this framework is at the moment limited to small data sets with several hundreds of points. Scripts and data sets used in this article are available on-line via the www.geomorphometry.org website and can be easily adopted/adjusted to any similar case study
Rheo-acoustic gels: Tuning mechanical and flow properties of colloidal gels with ultrasonic vibrations
Colloidal gels, where nanoscale particles aggregate into an elastic yet
fragile network, are at the heart of materials that combine specific optical,
electrical and mechanical properties. Tailoring the viscoelastic features of
colloidal gels in real-time thanks to an external stimulus currently appears as
a major challenge in the design of "smart" soft materials. Here we introduce
"rheo-acoustic" gels, a class of materials that are sensitive to ultrasonic
vibrations. By using a combination of rheological and structural
characterization, we evidence and quantify a strong softening in three widely
different colloidal gels submitted to ultrasonic vibrations (with submicron
amplitude and frequency 20-500 kHz). This softening is attributed to
micron-sized cracks within the gel network that may or may not fully heal once
vibrations are turned off depending on the acoustic intensity. Ultrasonic
vibrations are further shown to dramatically decrease the gel yield stress and
accelerate shear-induced fluidization. Ultrasound-assisted fluidization
dynamics appear to be governed by an effective temperature that depends on the
acoustic intensity. Our work opens the way to a full control of elastic and
flow properties by ultrasonic vibrations as well as to future theoretical and
numerical modeling of such rheo-acoustic gels.Comment: 21 pages, 14 figure
Recommended from our members
Combined use of satellite estimates and rain gauge observations to generate high-quality historical rainfall time series over Ethiopia
Climate data are used in a number of applications including climate risk management and adaptation to climate change. However, the availability of climate data, particularly throughout rural Africa, is very limited. Available weather stations are unevenly distributed and mainly located along main roads in cities and towns. This imposes severe limitations to the availability of climate information and services for the rural community where, arguably, these services are needed most. Weather station data also suffer from gaps in the time series. Satellite proxies, particularly satellite rainfall estimate, have been used as alternatives because of their availability even over remote parts of the world. However, satellite rainfall estimates also suffer from a number of critical shortcomings that include heterogeneous time series, short time period of observation, and poor accuracy particularly at higher temporal and spatial resolutions. An attempt is made here to alleviate these problems by combining station measurements with the complete spatial coverage of satellite rainfall estimates. Rain gauge observations are merged with a locally calibrated version of the TAMSAT satellite rainfall estimates to produce over 30-years (1983-todate) of rainfall estimates over Ethiopia at a spatial resolution of 10 km and a ten-daily time scale. This involves quality control of rain gauge data, generating locally calibrated version of the TAMSAT rainfall estimates, and combining these with rain gauge observations from national station network. The infrared-only satellite rainfall estimates produced using a relatively simple TAMSAT algorithm performed as good as or even better than other satellite rainfall products that use passive microwave inputs and more sophisticated algorithms. There is no substantial difference between the gridded-gauge and combined gauge-satellite products over the test area in Ethiopia having a dense station network; however, the combined product exhibits better quality over parts of the country where stations are sparsely distributed
A divide-and-conquer approach to analyze underdetermined biochemical models
Motivation: To obtain meaningful predictions from dynamic computational models, their uncertain parameter values need to be estimated from experimental data. Due to the usually large number of parameters compared to the available measurement data, these estimation problems are often underdetermined meaning that the solution is a multidimensional space. In this case, the challenge is yet to obtain a sound system understanding despite non-identifiable parameter values, e.g. through identifying those parameters that most sensitively determine the model’s behavior.
Results: Here, we present the so-called divide-and-conquer approach—a strategy to analyze underdetermined biochemical models. The approach draws on steady state omics measurement data and exploits a decomposition of the global estimation problem into independent subproblems. The solutions to these subproblems are joined to the complete space of global optima, which can be easily analyzed. We derive the conditions at which the decomposition occurs, outline strategies to fulfill these conditions and—using an example model—illustrate how the approach uncovers the most important parameters and suggests targeted experiments without knowing the exact parameter values.
Soil property maps of the agricultural land of Hungary
Continuous soil maps were prepared on the basis of three point databases, which cover the agricultural areas of Hungary.JRC.H.5 - Land Resources Managemen
In silico labeling reveals the time-dependent label half-life and transit-time in dynamical systems
Background: Mathematical models of dynamical systems facilitate the computation of characteristic properties that are not accessible experimentally. In cell biology, two main properties of interest are (1) the time-period a protein is accessible to other molecules in a certain state - its half-life - and (2) the time it spends when passing through a subsystem - its transit-time. We discuss two approaches to quantify the half-life, present the novel method of in silico labeling, and introduce the label half-life and label transit-time. The developed method has been motivated by laboratory tracer experiments. To investigate the kinetic properties and behavior of a substance of interest, we computationally label this species in order to track it throughout its life cycle. The corresponding mathematical model is extended by an additional set of reactions for the labeled species, avoiding any double-counting within closed circuits, correcting for the influences of upstream fluxes, and taking into account combinatorial multiplicity for complexes or reactions with several reactants or products. A profile likelihood approach is used to estimate confidence intervals on the label half-life and transit-time. Results: Application to the JAK-STAT signaling pathway in Epo-stimulated BaF3-EpoR cells enabled the calculation of the time-dependent label half-life and transit-time of STAT species. The results were robust against parameter uncertainties. Conclusions: Our approach renders possible the estimation of species and label half-lives and transit-times. It is applicable to large non-linear systems and an implementation is provided within the PottersWheel modeling framework (http://www.potterswheel.de)
- …
