759 research outputs found
Evolution of cosmological constant in effective gravity
In contrast to the phenomenon of nullification of the cosmological constant
in the equilibrium vacuum, which is the general property of any quantum vacuum,
there are many options in modifying the Einstein equation to allow the
cosmological constant to evolve in a non-equilibrium vacuum. An attempt is made
to extend the Einstein equation in the direction suggested by the
condensed-matter analogy of the quantum vacuum. Different scenarios are found
depending on the behavior of and the relation between the relaxation parameters
involved, some of these scenarios having been discussed in the literature. One
of them reproduces the scenario in which the effective cosmological constant
emerges as a constant of integration. The second one describes the situation,
when after the cosmological phase transition the cosmological constant drops
from zero to the negative value; this scenario describes the relaxation from
this big negative value back to zero and then to a small positive value. In the
third example the relaxation time is not a constant but depends on matter; this
scenario demonstrates that the vacuum energy (or its fraction) can play the
role of the cold dark matter.Comment: LaTeX file, 5 pages, no figures, version submitted to JETP Letter
The Revival of the Unified Dark Energy-Dark Matter Model ?
We consider the generalized Chaplygin gas (GCG) proposal for unification of
dark energy and dark matter and show that it admits an unique decomposition
into dark energy and dark matter components once phantom-like dark energy is
excluded. Within this framework, we study structure formation and show that
difficulties associated to unphysical oscillations or blow-up in the matter
power spectrum can be circumvented. Furthermore, we show that the dominance of
dark energy is related to the time when energy density fluctuations start
deviating from the linear behaviour.Comment: 6 pages, 4 eps figures, Revtex4 style. New References are added. Some
typos are corrected. Conclusions remain the sam
Both Ca2+ and Zn2+ are essential for S100A12 protein oligomerization and function
Background
Human S100A12 is a member of the S100 family of EF-hand calcium-modulated proteins that are associated with many diseases including cancer, chronic inflammation and neurological disorders. S100A12 is an important factor in host/parasite defenses and in the inflammatory response. Like several other S100 proteins, it binds zinc and copper in addition to calcium. Mechanisms of zinc regulation have been proposed for a number of S100 proteins e.g. S100B, S100A2, S100A7, S100A8/9. The interaction of S100 proteins with their targets is strongly dependent on cellular microenvironment.
Results
The aim of the study was to explore the factors that influence S100A12 oligomerization and target interaction. A comprehensive series of biochemical and biophysical experiments indicated that changes in the concentration of calcium and zinc led to changes in the oligomeric state of S100A12. Surface plasmon resonance confirmed that the presence of both calcium and zinc is essential for the interaction of S100A12 with one of its extracellular targets, RAGE – the Receptor for Advanced Glycation End products. By using a single-molecule approach we have shown that the presence of zinc in tissue culture medium favors both the oligomerization of exogenous S100A12 protein and its interaction with targets on the cell surface.
Conclusion
We have shown that oligomerization and target recognition by S100A12 is regulated by both zinc and calcium. Our present work highlighted the potential role of calcium-binding S100 proteins in zinc metabolism and, in particular, the role of S100A12 in the cross talk between zinc and calcium in cell signaling
SHREC 2011: robust feature detection and description benchmark
Feature-based approaches have recently become very popular in computer vision
and image analysis applications, and are becoming a promising direction in
shape retrieval. SHREC'11 robust feature detection and description benchmark
simulates the feature detection and description stages of feature-based shape
retrieval algorithms. The benchmark tests the performance of shape feature
detectors and descriptors under a wide variety of transformations. The
benchmark allows evaluating how algorithms cope with certain classes of
transformations and strength of the transformations that can be dealt with. The
present paper is a report of the SHREC'11 robust feature detection and
description benchmark results.Comment: This is a full version of the SHREC'11 report published in 3DO
Particle Counting Statistics of Time and Space Dependent Fields
The counting statistics give insight into the properties of quantum states of
light and other quantum states of matter such as ultracold atoms or electrons.
The theoretical description of photon counting was derived in the 1960s and was
extended to massive particles more recently. Typically, the interaction between
each particle and the detector is assumed to be limited to short time
intervals, and the probability of counting particles in one interval is
independent of the measurements in previous intervals. There has been some
effort to describe particle counting as a continuous measurement, where the
detector and the field to be counted interact continuously. However, no general
formula applicable to any time and space dependent field has been derived so
far. In our work, we derive a fully time and space dependent description of the
counting process for linear quantum many-body systems, taking into account the
back-action of the detector on the field. We apply our formalism to an
expanding Bose-Einstein condensate of ultracold atoms, and show that it
describes the process correctly, whereas the standard approach gives unphysical
results in some limits. The example illustrates that in certain situations, the
back-action of the detector cannot be neglected and has to be included in the
description
Exploring the origin of high optical absorption in conjugated polymers
The specific optical absorption of an organic semiconductor is critical to the performance of organic optoelectronic devices. For example, higher light-harvesting efficiency can lead to higher photocurrent in solar cells that are limited by sub-optimal electrical transport. Here, we compare over 40 conjugated polymers, and find that many different chemical structures share an apparent maximum in their extinction coefficients. However, a diketopyrrolopyrrole-thienothiophene copolymer shows remarkably high optical absorption at relatively low photon energies. By investigating its backbone structure and conformation with measurements and quantum chemical calculations, we find that the high optical absorption can be explained by the high persistence length of the polymer. Accordingly, we demonstrate high absorption in other polymers with high theoretical persistence length. Visible light harvesting may be enhanced in other conjugated polymers through judicious design of the structure
Deep Modeling of Growth Trajectories for Longitudinal Prediction of Missing Infant Cortical Surfaces
Charting cortical growth trajectories is of paramount importance for
understanding brain development. However, such analysis necessitates the
collection of longitudinal data, which can be challenging due to subject
dropouts and failed scans. In this paper, we will introduce a method for
longitudinal prediction of cortical surfaces using a spatial graph
convolutional neural network (GCNN), which extends conventional CNNs from
Euclidean to curved manifolds. The proposed method is designed to model the
cortical growth trajectories and jointly predict inner and outer cortical
surfaces at multiple time points. Adopting a binary flag in loss calculation to
deal with missing data, we fully utilize all available cortical surfaces for
training our deep learning model, without requiring a complete collection of
longitudinal data. Predicting the surfaces directly allows cortical attributes
such as cortical thickness, curvature, and convexity to be computed for
subsequent analysis. We will demonstrate with experimental results that our
method is capable of capturing the nonlinearity of spatiotemporal cortical
growth patterns and can predict cortical surfaces with improved accuracy.Comment: Accepted as oral presentation at IPMI 201
A Graph Theoretic Approach for Object Shape Representation in Compositional Hierarchies Using a Hybrid Generative-Descriptive Model
A graph theoretic approach is proposed for object shape representation in a
hierarchical compositional architecture called Compositional Hierarchy of Parts
(CHOP). In the proposed approach, vocabulary learning is performed using a
hybrid generative-descriptive model. First, statistical relationships between
parts are learned using a Minimum Conditional Entropy Clustering algorithm.
Then, selection of descriptive parts is defined as a frequent subgraph
discovery problem, and solved using a Minimum Description Length (MDL)
principle. Finally, part compositions are constructed by compressing the
internal data representation with discovered substructures. Shape
representation and computational complexity properties of the proposed approach
and algorithms are examined using six benchmark two-dimensional shape image
datasets. Experiments show that CHOP can employ part shareability and indexing
mechanisms for fast inference of part compositions using learned shape
vocabularies. Additionally, CHOP provides better shape retrieval performance
than the state-of-the-art shape retrieval methods.Comment: Paper : 17 pages. 13th European Conference on Computer Vision (ECCV
2014), Zurich, Switzerland, September 6-12, 2014, Proceedings, Part III, pp
566-581. Supplementary material can be downloaded from
http://link.springer.com/content/esm/chp:10.1007/978-3-319-10578-9_37/file/MediaObjects/978-3-319-10578-9_37_MOESM1_ESM.pd
Instrumentation-related uncertainty of reflectance and transmittance measurements with a two-channel spectrophotometer
Spectrophotometers are operated in numerous fields of science and industry for a variety of applications. In order to provide confidence for the measured data, analyzing the associated uncertainty is valuable. However, the uncertainty of the measurement results is often unknown or reduced to sample-related contributions. In this paper, we describe our approach for the systematic determination of the measurement uncertainty of the commercially available two-channel spectrophotometer Agilent Cary 5000 in accordance with the Guide to the expression of uncertainty in measurements. We focus on the instrumentation-related uncertainty contributions rather than the specific application and thus outline a general procedure which can be adapted for other instruments. Moreover, we discover a systematic signal deviation due to the inertia of the measurement amplifier and develop and apply a correction procedure. Thereby we increase the usable dynamic range of the instrument by more than one order of magnitude. We present methods for the quantification of the uncertainty contributions and combine them into an uncertainty budget for the device. © 2017 Author(s)
Fully Automatic Expression-Invariant Face Correspondence
We consider the problem of computing accurate point-to-point correspondences
among a set of human face scans with varying expressions. Our fully automatic
approach does not require any manually placed markers on the scan. Instead, the
approach learns the locations of a set of landmarks present in a database and
uses this knowledge to automatically predict the locations of these landmarks
on a newly available scan. The predicted landmarks are then used to compute
point-to-point correspondences between a template model and the newly available
scan. To accurately fit the expression of the template to the expression of the
scan, we use as template a blendshape model. Our algorithm was tested on a
database of human faces of different ethnic groups with strongly varying
expressions. Experimental results show that the obtained point-to-point
correspondence is both highly accurate and consistent for most of the tested 3D
face models
- …