91 research outputs found
Numeric and symbolic evaluation of the pfaffian of general skew-symmetric matrices
Evaluation of pfaffians arises in a number of physics applications, and for
some of them a direct method is preferable to using the determinantal formula.
We discuss two methods for the numerical evaluation of pfaffians. The first is
tridiagonalization based on Householder transformations. The main advantage of
this method is its numerical stability that makes unnecessary the
implementation of a pivoting strategy. The second method considered is based on
Aitken's block diagonalization formula. It yields to a kind of LU (similar to
Cholesky's factorization) decomposition (under congruence) of arbitrary
skew-symmetric matrices that is well suited both for the numeric and symbolic
evaluations of the pfaffian. Fortran subroutines (FORTRAN 77 and 90)
implementing both methods are given. We also provide simple implementations in
Python and Mathematica for purpose of testing, or for exploratory studies of
methods that make use of pfaffians.Comment: 13 pages, Download links:
http://gamma.ft.uam.es/robledo/Downloads.html and
http://www.phys.washington.edu/users/bertsch/computer.htm
Oral FluidâBased Biomarkers of Alveolar Bone Loss in Periodontitis
Periodontal disease is a bacteria-induced chronic inflammatory disease affecting the soft and hard supporting structures encompassing the teeth. When left untreated, the ultimate outcome is alveolar bone loss and exfoliation of the involved teeth. Traditional periodontal diagnostic methods include assessment of clinical parameters and radiographs. Though efficient, these conventional techniques are inherently limited in that only a historical perspective, not current appraisal, of disease status can be determined. Advances in the use of oral fluids as possible biological samples for objective measures of current disease state, treatment monitoring, and prognostic indicators have boosted saliva and other oral-based fluids to the forefront of technology. Oral fluids contain locally and systemically derived mediators of periodontal disease, including microbial, host-response, and bone-specific resorptive markers. Although most biomarkers in oral fluids represent inflammatory mediators, several specific collagen degradation and bone turnover-related molecules have emerged as possible measures of periodontal disease activity. Pyridinoline cross-linked carboxyterminal telopeptide (ICTP), for example, has been highly correlated with clinical features of the disease and decreases in response to intervention therapies, and has been shown to possess predictive properties for possible future disease activity. One foreseeable benefit of an oral fluidâbased periodontal diagnostic would be identification of highly susceptible individuals prior to overt disease. Timely detection and diagnosis of disease may significantly affect the clinical management of periodontal patients by offering earlier, less invasive, and more cost-effective treatment therapies.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/73247/1/annals.1384.028.pd
Automated Coronal Hole Detection using Local Intensity Thresholding Techniques
We identify coronal holes using a histogram-based intensity thresholding
technique and compare their properties to fast solar wind streams at three
different points in the heliosphere. The thresholding technique was tested on
EUV and X-ray images obtained using instruments onboard STEREO, SOHO and
Hinode. The full-disk images were transformed into Lambert equal-area
projection maps and partitioned into a series of overlapping sub-images from
which local histograms were extracted. The histograms were used to determine
the threshold for the low intensity regions, which were then classified as
coronal holes or filaments using magnetograms from the SOHO/MDI. For all three
instruments, the local thresholding algorithm was found to successfully
determine coronal hole boundaries in a consistent manner. Coronal hole
properties extracted using the segmentation algorithm were then compared with
in situ measurements of the solar wind at 1 AU from ACE and STEREO. Our results
indicate that flux tubes rooted in coronal holes expand super-radially within 1
AU and that larger (smaller) coronal holes result in longer (shorter) duration
high-speed solar wind streams
A Comparison of the Red and Green Coronal Line Intensities at the 29 March 2006 and the 1 August 2008 Total Solar Eclipses: Considerations of the Temperature of the Solar Corona
During the total solar eclipse at Akademgorodok, Siberia, Russia, in 1 August
2008, we imaged the flash spectrum with a slitless spectrograph. We have
spectroscopically determined the duration of totality, the epoch of the 2nd and
3rd contacts and the duration of the flash spectrum (63 s during ingress and 48
s during egress). Here we compare the 2008 flash spectra with those that we
similarly obtained from the total solar eclipse of 29 March 2006, at
Kastellorizo, Greece. Any changes of the intensity of the corona emission
lines, in particularly those of [Fe X] and [Fe XIV], could give us valuable
information about the energy content of the solar corona and the temperature
distribution of the corona. The results show that the high-ionization state,
the [Fe XIV] emission line, was much weaker during the 2008 eclipse, indicating
that following the long, inactive period during the solar minimum, there was a
drop in the overall temperature of the solar corona.Comment: 10 color figures of spectra, 3 b/w figure
Markov Properties of Electrical Discharge Current Fluctuations in Plasma
Using the Markovian method, we study the stochastic nature of electrical
discharge current fluctuations in the Helium plasma. Sinusoidal trends are
extracted from the data set by the Fourier-Detrended Fluctuation analysis and
consequently cleaned data is retrieved. We determine the Markov time scale of
the detrended data set by using likelihood analysis. We also estimate the
Kramers-Moyal's coefficients of the discharge current fluctuations and derive
the corresponding Fokker-Planck equation. In addition, the obtained Langevin
equation enables us to reconstruct discharge time series with similar
statistical properties compared with the observed in the experiment. We also
provide an exact decomposition of temporal correlation function by using
Kramers-Moyal's coefficients. We show that for the stationary time series, the
two point temporal correlation function has an exponential decaying behavior
with a characteristic correlation time scale. Our results confirm that, there
is no definite relation between correlation and Markov time scales. However
both of them behave as monotonic increasing function of discharge current
intensity. Finally to complete our analysis, the multifractal behavior of
reconstructed time series using its Keramers-Moyal's coefficients and original
data set are investigated. Extended self similarity analysis demonstrates that
fluctuations in our experimental setup deviates from Kolmogorov (K41) theory
for fully developed turbulence regime.Comment: 25 pages, 9 figures and 4 tables. V3: Added comments, references,
figures and major correction
Low Complexity Regularization of Linear Inverse Problems
Inverse problems and regularization theory is a central theme in contemporary
signal processing, where the goal is to reconstruct an unknown signal from
partial indirect, and possibly noisy, measurements of it. A now standard method
for recovering the unknown signal is to solve a convex optimization problem
that enforces some prior knowledge about its structure. This has proved
efficient in many problems routinely encountered in imaging sciences,
statistics and machine learning. This chapter delivers a review of recent
advances in the field where the regularization prior promotes solutions
conforming to some notion of simplicity/low-complexity. These priors encompass
as popular examples sparsity and group sparsity (to capture the compressibility
of natural signals and images), total variation and analysis sparsity (to
promote piecewise regularity), and low-rank (as natural extension of sparsity
to matrix-valued data). Our aim is to provide a unified treatment of all these
regularizations under a single umbrella, namely the theory of partial
smoothness. This framework is very general and accommodates all low-complexity
regularizers just mentioned, as well as many others. Partial smoothness turns
out to be the canonical way to encode low-dimensional models that can be linear
spaces or more general smooth manifolds. This review is intended to serve as a
one stop shop toward the understanding of the theoretical properties of the
so-regularized solutions. It covers a large spectrum including: (i) recovery
guarantees and stability to noise, both in terms of -stability and
model (manifold) identification; (ii) sensitivity analysis to perturbations of
the parameters involved (in particular the observations), with applications to
unbiased risk estimation ; (iii) convergence properties of the forward-backward
proximal splitting scheme, that is particularly well suited to solve the
corresponding large-scale regularized optimization problem
Track D Social Science, Human Rights and Political Science
Peer Reviewedhttps://deepblue.lib.umich.edu/bitstream/2027.42/138414/1/jia218442.pd
- âŚ