4,109 research outputs found
Modelling distributed lag effects in epidemiological time series studies
The paper argues that much of the existing literature on air pollution and mortality deals only with the transient effects of air pollution. Policy, on the other hand, needs to know when, whether and to what extent pollution-induced increases in mortality are reversed. This involves modelling the entire distributed lag effects of air pollution. Borrowing from econometrics this paper presents a method by which distributed lag effects can be estimated parsimoniously but plausibly estimated. The paper
presents a time series study into the relationship between ambient levels of air pollution and daily mortality counts for Manchester employing this technique. Black Smoke is shown to have a highly significant effect on mortality counts in the short term. Nevertheless we find that 80 percent of the deaths attributable to BS would have occurred anyway within one week whereas the remaining 20 percent of individuals would otherwise have enjoyed a normal life expectancy
The amenity value of the Italian climate
The hedonic price literature suggests that locations with more favourable characteristics should display compensating wage and house price differentials.
Estimates of the marginal willingness to pay for small changes in climate variables are derived using the hedonic price technique applied to Italian data. A hedonic price model was specified in terms of January and July averages. There exists considerable empirical support for the hypothesis that amenity values for climate are embedded in the labour and housing market. Italians would prefer a
drier climate during the winter months, but higher summertime temperatures are shown to reduce welfare. These results may have relevance to the task of determining the economic impact of future climate change
Modelling distributed lag effects in mortality and air pollution studies: the case of Santiago
Most of the epidemiological literature on air pollution and mortality deals only with single or dual pollutant models whose results are hard to interpret and of questionable value from the policy perspective. In addition, much of the existing literature deals only with the very short-term effects of air pollution whereas policy makers need to know when, whether and to what extent pollution-induced
increases in mortality counts are reversed. This involves modelling the infinite distributed lag effects of air pollution.
Borrowing from econometrics this paper presents a method by which the infinite distributed lag effects can be estimated parsimoniously but plausibly estimated. The paper presents a time series study into the relationship between ambient
levels of air pollution and daily mortality counts for Santiago employing this technique which confirms that the infinite lag effects are highly significant.
It is also shown that day to day variations in NO2 concentrations and in the concentrations of both fine and coarse particulates are associated with short-term
variations in death rates. These findings are made in the context of a model that simultaneously includes six different pollutants. Evidence is found pointing to the operation of a very short term harvesting effect
Valuing congestion costs in the British Museum
Museums are potentially congestible resources because the exhibits they contain are, in any relevant sense of the word, irreproducible. Insofar as visitor congestion
diminishes the value of individuals’ visits it constitutes an additional reason for charging for admission to museums, albeit one not previously considered. A policy of
free access to a museum containing unique treasures may dissipate the economic benefits of the museum.
Within the context of an empirical study undertaken for the British Museum using stated preference techniques it is shown that the congestion cost posed by the marginal visitor is quite high. Notwithstanding the argument that visits to the museum may possess external benefits, this points to the desirability of instigating charges for admission. Furthermore, it is shown that the marginal congestion cost decreases at least over a range as visitor numbers increase. In other words beyond certain levels introducing more visitors does not worsen congestion. This suggests that, contrary to what is often assumed, charging more during periods of high demand may be undesirable.
Insofar as congestion is a widespread phenomenon in important museums, galleries and sites of historical heritage the issues raised in this paper as well as the
methodology devised to determine congestion costs could have widespread application
Galactic Cannibalism: the Origin of the Magellanic Stream
We are in a privileged location in the Universe which allows us to observe
galactic interactions from close range -- the merger of our two nearest dwarf
satellite galaxies, the LMC and SMC. It is important to understand the local
merger process before we can have confidence in understanding mergers at high
redshift. We present high resolution Nbody+SPH simulations of the disruption of
the LMC and SMC and the formation of the Magellanic Stream, and discuss the
implications for galaxy formation and evolution.Comment: 2 pages, 1 figure, to appear in "The Evolution of Galaxies II: Basic
Building Blocks", (2002) ed. M. Sauvage et al. (Kluwer
High-resolution N-body Simulations of Galactic Cannibalism: The Magellanic Stream
Hierarchical clustering represents the favoured paradigm for galaxy formation
throughout the Universe; due to its proximity, the Magellanic system offers one
of the few opportunities for astrophysicists to decompose the full
six-dimensional phase-space history of a satellite in the midst of being
cannibalised by its host galaxy. The availability of improved observational
data for the Magellanic Stream and parallel advances in computational power has
led us to revisit the canonical tidal model describing the disruption of the
Small Magellanic Cloud and the consequent formation of the Stream. We suggest
improvements to the tidal model in light of these recent advances.Comment: 6 pages, 4 figures, LaTeX (gcdv.sty). Refereed contribution to the
5th Galactic Chemodynamics conference held in Swinburne, July 2003. Accepted
for publication in PASA. Version with high resolution figures available at
http://astronomy.swin.edu.au/staff/tconnors/publications.htm
Implementation of a geometrically informed and energetically constrained mesoscale eddy parameterization in an ocean circulation model
The global stratification and circulation of the ocean and their
sensitivities to changes in forcing depend crucially on the representation of
the mesoscale eddy field. Here, a geometrically informed and energetically
constrained parameterization framework for mesoscale eddies --- termed
GEOMETRIC --- is proposed and implemented in three-dimensional primitive
equation channel and sector models. The GEOMETRIC framework closes mesoscale
eddy fluxes according to the standard Gent--McWilliams scheme, but with the
eddy transfer coefficient constrained by the depth-integrated eddy energy
field, provided through a prognostic eddy energy budget evolving with the mean
state. It is found that coarse resolution calculations employing GEOMETRIC
broadly reproduce model sensitivities of the eddy permitting reference
calculations in the emergent circumpolar transport, meridional overturning
circulation profile and the depth-integrated eddy energy signature; in
particular, eddy saturation emerges in the sector configuration. Some
differences arise, attributed here to the simple prognostic eddy energy budget
employed, to be improved upon in future investigations. The GEOMETRIC framework
thus proposes a shift in paradigm, from a focus on how to close for eddy
fluxes, to focusing on the representation of eddy energetics.Comment: 19 pages, 9 figures, submitted to Journal of Physical Oceanography;
comments welcome. (Copyright statement: see section 7a of
https://www.ametsoc.org/ams/index.cfm/publications/ethical-guidelines-and-ams-policies/ams-copyright-policy/
Uneconomical Diagnosis of Cladograms: Comments on Wheeler and Nixon's Method for Sankoff Optimization
Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/74972/1/j.1096-0031.1997.tb00249.x.pd
Contrastive Learning Can Find An Optimal Basis For Approximately View-Invariant Functions
Contrastive learning is a powerful framework for learning self-supervised
representations that generalize well to downstream supervised tasks. We show
that multiple existing contrastive learning methods can be reinterpreted as
learning kernel functions that approximate a fixed positive-pair kernel. We
then prove that a simple representation obtained by combining this kernel with
PCA provably minimizes the worst-case approximation error of linear predictors,
under a straightforward assumption that positive pairs have similar labels. Our
analysis is based on a decomposition of the target function in terms of the
eigenfunctions of a positive-pair Markov chain, and a surprising equivalence
between these eigenfunctions and the output of Kernel PCA. We give
generalization bounds for downstream linear prediction using our Kernel PCA
representation, and show empirically on a set of synthetic tasks that applying
Kernel PCA to contrastive learning models can indeed approximately recover the
Markov chain eigenfunctions, although the accuracy depends on the kernel
parameterization as well as on the augmentation strength.Comment: Published at ICLR 202
Efficient FPT algorithms for (strict) compatibility of unrooted phylogenetic trees
In phylogenetics, a central problem is to infer the evolutionary
relationships between a set of species ; these relationships are often
depicted via a phylogenetic tree -- a tree having its leaves univocally labeled
by elements of and without degree-2 nodes -- called the "species tree". One
common approach for reconstructing a species tree consists in first
constructing several phylogenetic trees from primary data (e.g. DNA sequences
originating from some species in ), and then constructing a single
phylogenetic tree maximizing the "concordance" with the input trees. The
so-obtained tree is our estimation of the species tree and, when the input
trees are defined on overlapping -- but not identical -- sets of labels, is
called "supertree". In this paper, we focus on two problems that are central
when combining phylogenetic trees into a supertree: the compatibility and the
strict compatibility problems for unrooted phylogenetic trees. These problems
are strongly related, respectively, to the notions of "containing as a minor"
and "containing as a topological minor" in the graph community. Both problems
are known to be fixed-parameter tractable in the number of input trees , by
using their expressibility in Monadic Second Order Logic and a reduction to
graphs of bounded treewidth. Motivated by the fact that the dependency on
of these algorithms is prohibitively large, we give the first explicit dynamic
programming algorithms for solving these problems, both running in time
, where is the total size of the input.Comment: 18 pages, 1 figur
- …