70 research outputs found
Properties of Foreshocks and Aftershocks of the Non-Conservative SOC Olami-Feder-Christensen Model: Triggered or Critical Earthquakes?
Following Hergarten and Neugebauer [2002] who discovered aftershock and
foreshock sequences in the Olami-Feder-Christensen (OFC) discrete block-spring
earthquake model, we investigate to what degree the simple toppling mechanism
of this model is sufficient to account for the properties of earthquake
clustering in time and space. Our main finding is that synthetic catalogs
generated by the OFC model share practically all properties of real seismicity
at a qualitative level, with however significant quantitative differences. We
find that OFC catalogs can be in large part described by the concept of
triggered seismicity but the properties of foreshocks depend on the mainshock
magnitude, in qualitative agreement with the critical earthquake model and in
disagreement with simple models of triggered seismicity such as the Epidemic
Type Aftershock Sequence (ETAS) model [Ogata, 1988]. Many other features of OFC
catalogs can be reproduced with the ETAS model with a weaker clustering than
real seismicity, i.e. for a very small average number of triggered earthquakes
of first generation per mother-earthquake.Comment: revtex, 19 pages, 8 eps figure
Generation-by-Generation Dissection of the Response Function in Long Memory Epidemic Processes
In a number of natural and social systems, the response to an exogenous shock
relaxes back to the average level according to a long-memory kernel with . In the presence of an epidemic-like
process of triggered shocks developing in a cascade of generations at or close
to criticality, this "bare" kernel is renormalized into an even slower decaying
response function . Surprisingly, this means that the
shorter the memory of the bare kernel (the larger ), the longer the
memory of the response function (the smaller ). Here, we present a
detailed investigation of this paradoxical behavior based on a
generation-by-generation decomposition of the total response function, the use
of Laplace transforms and of "anomalous" scaling arguments. The paradox is
explained by the fact that the number of triggered generations grows
anomalously with time at so that the contributions of active
generations up to time more than compensate the shorter memory associated
with a larger exponent . This anomalous scaling results fundamentally
from the property that the expected waiting time is infinite for . The techniques developed here are also applied to the case
and we find in this case that the total renormalized response is a {\bf
constant} for followed by a cross-over to
for .Comment: 27 pages, 4 figure
Limitations of rupture forecasting exposed by instantaneously triggered earthquake doublet
Earthquake hazard assessments and rupture forecasts are based on the potential length of seismic rupture and whether or not slip is arrested at fault segment boundaries. Such forecasts do not generally consider that one earthquake can trigger a second large event, near-instantaneously, at distances greater than a few kilometers. Here we present a geodetic and seismological analysis of a magnitude 7.1 intra-continental earthquake that occurred in Pakistan in 1997. We find that the earthquake, rather than a single event as hitherto assumed, was in fact an earthquake doublet: initial rupture on a shallow, blind 2 reverse fault was followed just 19 seconds later by a second rupture on a separate reverse fault 50 km away. Slip on the second fault increased the total seismic moment by half, and doubled both the combined event duration and the area of maximum ground shaking. We infer that static Coulomb stresses at the initiation location of the second earthquake were probably reduced as a result of the first. Instead, we suggest that a dynamic triggering mechanism is likely, although the responsible seismic wave phase is unclear. Our results expose a flaw in earthquake rupture forecasts that disregard cascading, multiple-fault ruptures of this type
Is the northern high-latitude land-based CO2 sink weakening?
Author Posting. © American Geophysical Union, 2011. This article is posted here by permission of American Geophysical Union for personal use, not for redistribution. The definitive version was published in Global Biogeochemical Cycles 25 (2011): GB3018, doi:10.1029/2010GB003813.Studies indicate that, historically, terrestrial ecosystems of the northern high-latitude region may have been responsible for up to 60% of the global net land-based sink for atmospheric CO2. However, these regions have recently experienced remarkable modification of the major driving forces of the carbon cycle, including surface air temperature warming that is significantly greater than the global average and associated increases in the frequency and severity of disturbances. Whether Arctic tundra and boreal forest ecosystems will continue to sequester atmospheric CO2 in the face of these dramatic changes is unknown. Here we show the results of model simulations that estimate a 41 Tg C yrâ1 sink in the boreal land regions from 1997 to 2006, which represents a 73% reduction in the strength of the sink estimated for previous decades in the late 20th century. Our results suggest that CO2 uptake by the region in previous decades may not be as strong as previously estimated. The recent decline in sink strength is the combined result of (1) weakening sinks due to warming-induced increases in soil organic matter decomposition and (2) strengthening sources from pyrogenic CO2 emissions as a result of the substantial area of boreal forest burned in wildfires across the region in recent years. Such changes create positive feedbacks to the climate system that accelerate global warming, putting further pressure on emission reductions to achieve atmospheric stabilization targets.This study was supported through grants
provided as part of the Arctic System Science Program (NSF OPPâ
0531047), the North American Carbon Program (NASA NNG05GD25G),
and the Bonanza Creek LongâTerm Ecological Program (funded jointly by
NSF grant DEBâ0423442 and USDA Forest Service, Pacific Northwest
Research Station grant PNW01âJV11261952â231)
Future effects of ozone on carbon sequestration and climate change policy using a global biogeochemical model
Author Posting. © The Authors, 2004. This is the author's version of the work. It is posted here by permission of Springer for personal use, not for redistribution. The definitive version was published in Climatic Change 73 (2005): 345-373, doi:10.1007/s10584-005-6776-4.Exposure of plants to ozone inhibits photosynthesis and therefore reduces vegetation production and carbon sequestration. The reduced carbon storage would then require further reductions in fossil fuel emissions to meet a given CO2 concentration target, thereby increasing the cost of meeting the target. Simulations with the Terrestrial Ecosystem Model (TEM) for the historical period (1860-1995) show the largest damages occur in the Southeast and Midwestern regions of the United States, eastern Europe, and eastern China. The largest reductions in carbon storage for the period 1950-1995, 41%, occur in eastern Europe. Scenarios for the 21st century developed with the MIT Integrated Global Systems Model (IGSM) lead to even greater negative effects on carbon storage in the future. In some regions, current land carbon sinks become carbon sources, and this change leads to carbon sequestration decreases of up to 0.4 Pg C yr-1 due to damage in some regional ozone hot spots. With a climate policy, failing to consider the effects of ozone damage on carbon sequestration would raise the global costs over the next century of stabilizing atmospheric concentrations of CO2 equivalents at 550 ppm by 6 to 21%. Because stabilization at 550 ppm will reduce emission of other gases that cause ozone, these additional benefits are estimated to be between 5 and 25% of the cost of the climate policy. Tropospheric ozone effects on terrestrial ecosystems thus produce a surprisingly large feedback in estimating climate policy costs that, heretofore, has not been included in cost estimates.This study was funded by the Biocomplexity Program of the U.S. National Science Foundation (ATM-0120468), the Methods and Models for Integrated Assessment Program of the U.S. National Science Foundation (DEB-9711626) and the Earth Observing System Program of the U.S. National Aeronautics and Space Administration (NAG5-10135). The IGSM has been developed as part of the Joint Program on the Science and Policy of Global Change with the support of a government-industry partnership including in addition to the above the US Department of Energy (901214-HAR; DE-FG02-94ER61937; DE-FG0293ER61713), the US Environmental Protection Agency (X-827703-01-0; XA-83042801-0), the National Aeronautics and Atmospheric Administration (NA16GP2290) and a group of corporate sponsors from the United States, Japan, United Kingdom, Germany, France, and Norway
Recommended from our members
Uniform California Earthquake Rupture Forecast, Version 3 (UCERF3) âThe Time-Independent Model
The 2014 Working Group on California Earthquake Probabilities
(WGCEP14) present the time-independent component of the Uniform California
Earthquake Rupture Forecast, Version 3 (UCERF3), which provides authoritative
estimates of the magnitude, location, and time-averaged frequency of potentially
damaging earthquakes in California. The primary achievements have been to relax
fault segmentation and include multifault ruptures, both limitations of UCERF2.
The rates of all earthquakes are solved for simultaneously and from a broader range
of data, using a system-level inversion that is both conceptually simple and extensible.
The inverse problem is large and underdetermined, so a range of models is
sampled using an efficient simulated annealing algorithm. The approach is more
derivative than prescriptive (e.g., magnitudeâfrequency distributions are no longer
assumed), so new analysis tools were developed for exploring solutions. Epistemic
uncertainties were also accounted for using 1440 alternative logic-tree branches,
necessitating access to supercomputers. The most influential uncertainties include
alternative deformation models (fault slip rates), a new smoothed seismicity algorithm,
alternative values for the total rate of M[subscript w] â„ 5 events, and different scaling
relationships, virtually all of which are new. As a notable first, three deformation
models are based on kinematically consistent inversions of geodetic and geologic
data, also providing slip-rate constraints on faults previously excluded due to lack
of geologic data. The grand inversion constitutes a system-level framework for
testing hypotheses and balancing the influence of different experts. For example,
we demonstrate serious challenges with the GutenbergâRichter hypothesis for
individual faults. UCERF3 is still an approximation of the system, however, and
the range of models is limited (e.g., constrained to stay close to UCERF2). Nevertheless,
UCERF3 removes the apparent UCERF2 overprediction of M 6.5â7 earthquake
rates and also includes types of multifault ruptures seen in nature. Although
UCERF3 fits the data better than UCERF2 overall, there may be areas that warrant
further site-specific investigation. Supporting products may be of general interest,
and we list key assumptions and avenues for future model improvements
Recommended from our members
Long-Term Time-Dependent Probabilities for the Third Uniform California Earthquake Rupture Forecast (UCERF3)
The 2014 Working Group on California Earthquake Probabilities (WGCEP 2014) presents time-dependent earthquake probabilities for the third Uniform California Earthquake Rupture Forecast (UCERF3). Building on the UCERF3 time-independent model published previously, renewal models are utilized to represent elastic-rebound-implied probabilities. A new methodology has been developed that solves applicability issues in the previous approach for unsegmented models. The new methodology also supports magnitude-dependent aperiodicity and accounts for the historic open interval on faults that lack a date-of-last-event constraint. Epistemic uncertainties are represented with a logic tree, producing 5760 different forecasts. Results for a variety of evaluation metrics are presented, including logic-tree sensitivity analyses and comparisons to the previous model (UCERF2). For 30 yr M â„ 6.7 probabilities, the most significant changes from UCERF2 are a threefold increase on the Calaveras fault and a threefold decrease on the San Jacinto fault. Such changes are due mostly to differences in the time-independent models (e.g., fault-slip rates), with relaxation of segmentation and inclusion of multifault ruptures being particularly influential. In fact, some UCERF2 faults were simply too long to produce M 6.7 size events given the segmentation assumptions in that study. Probability model differences are also influential, with the implied gains (relative to a Poisson model) being generally higher in UCERF3. Accounting for the historic open interval is one reason. Another is an effective 27% increase in the total elastic-rebound-model weight. The exact factors influencing differences between UCERF2 and UCERF3, as well as the relative importance of logic-tree branches, vary throughout the region and depend on the evaluation metric of interest. For example, M â„ 6.7 probabilities may not be a good proxy for other hazard or loss measures. This sensitivity, coupled with the approximate nature of the model and known limitations, means the applicability of UCERF3 should be evaluated on a case-by-case basis
Methane fluxes between terrestrial ecosystems and the atmosphere at northern high latitudes during the past century : a retrospective analysis with a process-based biogeochemistry model
Author Posting. © American Geophysical Union, 2004. This article is posted here by permission of American Geophysical Union for personal use, not for redistribution. The definitive version was published in Global Biogeochemical Cycles 18 (2004): GB3010, doi:10.1029/2004GB002239.We develop and use a new version of the Terrestrial Ecosystem Model (TEM) to study how rates of methane (CH4) emissions and consumption in high-latitude soils of the Northern Hemisphere have changed over the past century in response to observed changes in the region's climate. We estimate that the net emissions of CH4 (emissions minus consumption) from these soils have increased by an average 0.08 Tg CH4 yrâ1 during the twentieth century. Our estimate of the annual net emission rate at the end of the century for the region is 51 Tg CH4 yrâ1. Russia, Canada, and Alaska are the major CH4 regional sources to the atmosphere, responsible for 64%, 11%, and 7% of these net emissions, respectively. Our simulations indicate that large interannual variability in net CH4 emissions occurred over the last century. Our analyses of the responses of net CH4 emissions to the past climate change suggest that future global warming will increase net CH4 emissions from the Pan-Arctic region. The higher net CH4 emissions may increase atmospheric CH4 concentrations to provide a major positive feedback to the climate system.This study was supported by a
NSF biocomplexity grant (ATM-0120468), the NASA Land Cover and
Land Use Change Program (NAG5-6257), and by funding from MIT Joint
Program on the Science and Policy of Global Change, which is supported
by a consortium of government, industry, and foundation sponsors
- âŠ