951 research outputs found
GPS source solution of the 2004 Parkfield earthquake
We compute a series of finite-source parameter inversions of the fault
rupture of the 2004 Parkfield earthquake based on 1 Hz GPS records only. We
confirm that some of the co-seismic slip at shallow depth (<5 km) constrained
by InSAR data processing results from early post-seismic deformation. We also
show 1) that if located very close to the rupture, a GPS receiver can saturate
while it remains possible to estimate the ground velocity (~1.2 m/s) near the
fault, 2) that GPS waveforms inversions constrain that the slip distribution at
depth even when GPS monuments are not located directly above the ruptured areas
and 3) the slip distribution at depth from our best models agree with that
recovered from strong motion data. The 95th percentile of the slip amplitudes
for rupture velocities ranging from 2 to 5 km/s is, 55 +/- 6 cm.Comment: 24 pages including supp. material
Source parameters of the 23 April 1992 M 6.1 Joshua Tree, California, earthquake and its aftershocks: Empirical Green's function analysis of GEOS and TERRAscope data
Source parameters of the M 6.1 23 April 1992 Joshua Tree mainshock and 86 M 1.8 to 4.9 aftershocks are determined using an empirical Green's function methodology. For the aftershocks, deconvolved P- and S-wave spectra are calculated for 126 pairs of closely spaced events recorded on portable GEOS stations; S-wave spectra from the two horizontal components are averaged. The deconvolved spectra are fit by a ratio of omega-square source models, yielding an optimal (least-squares) corner frequency for both the large and the small event in each pair. We find no resolved difference between the inferred P- and S-wave corner frequencies. Using the standard Brune model for stress drop, we also find no resolved nonconstant scaling of stress drop with moment, although we also conclude that detailed scaling systematics would be difficult to resolve. In particular, a weak increase of stress drop with moment over a limited moment/magnitude cannot be ruled out. For magnitudes smaller than M 3 to 3.5, the inferred stress-drop values will be limited by the maximum observable corner frequency value of 60 Hz. For the mainshock, source-time functions are obtained from mainshock recordings at three TERRAscope stations (PFO, PAS, and GSC) using an M 4.3 foreshock as an empirical Green's function. The results indicate a fairly simple, single-pulse source-time function, with clear south-to-north directivity and an inferred rupture radius of 5 to 6 km. The deconvolved source-time functions are inverted to obtain a finite-rupture model that gives a robust estimate of rupture dimension. Early aftershocks are found to lie along the perimeters of regions with high mainshock slip. The inferred mainshock stress-drop value, 56 bars, is within the range determined for the aftershocks. Our derived mainshock source spectra do not show resolvable deviation from the omega-square model
Near real time regional moment tensor estimation using italian broadband stations
Since 2002, the Istituto Nazionale di Geofisica e Vulcanologia (INGV) in Rome has started the installation of a high quality
regional broadband network throughout the Italian territory. Up today, the network consists of 125 stations equipped with 40 s
natural period instruments. The dense station coverage allows for the implementation of real-time regional moment tensor (MT)
estimation procedures such as that proposed by Dreger and Helmberger (1993).
The automatic MT algorithm uses real-time broadband waveforms continuously telemetered to INGV, and it is triggered for events
with magnitude greater than Ml 3.5. This is the lowermost value for which we have found it possible to obtain reliable MT determination
in the frequency band used in the inversion. The automatic solution is available within about 3-5 minutes after the earthquake
location. Each solution has an assigned quality factor dependent on the number of the station used in the inversion, and the godness
of fit between synthetic and observed data. MT is published on the web after revision by a seismologist.
Efforts are also made to evaluate MT solutions for earthquakes occurring in Italy and neighboring regions in the last years. The
results are compared to those obtained from application of other moment tensor methods. It is always found a good agreement
between the newly determined solutions and those from other methods.
Overall, fast and accurate moment tensor solutions are an important ingredient when attempting to estimate the recorded ground
shaking. Overall, in Italy, earthquakes in the magnitude range 3.5 – 5 are very common; the availability of their focal mechanisms
allows the mapping of the principal stress field axes leading to a better understanding of the ongoing tectonics
How to keep drivers engaged while supervising driving automation? A literature survey and categorization of six solution areas
This work aimed to organise recommendations for keeping people engaged during human supervision of driving automation, encouraging a safe and acceptable introduction of automated driving systems. First, heuristic knowledge of human factors, ergonomics, and psychological theory was used to propose solution areas to human supervisory control problems of sustained attention. Driving and non-driving research examples were drawn to substantiate the solution areas. Automotive manufacturers might (1) avoid this supervisory role altogether, (2) reduce it in objective ways or (3) alter its subjective experiences, (4) utilize conditioning learning principles such as with gamification and/or selection/training techniques, (5) support internal driver cognitive processes and mental models and/or (6) leverage externally situated information regarding relations between the driver, the driving task, and the driving environment. Second, a cross-domain literature survey of influential human-automation interaction research was conducted for how to keep engagement/attention in supervisory control. The solution areas (via numeric theme codes) were found to be reliably applied from independent rater categorisations of research recommendations. Areas (5) and (6) were addressed by around 70% or more of the studies, areas (2) and (4) in around 50% of the studies, and areas (3) and (1) in less than around 20% and 5%, respectively. The present contribution offers a guiding organisational framework towards improving human attention while supervising driving automation.submittedVersio
Recommended from our members
Berkeley Seismological Laboratory Seismic Moment Tensor Report for the August 6, 2007 M3.9 Seismic event in central Utah
We have performed a complete moment tensor analysis of the seismic event, which occurred on Monday August 6, 2007 at 08:48:40 UTC 21 km from Mt.Pleasant, Utah. In our analysis we utilized complete three-component seismic records recorded by the USArray, University of Utah, and EarthScope seismic arrays. The seismic waveform data was integrated to displacement and filtered between 0.02 to 0.10 Hz following instrument removal. We used the Song et al. (1996) velocity model to compute Green's functions used in the moment tensor inversion. A map of the stations we used and the location of the event is shown in Figure 1. In our moment tensor analysis we assumed a shallow source depth of 1 km consistent with the shallow depth reported for this event. As shown in Figure 2 the results point to a source mechanism with negligible double-couple radiation and is composed of dominant CLVD and implosive isotropic components. The total scalar seismic moment is 2.12e22 dyne cm corresponding to a moment magnitude (Mw) of 4.2. The long-period records are very well matched by the model (Figure 2) with a variance reduction of 73.4%. An all dilational (down) first motion radiation pattern is predicted by the moment tensor solution, and observations of first motions are in agreement
CD4+CD25highCD127low Regulatory T Cells in Peripheral Blood Are Not an Independent Factor for Chronic Graft-versus-Host Disease after Allogeneic Stem Cell Transplantation
Background. The therapeutic efficacy of allogeneic hemopoietic stem cell transplantation (HSCT) largely relies on the graft-versus-leukemia (GVL) effect. Uncontrolled graft-versus-host disease (GVHD) is a feared complication of HSCT. Regulatory T cells (Treg) are a subset of CD4+ T-helper cells believed to maintain tolerance after HSCT. It remains unclear whether low peripheral blood Treg have an impact on the risk for acute (aGVHD) and chronic GVHD (cGVHD). Methods. In this paper we enumerated the CD4+CD25highCD127low Treg in the peripheral blood of 84 patients after at least 150 days from HSCT and in 20 healthy age-matched controls. Results. Although similar mean lymphocyte counts were found in patients and controls, CD3+CD4+ T-cell counts were significantly lower in patients. Patients also had significantly lower Treg percentages among lymphocytes as compared to controls. Patients with cGVHD had even higher percentages of Treg if compared to patients without cGVHD. In multivariate analysis, Treg percentages were not an independent factor for cGVHD. Conclusions. This paper did not show a relation between deficient peripheral blood Treg and cGVHD, therefore cGVHD does not seem to occur as a result of peripheral Treg paucity
Recommended from our members
Identifying Isotropic Events Using an Improved Regional Moment Tensor Inversion Technique
Using a regional time-domain waveform inversion for the complete moment tensor we calculate the deviatoric and isotropic source components for several explosions at the Nevada Test Site as well as earthquakes, and collapses in the surrounding region of the western US. The events separate into specific populations according to their deviation from a pure double-couple and ratio of isotropic to deviatoric energy. The separation allows for anomalous event identification and discrimination between explosions, earthquakes, and collapses. Error in the moment tensor solutions and source parameters is also calculated. We investigate the sensitivity of the moment tensor solutions to Green's functions calculated with imperfect Earth models, inaccurate event locations, and data with a low signal-to-noise ratio. We also test the performance of the method under a range of recording conditions from excellent azimuthal coverage to cases of sparse station availability, as might be expected for smaller events. Finally, we assess the depth and frequency dependence upon event size. This analysis will be used to determine the range where well-constrained solutions can be obtained
Microseismic joint location and anisotropic velocity inversion for hydraulic fracturing in a tight Bakken reservoir
To improve the accuracy of microseismic event locations, we developed a new inversion method with double-difference constraints for determining the hypocenters and the anisotropic velocity model for unconventional reservoirs. We applied this method to a microseismic data set monitoring a Middle Bakken completion in the Beaver Lodge area of North Dakota. Geophone arrays in four observation wells improved the ray coverage for the velocity inversion. Using an accurate anisotropic velocity model is important to correctly assess the height growth of the hydraulically induced fractures in the Middle Bakken. Our results showed that (1) moderate-to-strong anisotropy exists in all studied sedimentary layers, especially in the Upper and Lower Bakken shale formations, where the Thomsen parameters (ϵ and γ) can be greater than 0.4, (2) all the events selected for high signal-to-noise ratio and used for the joint velocity inversion are located in the Bakken and overlying Lodgepole formations, i.e., no events are detected in the Three Forks formation below the Bakken, and (3) more than half of the strong events are in two clusters at approximately 100 and 150 m above the Middle Bakken. Reoccurrence of strong, closely clustered events suggested activation of natural fractures or faults in the Lodgepole formation. The sensitivity analysis for the inversion results showed that the relative uncertainty in parameter δ is larger than other anisotropy parameters. The microseismic event locations and the anisotropic velocity model are validated by comparing synthetic and observed seismic waveforms and by S-wave splitting.Shell Oil Compan
Recommended from our members
2-D or not 2-D, that is the question: A Northern California test
Reliable estimates of the seismic source spectrum are necessary for accurate magnitude, yield, and energy estimation. In particular, how seismic radiated energy scales with increasing earthquake size has been the focus of recent debate within the community and has direct implications on earthquake source physics studies as well as hazard mitigation. The 1-D coda methodology of Mayeda et al. has provided the lowest variance estimate of the source spectrum when compared against traditional approaches that use direct S-waves, thus making it ideal for networks that have sparse station distribution. The 1-D coda methodology has been mostly confined to regions of approximately uniform complexity. For larger, more geophysically complicated regions, 2-D path corrections may be required. The complicated tectonics of the northern California region coupled with high quality broadband seismic data provides for an ideal ''apples-to-apples'' test of 1-D and 2-D path assumptions on direct waves and their coda. Using the same station and event distribution, we compared 1-D and 2-D path corrections and observed the following results: (1) 1-D coda results reduced the amplitude variance relative to direct S-waves by roughly a factor of 8 (800%); (2) Applying a 2-D correction to the coda resulted in up to 40% variance reduction from the 1-D coda results; (3) 2-D direct S-wave results, though better than 1-D direct waves, were significantly worse than the 1-D coda. We found that coda-based moment-rate source spectra derived from the 2-D approach were essentially identical to those from the 1-D approach for frequencies less than {approx}0.7-Hz, however for the high frequencies (0.7{le} f {le} 8.0-Hz), the 2-D approach resulted in inter-station scatter that was generally 10-30% smaller. For complex regions where data are plentiful, a 2-D approach can significantly improve upon the simple 1-D assumption. In regions where only 1-D coda correction is available it is still preferable over 2-D direct wave-based measures
- …