398 research outputs found
Masses and Mixings in a Grand Unified Toy Model
The generation of the fermion mass hierarchy in the standard model of
particle physics is a long-standing puzzle. The recent discoveries from
neutrino physics suggests that the mixing in the lepton sector is large
compared to the quark mixings. To understand this asymmetry between the quark
and lepton mixings is an important aim for particle physics. In this regard,
two promising approaches from the theoretical side are grand unified theories
and family symmetries. In this note we try to understand certain general
features of grand unified theories with Abelian family symmetries by taking the
simplest SU(5) grand unified theory as a prototype. We construct an SU(5) toy
model with family symmetry
that, in a natural way, duplicates the observed mass hierarchy and mixing
matrices to lowest approximation. The system for generating the mass hierarchy
is through a Froggatt-Nielsen type mechanism. One idea that we use in the model
is that the quark and charged lepton sectors are hierarchical with small mixing
angles while the light neutrino sector is democratic with larger mixing angles.
We also discuss some of the difficulties in incorporating finer details into
the model without making further assumptions or adding a large scalar sector.Comment: 21 pages, 2 figures, RevTeX, v2: references updated and typos
corrected, v3: updated top quark mass, comments on MiniBooNE result, and
typos correcte
Promoting motor skills in low-income, ethnic children: The Physical Activity in Linguistically Diverse Communities (PALDC) nonrandomized trial
This study reports the long-term effects of a professional learning program for classroom teachers on fundamental motor skill (FMS) proficiency of primary school students from ethnically diverse backgrounds. Design: A cluster non-randomized trial using a nested cross-sectional design. Methods: The study was conducted in 8 primary schools located in disadvantaged and culturally diverse areas in Sydney, Australia. The intervention used an action learning framework, with each school developing and implementing an action plan for enhancing the teaching of FMS in their school. School teams comprised 4-5 teachers and were supported by a member of the research team. The primary outcome was total proficiency score for 7 FMS (run, jump, catch, throw, kick, leap, side gallop). Outcome data were analyzed using mixed effects models. Results: Eight-hundred and sixty-two students (82% response rate) were assessed at baseline in 2006 and 830 (82%) at follow-up in 2010. Compared with students in the control schools, there was a significantly greater increase in total motor skill proficiency among children in the intervention schools at follow-up (adjusted difference = 5.2 components, 95%CI [1.65, 8.75]; p. = 0.01) and in four of the seven motor skills. Conclusions: Training classroom teachers to develop and implement units of work based around individual FMS is a promising strategy for increasing FMS among ethnically diverse children over an extended period of time
Quantification of Spatial Parameters in 3D Cellular Constructs Using Graph Theory
Multispectral three-dimensional (3D) imaging provides spatial information for biological structures that cannot be measured by traditional methods. This work presents a method of tracking 3D biological structures to quantify changes over time using graph theory. Cell-graphs were generated based on the pairwise distances, in 3D-Euclidean space, between nuclei during collagen I gel compaction. From these graphs quantitative features are extracted that measure both the global topography and the frequently occurring local structures of the “tissue constructs.” The feature trends can be controlled by manipulating compaction through cell density and are significant when compared to random graphs. This work presents a novel methodology to track a simple 3D biological event and quantitatively analyze the underlying structural change. Further application of this method will allow for the study of complex biological problems that require the quantification of temporal-spatial information in 3D and establish a new paradigm in understanding structure-function relationships
Recommended from our members
Statistical inference of OH concentrations and air mass dilution rates from successive observations of non-methane hydrocarbons in single air masses
Bayesian inference has been used to determine rigorous estimates of hydroxyl radical concentrations () and air mass dilution rates (K) averaged following air masses between linked observations of nonmethane hydrocarbons (NMHCs) spanning the North Atlantic during the Intercontinental Transport and Chemical Transformation (ITCT)-Lagrangian-2K4 experiment. The Bayesian technique obtains a refined (posterior) distribution of a parameter given data related to the parameter through a model and prior beliefs about the parameter distribution. Here, the model describes hydrocarbon loss through OH reaction and mixing with a background concentration at rate K. The Lagrangian experiment provides direct observations of hydrocarbons at two time points, removing assumptions regarding composition or sources upstream of a single observation. The estimates are sharpened by using many hydrocarbons with different reactivities and accounting for their variability and measurement uncertainty. A novel technique is used to construct prior background distributions of many species, described by variation of a single parameter . This exploits the high correlation of species, related by the first principal component of many NMHC samples. The Bayesian method obtains posterior estimates of , K and following each air mass. Median values are typically between 0.5 and 2.0 × 106 molecules cm−3, but are elevated to between 2.5 and 3.5 × 106 molecules cm−3, in low-level pollution. A comparison of estimates from absolute NMHC concentrations and NMHC ratios assuming zero background (the “photochemical clock” method) shows similar distributions but reveals systematic high bias in the estimates from ratios. Estimates of K are ∼0.1 day−1 but show more sensitivity to the prior distribution assumed
Recommended from our members
Sources and characteristics of summertime organic aerosol in the Colorado Front Range: perspective from measurements and WRF-Chem modeling
Abstract. The evolution of organic aerosols (OAs) and their precursors in the boundary layer (BL) of the Colorado Front Range during the Front Range Air Pollution and Photochemistry Éxperiment (FRAPPÉ, July–August 2014) was analyzed by in situ measurements and chemical transport modeling. Measurements indicated significant production of secondary OA (SOA), with enhancement ratio of OA with respect to carbon monoxide (CO) reaching 0.085±0.003 µg m−3 ppbv−1. At background mixing ratios of CO, up to ∼ 1.8 µg m−3 background OA was observed, suggesting significant non-combustion contribution to OA in the Front Range. The mean concentration of OA in plumes with a high influence of oil and natural gas (O&G) emissions was ∼ 40 % higher than in urban-influenced plumes. Positive matrix factorization (PMF) confirmed a dominant contribution of secondary, oxygenated OA (OOA) in the boundary layer instead of fresh, hydrocarbon-like OA (HOA). Combinations of primary OA (POA) volatility assumptions, aging of semi-volatile species, and different emission estimates from the O&G sector were used in the Weather Research and Forecasting model coupled with Chemistry (WRF-Chem) simulation scenarios. The assumption of semi-volatile POA resulted in greater than a factor of 10 lower POA concentrations compared to PMF-resolved HOA. Including top-down modified O&G emissions resulted in substantially better agreements in modeled ethane, toluene, hydroxyl radical, and ozone compared to measurements in the high-O&G-influenced plumes. By including emissions from the O&G sector using the top-down approach, it was estimated that the O&G sector contributed to < 5 % of total OA, but up to 38 % of anthropogenic SOA (aSOA) in the region. The best agreement between the measured and simulated median OA was achieved by limiting the extent of biogenic hydrocarbon aging and consequently biogenic SOA (bSOA) production. Despite a lower production of bSOA in this scenario, contribution of bSOA to total SOA remained high at 40–54 %. Future studies aiming at a better emissions characterization of POA and intermediate-volatility organic compounds (IVOCs) from the O&G sector are valuable
Assignment of weight-based antibody units for four additional serotypes to a human anti-pneumococcal standard reference serum 007sp
The pneumococcal ELISA reference standard serum, Lot 89SF, has been in use since 1990 and was replaced with a new reference standard serum, 007sp in 2013. This serum was generated under an FDA-approved clinical protocol, where 278 adult volunteers were immunized with the 23-valent unconjugated polysaccharide vaccine, Pneumovax II®, and a unit of blood was obtained twice within 120 days following immunization. Pooled serum was prepared from the plasma, filled at 6ml per vial and lyophilised. Five independent laboratories participated in bridging the serotype specific IgG assignments for 89SF to 007sp to establish equivalent reference values for 13 pneumococcal capsular serotypes (1,3, 4, 5, 6A, 6B, 7F, 9V, 14, 18C, 19A, 19F, 23F) using the WHO reference ELISA. A subsequent follow up study established equivalent reference values for an additional seven serotypes (8, 10A, 11A, 12F, 15B, 22F, 33F). In this study three laboratories assigned weight-based IgG concentrations in mcg/mL of serum to 007sp for four additional serotypes; 2, 9N, 17F and 20A. This study completes the assignment of serotypes in 89SF to 007sp. In addition, the IgG antibody assignments for a 12 member WHO QC serum panel were extended to cover the four additional serotypes. Agreement was excellent with a concordance correlation coefficient (rc) > 0.996 when each laboratory was compared to the assigned values for the 12 WHO QC sera. The 007sp preparation has replaced 89SF as the pneumococcal reference standard. Sufficient quantities of 007sp are projected to be available for the next 25 years
How the recent BABAR data for P to \gamma\gamma* affect the Standard Model predictions for the rare decays P to l+l-
Measuring the lepton anomalous magnetic moments and the rare decays
of light pseudoscalar mesons into lepton pairs , serve as
important tests of the Standard Model. To reduce the theoretical uncertainty in
the standard model predictions, the data on the charge and transition form
factors of the light pseudoscalar mesons play a significant role. Recently, new
data on the behavior of the transition form factors at
large momentum transfer were supplied by the BABAR collaboration. There are
several problems with the theoretical interpretation of these data: 1) An
unexpectedly slow decrease of the pion transition form factor at high momenta,
2) the qualitative difference in the behavior of the pion form factor and the
and form factors at high momenta, 3) the inconsistency of
the measured ratio of the and form factors with the
predicted one. We comment on the influence of the new BABAR data on the rare
decay branchings.Comment: 11 pages, 3 figure
- …