1,067,458 research outputs found

    Is the magnitude of the Peccei-Quinn scale set by the landscape?

    Full text link
    Rather general considerations of the string theory landscape imply a mild statistical draw towards large soft SUSY breaking terms tempered by the requirement of proper electroweak symmetry breaking where SUSY contributions to the weak scale are not too far from m(weak)~ 100 GeV. Such a picture leads to the prediction that m_h~ 125 GeV while most sparticles are beyond current LHC reach. Here we explore the possibility that the magnitude of the Peccei-Quinn (PQ) scale f_a is also set by string landscape considerations within the framework of a compelling SUSY axion model. First, we examine the case where the PQ symmetry arises as an accidental approximate global symmetry from a more fundamental gravity-safe Z(24)^R symmetry and where the SUSY mu parameter arises from a Kim-Nilles operator. The pull towards large soft terms then also pulls the PQ scale as large as possible. Unless this is tempered by rather severe (unknown) cosmological or anthropic bounds on the density of dark matter, then we would expect a far greater abundance of dark matter than is observed. This conclusion cannot be negated by adopting a tiny axion misalignment angle theta_i because WIMPs are also overproduced at large f_a. Hence, we conclude that setting the PQ scale via anthropics is highly unlikely. Instead, requiring soft SUSY breaking terms of order the gravity-mediation scale m_{3/2}~ 10-100 TeV places the mixed axion-neutralino dark matter abundance into the intermediate scale sweet zone where f_a~ 10^{11}-10^{12} GeV. We compare our analysis to the more general case of a generic SUSY DFSZ axion model with uniform selection on theta_i but leading to the measured dark matter abundance: this approach leads to a preference for f_a~ 10^{12} GeV.Comment: 24 pages plus 10 figure

    Variable Selection in Latent Regression IRT Models via Knockoffs: An Application to International Large-scale Assessment in Education

    Full text link
    International large-scale assessments (ILSAs) play an important role in educational research and policy making. They collect valuable data on education quality and performance development across many education systems, giving countries the opportunity to share techniques, organizational structures, and policies that have proven efficient and successful. To gain insights from ILSA data, we identify non-cognitive variables associated with students' academic performance. This problem has three analytical challenges: 1) academic performance is measured by cognitive items under a matrix sampling design; 2) there are many missing values in the non-cognitive variables; and 3) multiple comparisons due to a large number of non-cognitive variables. We consider an application to the Programme for International Student Assessment (PISA), aiming to identify non-cognitive variables associated with students' performance in science. We formulate it as a variable selection problem under a general latent variable model framework and further propose a knockoff method that conducts variable selection with a controlled error rate for false selections

    Using the bayesmeta R package for Bayesian random-effects meta-regression

    Full text link
    BACKGROUND: Random-effects meta-analysis within a hierarchical normal modeling framework is commonly implemented in a wide range of evidence synthesis applications. More general problems may even be tackled when considering meta-regression approaches that in addition allow for the inclusion of study-level covariables. METHODS: We describe the Bayesian meta-regression implementation provided in the bayesmeta R package including the choice of priors, and we illustrate its practical use. RESULTS: A wide range of example applications are given, such as binary and continuous covariables, subgroup analysis, indirect comparisons, and model selection. Example R code is provided. CONCLUSIONS: The bayesmeta package provides a flexible implementation. Due to the avoidance of MCMC methods, computations are fast and reproducible, facilitating quick sensitivity checks or large-scale simulation studies.Comment: 17 pages, 8 figure

    Dealing with uncertainty in test assembly

    Get PDF
    The recent development of computer technologies enabled test institutes to improve their process of item selection for test administration by means of automated test assembly (ATA). A general framework for ATA consists in adopting mixed-integer programming models which are commonly intended to be solved by commercial solvers. Those softwares, notwithstanding their success in handling most of the basic ATA problems, are not always able to find solutions for highly constrained and large-sized instances. Moreover, all the model coefficients are assumed to be fixed and known, an hypothesis that is not true for the item information functions which are derived from the estimates of item response theory parameters. These restrictions motivated us to find an alternative way to specify and solve ATA models. First, we suggest an application of the chance-constrained (CC) approach (see Charnes and Cooper, 1959) which allows to maximize the α-quantile (usually smaller than 0.05) of the sampling distribution function of the test information function obtained by bootstrapping the calibration process. Secondly, for solving the ATA models, CC or not, we adapted a stochastic meta-heuristic called simulated annealing (SA) proposed by Goffe (1996). This technique can handle large-scale models and non-linear functions. A reformulation of the model by the Lagrangian relaxation helps to find the most feasible/optimal solution and, thanks to the SA, more than one neighbourhood of the space is explored avoiding to be trapped in a local optimum. Several simulations on ATA problems are performed and the solutions are compared to CPLEX 12.8.0 Optimizer, a benchmark solver in the linear programming field. Finally, a real data application shows the potential of our approach in practical situations. The algorithms are coded in the open-source framework Julia. Two packages written in Julia are released for solving the estimation and assembly problems described in this dissertation

    On sample selection models and skew distributions

    Get PDF
    This thesis is concerned with methods for dealing with missing data in nonrandom samples and recurrent events data. The first part of this thesis is motivated by scores arising from questionnaires which often follow asymmetric distributions, on a fixed range. This can be due to scores clustering at one end of the scale or selective reporting. Sometimes, the scores are further subjected to sample selection resulting in partial observability. Thus, methods based on complete cases for skew data are inadequate for the analysis of such data and a general sample selection model is required. Heckman proposed a full maximum likelihood estimation method under the normality assumption for sample selection problems, and parametric and non-parametric extensions have been proposed. A general selection distribution for a vector Y 2 Rp has a PDF fY given by fY(y) = fY?(y) P(S? 2 CjY? = y) P(S? 2 C) ; where S? 2 Rq and Y? 2 Rp are two random vectors, and C is a measurable subset of Rq. We use this generalization to develop a sample selection model with underlying skew-normal distribution. A link is established between the continuous component of our model log-likelihood function and an extended version of a generalized skewnormal distribution. This link is used to derive the expected value of the model, which extends Heckman's two-step method. The general selection distribution is also used to establish the closed skew-normal distribution as the continuous component of the usual multilevel sample selection models. Finite sample performances of the maximum likelihood estimator of the models are studied via Monte Carlo simulation. The model parameters are more precisely estimated under the new models, even in the presence of moderate to extreme skewness, than the Heckman selection models. Application to data from a study of neck injuries where the responses are substantially skew successfully discriminates between selection and inherent skewness, and the multilevel model is used to analyze jointly unit and item non-response. We also discuss computational and identification issues, and provide an extension of the model using copula-based sample selection models with truncated marginals. The second part of this thesis is motivated by studies that seek to analyze processes that generate events repeatedly over time. We consider the number of events per subject within a specified study period as the primary outcome of interest. One considerable challenge in the analysis of this type of data is the large proportion of patients that might discontinue before the end of the study, leading to partially observed data. Sophisticated sensitivity analyses tools are therefore necessary for the analysis of such data. We propose the use of two frequentist based imputation methods for dealing with missing data in recurrent event data framework. The recurrent events are modeled as over-dispersed Poisson data, with constant rate function. Different assumptions about future behavior of dropouts depending on reasons for dropout and treatment received are made and evaluated in a simulation study. We illustrate our approach with a clinical trial in patients who suffer from bladder cancer
    • …
    corecore