746 research outputs found

    Large-scale Spatial Distribution Identification of Base Stations in Cellular Networks

    Full text link
    The performance of cellular system significantly depends on its network topology, where the spatial deployment of base stations (BSs) plays a key role in the downlink scenario. Moreover, cellular networks are undergoing a heterogeneous evolution, which introduces unplanned deployment of smaller BSs, thus complicating the performance evaluation even further. In this paper, based on large amount of real BS locations data, we present a comprehensive analysis on the spatial modeling of cellular network structure. Unlike the related works, we divide the BSs into different subsets according to geographical factor (e.g. urban or rural) and functional type (e.g. macrocells or microcells), and perform detailed spatial analysis to each subset. After examining the accuracy of Poisson point process (PPP) in BS locations modeling, we take into account the Gibbs point processes as well as Neyman-Scott point processes and compare their accuracy in view of large-scale modeling test. Finally, we declare the inaccuracy of the PPP model, and reveal the general clustering nature of BSs deployment, which distinctly violates the traditional assumption. This paper carries out a first large-scale identification regarding available literatures, and provides more realistic and more general results to contribute to the performance analysis for the forthcoming heterogeneous cellular networks

    ppmlhdfe: Fast Poisson Estimation with High-Dimensional Fixed Effects

    Get PDF
    In this paper we present ppmlhdfe, a new Stata command for estimation of (pseudo) Poisson regression models with multiple high-dimensional fixed effects (HDFE). Estimation is implemented using a modified version of the iteratively reweighted least-squares (IRLS) algorithm that allows for fast estimation in the presence of HDFE. Because the code is built around the reghdfe package, it has similar syntax, supports many of the same functionalities, and benefits from reghdfe's fast convergence properties for computing high-dimensional least squares problems. Performance is further enhanced by some new techniques we introduce for accelerating HDFE-IRLS estimation specifically. ppmlhdfe also implements a novel and more robust approach to check for the existence of (pseudo) maximum likelihood estimates.Comment: For associated code and data repository, see https://github.com/sergiocorreia/ppmlhdf

    Communication-Avoiding Optimization Methods for Distributed Massive-Scale Sparse Inverse Covariance Estimation

    Full text link
    Across a variety of scientific disciplines, sparse inverse covariance estimation is a popular tool for capturing the underlying dependency relationships in multivariate data. Unfortunately, most estimators are not scalable enough to handle the sizes of modern high-dimensional data sets (often on the order of terabytes), and assume Gaussian samples. To address these deficiencies, we introduce HP-CONCORD, a highly scalable optimization method for estimating a sparse inverse covariance matrix based on a regularized pseudolikelihood framework, without assuming Gaussianity. Our parallel proximal gradient method uses a novel communication-avoiding linear algebra algorithm and runs across a multi-node cluster with up to 1k nodes (24k cores), achieving parallel scalability on problems with up to ~819 billion parameters (1.28 million dimensions); even on a single node, HP-CONCORD demonstrates scalability, outperforming a state-of-the-art method. We also use HP-CONCORD to estimate the underlying dependency structure of the brain from fMRI data, and use the result to identify functional regions automatically. The results show good agreement with a clustering from the neuroscience literature.Comment: Main paper: 15 pages, appendix: 24 page

    Hidden Gibbs random fields model selection using Block Likelihood Information Criterion

    Full text link
    Performing model selection between Gibbs random fields is a very challenging task. Indeed, due to the Markovian dependence structure, the normalizing constant of the fields cannot be computed using standard analytical or numerical methods. Furthermore, such unobserved fields cannot be integrated out and the likelihood evaluztion is a doubly intractable problem. This forms a central issue to pick the model that best fits an observed data. We introduce a new approximate version of the Bayesian Information Criterion. We partition the lattice into continuous rectangular blocks and we approximate the probability measure of the hidden Gibbs field by the product of some Gibbs distributions over the blocks. On that basis, we estimate the likelihood and derive the Block Likelihood Information Criterion (BLIC) that answers model choice questions such as the selection of the dependency structure or the number of latent states. We study the performances of BLIC for those questions. In addition, we present a comparison with ABC algorithms to point out that the novel criterion offers a better trade-off between time efficiency and reliable results

    Measuring industrial agglomeration with inhomogeneous K-function: the case of ICT firms in Milan (Italy)

    Get PDF
    Why do industrial clusters occur in space? Is it because industries need to stay close together to interact or, conversely, because they concentrate in certain portions of space to exploit favourable conditions like public incentives, proximity to communication networks, to big population concentrations or to reduce transport costs? This is a fundamental question and the attempt to answer to it using empirical data is a challenging statistical task. In economic geography scientists refer to this dichotomy using the two categories of spatial interaction and spatial reaction to common factors. In economics we can refer to a distinction between exogenous causes and endogenous effects. In spatial econometrics and statistics we use the terms of spatial dependence and spatial heterogeneity. A series of recent papers introduced explorative methods to analyses the spatial patterns of firms using micro data and characterizing each firm by its spatial coordinates. In such a setting a spatial distribution of firms is seen as a point pattern and an industrial cluster as the phenomenon of extra-concentration of one industry with respect to the concentration of a benchmarking spatial distribution. Often the benchmarking distribution is that of the whole economy on the ground that exogenous factors affect in the same way all branches. Using such an approach a positive (or negative) spatial dependence between firms is detected when the pattern of a specific sector is more aggregated (or more dispersed) than the one of the whole economy. In this paper we suggest a parametric approach to the analysis of spatial heterogeneity, based on the socalled inhomogeneous K-function (Baddeley et al., 2000). We present an empirical application of the method to the spatial distribution of high-tech industries in Milan (Italy) in 2001. We consider the economic space to be non homogenous, we estimate the pattern of inhomogeneity and we use it to separate spatial heterogeneity from spatial dependence.
    • …
    corecore