535 research outputs found

    A Centered Index of Spatial Concentration: Axiomatic Approach with an Application to Population and Capital Cities

    Get PDF
    We construct an axiomatic index of spatial concentration around a center or capital point of interest, a concept with wide applicability from urban economics, economic geography and trade, to political economy and industrial organization. We propose basic axioms (decomposability and monotonicity) and refinement axioms (order preservation, convexity, and local monotonicity) for how the index should respond to changes in the underlying distribution. We obtain a unique class of functions satisfying all these properties, defined over any n-dimensional Euclidian space: the sum of a decreasing, isoelastic function of individual distances to the capital point of interest, with specific boundaries for the elasticity coefficient that depend on n. We apply our index to measure the concentration of population around capital cities across countries and US states, and also in US metropolitan areas. We show its advantages over alternative measures, and explore its correlations with many economic and political variables of interest.Spatial Concentration, Population Concentration, Capital Cities, Gravity, CRRA, Harmonic Functions, Axiomatics.

    A Centered Index of Spatial Concentration: Axiomatic Approach with an Application to Population and Capital Cities

    Get PDF
    We construct an axiomatic index of spatial concentration around a center or capital point of interest, a concept with wide applicability from urban economics, economic geography and trade, to political economy and industrial organization. We propose basic axioms (decomposability and monotonicity) and refinement axioms (order preservation, convexity, and local monotonicity) for how the index should respond to changes in the underlying distribution. We obtain a unique class of functions satisfying all these properties, defined over any n-dimensional Euclidian space: the sum of a decreasing, isoelastic function of individual distances to the capital point of interest, with specific boundaries for the elasticity coefficient that depend on n. We apply our index to measure the concentration of population around capital cities across countries and US states, and also in US metropolitan areas. We show its advantages over alternative measures, and explore its correlations with many economic and political variables of interest.

    A Centered Index of Spatial Concentration : Axiomatic Approach with an Application to Population and Capital Cities

    Get PDF
    We construct an axiomatic index of spatial concentration around a center or capital point of interest, a concept with wide applicability from urban economics, economic geography and trade, to political economy and industrial organization. We propose basic axioms (decomposability and monotonicity) and renement axioms (order preservation, convexity, and local monotonicity) for how the index should respond to changes in the underlying distribution. We obtain a unique class of functions satisfying all these properties, defined over any n-dimensional Euclidian space : the sum of a decreasing, isoelastic function of individual distances to the capital point of interest, with specifc boundaries for the elasticity coecient that depend on n. We apply our index to measure the concentration of population around capital cities across countries and US states, and also in US metropolitan areas. We show its advantages over alternative measures, and explore its correlations with many economic and political variables of interest.Spatial Concentration, Population Concentration, Capital Cities, Gravity, CRRA, Harmonic Functions, Axiomatics

    Apprentissage discriminant des modèles continus en traduction automatique

    Get PDF
    Over the past few years, neural network (NN) architectures have been successfully applied to many Natural Language Processing (NLP) applications, such as Automatic Speech Recognition (ASR) and Statistical Machine Translation (SMT).For the language modeling task, these models consider linguistic units (i.e words and phrases) through their projections into a continuous (multi-dimensional) space, and the estimated distribution is a function of these projections. Also qualified continuous-space models (CSMs), their peculiarity hence lies in this exploitation of a continuous representation that can be seen as an attempt to address the sparsity issue of the conventional discrete models. In the context of SMT, these echniques have been applied on neural network-based language models (NNLMs) included in SMT systems, and oncontinuous-space translation models (CSTMs). These models have led to significant and consistent gains in the SMT performance, but are also considered as very expensive in training and inference, especially for systems involving large vocabularies. To overcome this issue, Structured Output Layer (SOUL) and Noise Contrastive Estimation (NCE) have been proposed; the former modifies the standard structure on vocabulary words, while the latter approximates the maximum-likelihood estimation (MLE) by a sampling method. All these approaches share the same estimation criterion which is the MLE ; however using this procedure results in an inconsistency between theobjective function defined for parameter stimation and the way models are used in the SMT application. The work presented in this dissertation aims to design new performance-oriented and global training procedures for CSMs to overcome these issues. The main contributions lie in the investigation and evaluation of efficient training methods for (large-vocabulary) CSMs which aim~:(a) to reduce the total training cost, and (b) to improve the efficiency of these models when used within the SMT application. On the one hand, the training and inference cost can be reduced (using the SOUL structure or the NCE algorithm), or by reducing the number of iterations via a faster convergence. This thesis provides an empirical analysis of these solutions on different large-scale SMT tasks. On the other hand, we propose a discriminative training framework which optimizes the performance of the whole system containing the CSM as a component model. The experimental results show that this framework is efficient to both train and adapt CSM within SMT systems, opening promising research perspectives.Durant ces dernières années, les architectures de réseaux de neurones (RN) ont été appliquées avec succès à de nombreuses applications en Traitement Automatique de Langues (TAL), comme par exemple en Reconnaissance Automatique de la Parole (RAP) ainsi qu'en Traduction Automatique (TA).Pour la tâche de modélisation statique de la langue, ces modèles considèrent les unités linguistiques (c'est-à-dire des mots et des segments) à travers leurs projections dans un espace continu (multi-dimensionnel), et la distribution de probabilité à estimer est une fonction de ces projections.Ainsi connus sous le nom de "modèles continus" (MC), la particularité de ces derniers se trouve dans l'exploitation de la représentation continue qui peut être considérée comme une solution au problème de données creuses rencontré lors de l'utilisation des modèles discrets conventionnels.Dans le cadre de la TA, ces techniques ont été appliquées dans les modèles de langue neuronaux (MLN) utilisés dans les systèmes de TA, et dans les modèles continus de traduction (MCT).L'utilisation de ces modèles se sont traduit par d'importantes et significatives améliorations des performances des systèmes de TA. Ils sont néanmoins très coûteux lors des phrases d'apprentissage et d'inférence, notamment pour les systèmes ayant un grand vocabulaire.Afin de surmonter ce problème, l'architecture SOUL (pour "Structured Output Layer" en anglais) et l'algorithme NCE (pour "Noise Contrastive Estimation", ou l'estimation contrastive bruitée) ont été proposés: le premier modifie la structure standard de la couche de sortie, alors que le second cherche à approximer l'estimation du maximum de vraisemblance (MV) par une méthode d’échantillonnage.Toutes ces approches partagent le même critère d'estimation qui est la log-vraisemblance; pourtant son utilisation mène à une incohérence entre la fonction objectif définie pour l'estimation des modèles, et la manière dont ces modèles seront utilisés dans les systèmes de TA.Cette dissertation vise à concevoir de nouvelles procédures d'entraînement des MC, afin de surmonter ces problèmes.Les contributions principales se trouvent dans l'investigation et l'évaluation des méthodes d'entraînement efficaces pour MC qui visent à: (i) réduire le temps total de l'entraînement, et (ii) améliorer l'efficacité de ces modèles lors de leur utilisation dans les systèmes de TA.D'un côté, le coût d'entraînement et d'inférence peut être réduit (en utilisant l'architecture SOUL ou l'algorithme NCE), ou la convergence peut être accélérée.La dissertation présente une analyse empirique de ces approches pour des tâches de traduction automatique à grande échelle.D'un autre côté, nous proposons un cadre d'apprentissage discriminant qui optimise la performance du système entier ayant incorporé un modèle continu.Les résultats expérimentaux montrent que ce cadre d'entraînement est efficace pour l'apprentissage ainsi que pour l'adaptation des MC au sein des systèmes de TA, ce qui ouvre de nouvelles perspectives prometteuses

    Keeping Dictators Honest: the Role of Population Concentration

    Get PDF
    In order to explain the apparently paradoxical presence of acceptable governance in many non-democratic regimes, economists and political scientists have focused mostly on institutions acting as de facto checks and balances. In this paper, we propose that population plays a similar role in guaranteeing the quality of governance and redistribution. around the policy making center serves as an insurgency threat to a dictatorship, inducing it to yield to more redistribution and better governance. We bring this centered concept of population concentration to the data through the Centered Index of Spatial Concentration developed by Do & Campante (2008). The evidence supports our predictions: only in the sample of autocracies, population concentration around the capital city is positively associated with better governance and more redistribution (proxied by post-tax inequality), in OLS and IV regressions. Finally, we provide arguments to dismiss possible reverse causation as well as alternative, non-political economy explanations of such regularity, discuss the general applicability of our index and conclude with policy implications.Capital Cities, Gravity, Governance, Inequality, Redistribution, Population Concentration, Revolutions, Harmonic Functions, Axiomatics

    The spillover effects of target interest rate news from the U.S. Fed and the European Central Bank on the Asia-Pacific stock markets

    Get PDF
    This paper provides comprehensive evidence on the spillover effects of the U.S. Fed’s and the European Central Bank (ECB)’s target interest rate news on the market returns and return volatilities of 12 stock markets in the Asia-Pacific over the period 1999–2006. The news spillover effects on the returns are generally consistent with the literature where amajority of stock markets shows significant negative returns in response to unexpected rate rises. While the results of the speed of adjustment for the Fed’s news are mixed across the markets, the ECB news was absorbed slowly, in general. The return volatilities were higher in response to the interest rate news from both sources. In addition, both the Fed and the ECB news elicited tardy or persisting volatility responses. These findings have important implications for all levels of market participants in the Asia-Pacific stock markets.Target interest rate news; Spillover effects; U.S. Fed; ECB

    Approximate Data Analytics Systems

    Get PDF
    Today, most modern online services make use of big data analytics systems to extract useful information from the raw digital data. The data normally arrives as a continuous data stream at a high speed and in huge volumes. The cost of handling this massive data can be significant. Providing interactive latency in processing the data is often impractical due to the fact that the data is growing exponentially and even faster than Moore’s law predictions. To overcome this problem, approximate computing has recently emerged as a promising solution. Approximate computing is based on the observation that many modern applications are amenable to an approximate, rather than the exact output. Unlike traditional computing, approximate computing tolerates lower accuracy to achieve lower latency by computing over a partial subset instead of the entire input data. Unfortunately, the advancements in approximate computing are primarily geared towards batch analytics and cannot provide low-latency guarantees in the context of stream processing, where new data continuously arrives as an unbounded stream. In this thesis, we design and implement approximate computing techniques for processing and interacting with high-speed and large-scale stream data to achieve low latency and efficient utilization of resources. To achieve these goals, we have designed and built the following approximate data analytics systems: • StreamApprox—a data stream analytics system for approximate computing. This system supports approximate computing for low-latency stream analytics in a transparent way and has an ability to adapt to rapid fluctuations of input data streams. In this system, we designed an online adaptive stratified reservoir sampling algorithm to produce approximate output with bounded error. • IncApprox—a data analytics system for incremental approximate computing. This system adopts approximate and incremental computing in stream processing to achieve high-throughput and low-latency with efficient resource utilization. In this system, we designed an online stratified sampling algorithm that uses self-adjusting computation to produce an incrementally updated approximate output with bounded error. • PrivApprox—a data stream analytics system for privacy-preserving and approximate computing. This system supports high utility and low-latency data analytics and preserves user’s privacy at the same time. The system is based on the combination of privacy-preserving data analytics and approximate computing. • ApproxJoin—an approximate distributed joins system. This system improves the performance of joins — critical but expensive operations in big data systems. In this system, we employed a sketching technique (Bloom filter) to avoid shuffling non-joinable data items through the network as well as proposed a novel sampling mechanism that executes during the join to obtain an unbiased representative sample of the join output. Our evaluation based on micro-benchmarks and real world case studies shows that these systems can achieve significant performance speedup compared to state-of-the-art systems by tolerating negligible accuracy loss of the analytics output. In addition, our systems allow users to systematically make a trade-off between accuracy and throughput/latency and require no/minor modifications to the existing applications

    Studies of Group Fused Lasso and Probit Model for Right-Censored Data

    Get PDF
    This document is composed of three main chapters. In the first chapter, we study the mixture of experts, a powerful machine learning model in which each expert handles a different region of the covariate space. However, it is crucial to choose an appropriate number of experts to avoid overfitting or underfitting. A group fused lasso (GFL) term is added to the model with the goal of making the coefficients of the experts and the gating network closer together. An algorithm to optimize the problem is also developed using block-wise coordinate descent in the dual counterpart. Numerical results on simulated and real world datasets show that the penalized model outperforms the unpenalized one and performs on par with many well-known models. The second chapter studies GFL on its own and methods to solve it efficiently. In GFL, the response and the coefficient of each observation are not scalars but vectors. Thus, many fast solvers of the fused lasso cannot be applied to the GFL. Two algorithms are proposed to solve the GFL, namely Alternating Minimization and Dual Path. Results from speed trial show that our algorithms are competitive compared to other existing methods. The third chapter proposes a better alternative to the Box-Cox transformation, a popular method to transform the response variable to have an approximately normal distribution in many cases. The Box-Cox transformation is widely applied in regression, ANOVA and machine learning for both complete and censored data. However, since it is parametric, it can be too restrictive in many cases. Our proposed method is nonparametric, more flexible and can be fitted efficiently by our novel EM algorithms which accommodate both complete and right-censored data
    corecore