10,538 research outputs found

    Structural Regularities in Text-based Entity Vector Spaces

    Get PDF
    Entity retrieval is the task of finding entities such as people or products in response to a query, based solely on the textual documents they are associated with. Recent semantic entity retrieval algorithms represent queries and experts in finite-dimensional vector spaces, where both are constructed from text sequences. We investigate entity vector spaces and the degree to which they capture structural regularities. Such vector spaces are constructed in an unsupervised manner without explicit information about structural aspects. For concreteness, we address these questions for a specific type of entity: experts in the context of expert finding. We discover how clusterings of experts correspond to committees in organizations, the ability of expert representations to encode the co-author graph, and the degree to which they encode academic rank. We compare latent, continuous representations created using methods based on distributional semantics (LSI), topic models (LDA) and neural networks (word2vec, doc2vec, SERT). Vector spaces created using neural methods, such as doc2vec and SERT, systematically perform better at clustering than LSI, LDA and word2vec. When it comes to encoding entity relations, SERT performs best.Comment: ICTIR2017. Proceedings of the 3rd ACM International Conference on the Theory of Information Retrieval. 201

    Null Models of Economic Networks: The Case of the World Trade Web

    Get PDF
    In all empirical-network studies, the observed properties of economic networks are informative only if compared with a well-defined null model that can quantitatively predict the behavior of such properties in constrained graphs. However, predictions of the available null-model methods can be derived analytically only under assumptions (e.g., sparseness of the network) that are unrealistic for most economic networks like the World Trade Web (WTW). In this paper we study the evolution of the WTW using a recently-proposed family of null network models. The method allows to analytically obtain the expected value of any network statistic across the ensemble of networks that preserve on average some local properties, and are otherwise fully random. We compare expected and observed properties of the WTW in the period 1950-2000, when either the expected number of trade partners or total country trade is kept fixed and equal to observed quantities. We show that, in the binary WTW, node-degree sequences are sufficient to explain higher-order network properties such as disassortativity and clustering-degree correlation, especially in the last part of the sample. Conversely, in the weighted WTW, the observed sequence of total country imports and exports are not sufficient to predict higher-order patterns of the WTW. We discuss some important implications of these findings for international-trade models.Comment: 39 pages, 46 figures, 2 table

    Creativity as Cognitive design \ud The case of mesoscopic variables in Meta-Structures\ud

    Get PDF
    Creativity is an open problem which has been differently approached by several disciplines since a long time. In this contribution we consider as creative the constructivist design an observer does on the description levels of complex phenomena, such as the self-organized and emergent ones ( e.g., Bènard rollers, Belousov-Zhabotinsky reactions, flocks, swarms, and more radical cognitive and social emergences). We consider this design as related to the Gestaltian creation of a language fit for representing natural processes and the observer in an integrated way. Organised systems, both artificial and most of the natural ones are designed/ modelled according to a logical closed model which masters all the inter-relation between their constitutive elements, and which can be described by an algorithm or a single formal model. We will show there that logical openness and DYSAM (Dynamical Usage of Models) are the proper tools for those phenomena which cannot be described by algorithms or by a single formal model. The strong correlation between emergence and creativity suggests that an open model is the best way to provide a formal definition of creativity. A specific application relates to the possibility to shape the emergence of Collective Behaviours. Different modelling approaches have been introduced, based on symbolic as well as sub-symbolic rules of interaction to simulate collective phenomena by means of computational emergence. Another approach is based on modelling collective phenomena as sequences of Multiple Systems established by percentages of conceptually interchangeable agents taking on the same roles at different times and different roles at the same time. In the Meta-Structures project we propose to use mesoscopic variables as creative design, invention, good continuity and imitation of the description level. In the project we propose to define the coherence of sequences of Multiple Systems by using the values taken on by the dynamic mesoscopic clusters of its constitutive elements, such as the instantaneous number of elements having, in a flock, the same speed, distance from their nearest neighbours, direction and altitude. In Meta-Structures the collective behaviour’s coherence corresponds, for instance, to the scalar values taken by speed, distance, direction and altitude along time, through statistical strategies of interpolation, quasi-periodicity, levels of ergodicity and their reciprocal relationship. In this case the constructivist role of the observer is considered creative as it relates to neither non-linear replication nor transposition of levels of description and models used for artificial systems, like reductionism. Creativity rather lies in inventing new mesoscopic variables able to identify coherent patterns in complex systems. As it is known, mesoscopic variables represent partial macroscopic properties of a system by using some of the microscopic degrees of freedom possessed by composing elements. Such partial usage of microscopic as well as macroscopic properties allows a kind of Gestaltian continuity and imitation between levels of descriptions for mesoscopic modelling. \ud \u

    Using Laboratory Experiments to Design Efficient Market Institutions: The case of wholesale electricity markets

    Get PDF
    This paper assesses the contribution of laboratory experiments to the economics of design applied to the electricity industry. The analysis is dedicated to wholesale markets, and reviews the results accumulated to date concerning both the general architecture of power markets and the very details of the market rules or institution, that is the auction rule. We argue that these experimental results contribute to a better understanding of the performances properties and implementation features of competitive market designs and that experimental economics has proven very useful to public authorities to inform the restructuring of electricity industry. It thus confirms the role of experimental economics as a complement to theoretical approaches in the design effort.Experimental economics; market design; design economics; electricity auction;

    Patterns in syntactic dependency networks

    Get PDF
    Many languages are spoken on Earth. Despite their diversity, many robust language universals are known to exist. All languages share syntax, i.e., the ability of combining words for forming sentences. The origin of such traits is an issue of open debate. By using recent developments from the statistical physics of complex networks, we show that different syntactic dependency networks (from Czech, German, and Romanian) share many nontrivial statistical patterns such as the small world phenomenon, scaling in the distribution of degrees, and disassortative mixing. Such previously unreported features of syntax organization are not a trivial consequence of the structure of sentences, but an emergent trait at the global scale.Peer ReviewedPostprint (published version

    Training-free Measures Based on Algorithmic Probability Identify High Nucleosome Occupancy in DNA Sequences

    Full text link
    We introduce and study a set of training-free methods of information-theoretic and algorithmic complexity nature applied to DNA sequences to identify their potential capabilities to determine nucleosomal binding sites. We test our measures on well-studied genomic sequences of different sizes drawn from different sources. The measures reveal the known in vivo versus in vitro predictive discrepancies and uncover their potential to pinpoint (high) nucleosome occupancy. We explore different possible signals within and beyond the nucleosome length and find that complexity indices are informative of nucleosome occupancy. We compare against the gold standard (Kaplan model) and find similar and complementary results with the main difference that our sequence complexity approach. For example, for high occupancy, complexity-based scores outperform the Kaplan model for predicting binding representing a significant advancement in predicting the highest nucleosome occupancy following a training-free approach.Comment: 8 pages main text (4 figures), 12 total with Supplementary (1 figure
    • …
    corecore