39 research outputs found

    ¿Cómo cerrar el círculo de la innovación? Abriéndolo

    Full text link
    Recuerdo la primera vez que vi la tercera "i", en minúscula, acompañando a las letras I + D, las ya conocidas como "Investigación+ Desarrollo". Me sorprendió, no sabía qué significaba, y me sentía que no iba al ritmo de los tiempos. El nuevo mantra era ahora la I + D+ i. La idea detrás de esa nueva letra era que los resultados de investigación no debían quedarse en el centro que los generara (universidad, instituto de investigación), sino que había que intentar transferir esos resultados a la sociedad, sacarlos del cajón del investigador, y permitir que fueran de provecho general. Esto normalmente se hace a través de empresas, que se encargan de productizar y comercializar el resultado de investigación

    Temporal characterization of the requests to Wikipedia

    Get PDF
    This paper presents an empirical study about the temporal patterns characterizing the requests submitted by users to Wikipedia. The study is based on the analysis of the log lines registered by the Wikimedia Foundation Squid servers after having sent the appropriate content in response to users' requests. The analysis has been conducted regarding the ten most visited editions of Wikipedia and has involved more than 14,000 million log lines corresponding to the traffic of the entire year 2009. The conducted methodology has mainly consisted in the parsing and filtering of users' requests according to the study directives. As a result, relevant information fields have been finally stored in a database for persistence and further characterization. In this way, we, first, assessed, whether the traffic to Wikipedia could serve as a reliable estimator of the overall traffic to all the Wikimedia Foundation projects. Our subsequent analysis of the temporal evolutions corresponding to the different types of requests to Wikipedia revealed interesting differences and similarities among them that can be related to the users' attention to the Encyclopedia. In addition, we have performed separated characterizations of each Wikipedia edition to compare their respective evolutions over time

    A quantitative examination of the impact of featured articles in Wikipedia

    Get PDF
    This paper presents a quantitative examination of the impact of the presentation of featured articles as quality content in the main page of several Wikipedia editions. Moreover, the paper also presents the analysis performed to determine the number of visits received by the articles promoted to the featured status. We have analyzed the visits not only in the month when articles awarded the promotion or were included in the main page, but also in the previous and following ones. The main aim for this is to assess the attention attracted by the featured content and the different dynamics exhibited by each community of users in respect to the promotion process. The main results of this paper are twofold: it shows how to extract relevant information related to the use of Wikipedia, which is an emerging research topic, and it analyzes whether the featured articles mechanism achieve to attract more attention

    On the distribution of source code file sizes

    Get PDF
    Source code size is an estimator of software effort. Size is also often used to calibrate models and equations to estimate the cost of software. The distribution of source code file sizes has been shown in the literature to be a lognormal distribution. In this paper, we measure the size of a large collection of software (the Debian GNU/Linux distribution version 5.0.2), and we find that the statistical distribution of its source code file sizes follows a double Pareto distribution. This means that large files are to be found more often than predicted by the lognormal distribution, therefore the previously proposed models underestimate the cost of software

    Studying the laws of software evolution in a long-lived FLOSS project

    Get PDF
    ome free, open-source software projects have been around for quite a long time, the longest living ones dating from the early 1980s. For some of them, detailed information about their evolution is available in source code management systems tracking all their code changes for periods of more than 15 years. This paper examines in detail the evolution of one of such projects, glibc, with the main aim of understanding how it evolved and how it matched Lehman's laws of software evolution. As a result, we have developed a methodology for studying the evolution of such long-lived projects based on the information in their source code management repository, described in detail several aspects of the history of glibc, including some activity and size metrics, and found how some of the laws of software evolution may not hold in this cas

    Characterization of the Wikipedia Traffic

    Get PDF
    Since its inception, Wikipedia has grown to a solid and stable project and turned into a mass collaboration tool that allows the sharing and distribution of knowledge. The wiki approach that basis this initiative promotes the participation and collaboration of users. In addition to visits for browsing its contents, Wikipedia also receives the contributions of users to improve them. In the past, researchers paid attention to different aspects concerning authoring and quality of contents. However, little effort has been made to study the nature of the visits that Wikipedia receives. We conduct such an study using a sample of users' requests provided by the Wikimedia Foundation in the form of Squid log lines. Our sample contains more that 14,000 million requests from users all around the world and directed to all the projects maintained by the Wikimedia Foundation, including different editions of Wikipedia. This papers describes the work made to characterize the traffic directed to Wikipedia and consisting of the requests sent by its users. Our main aim is to obtain a detailed description of its composition in terms of the percentages corresponding to the different types of requests making part of it. The benefits from our work may range from the prediction of traffic peaks to the determination of the kind of resources most often requested, which can be useful for scalability considerations

    Do Subsidies Provided to Public Transport in Madrid Favor Vertical Equity?

    Get PDF
    Despite the widespread implementation of subsidy policies for urban transport in many cities, the equity evaluation of these policies still remains limited. There is scarce quantitative assessment of the distributional incidence of transport subsidy policies. This paper contributes to fill this research gap by developing a practical approach to evaluate the impact of fare subsidization on vertical equity. In the paper we implement a two-step methodology. First, we develop two main indicators to measure the social impacts of the ?travel pass?, which is a highly subsidized fare in order to examine the policy for its effectiveness in reaching the poor. Second, by using the latest disaggregated data from Madrid?s Transportation Survey, we fitted a multiple regression model which found out that the use of the travel pass depends fundamentally on income level and accessibility to public transport. Since the quality of accessibility in the city is quite homogeneous, the subsidy policy associated with the travel pass is shown to be progressive because it is well targeted towards economically disadvantaged groups. Consequently, there seems to be evidence that subsidies provided to public transport in Madrid tend to favor vertical equity

    The evolution of the laws of software evolution. A discussion based on a systematic literature review

    Get PDF
    After more than 40 years of life, software evolution should be considered as a mature field. However, despite such a long history, many research questions still remain open, and controversial studies about the validity of the laws of software evolution are common. During the first part of these 40 years the laws themselves evolved to adapt to changes in both the research and the software industry environments. This process of adaption to new paradigms, standards, and practices stopped about 15 years ago, when the laws were revised for the last time. However, most controversial studies have been raised during this latter period. Based on a systematic and comprehensive literature review, in this paper we describe how and when the laws, and the software evolution field, evolved. We also address the current state of affairs about the validity of the laws, how they are perceived by the research community, and the developments and challenges that are likely to occur in the coming years

    Adapting the “Staged Model for Software Evolution” to FLOSS

    Get PDF
    Research into traditional software evolution has been tackled from two broad perspectives: that focused on the how, which looks at the processes, methods and techniques to implement and evolve software; and that focused on the what/why perspective, aiming at achieving an understanding of the drivers and general characteristics of the software evolution phenomenon. The two perspectives are related in various ways: the study of the what/why is for instance essential to achieve an appropriate management of software engineering activities, and to guide innovation in processes, methods and tools, that is, the how. The output of the what/why studies is exemplified by empirical hypotheses, such as the staged model of software evolution,. This paper focuses on the commonalities and differences between the evolution and patterns in the lifecycles of traditional commercial systems and free/libre/open source software (FLOSS) systems. The existing staged model for software evolution is therefore revised for its applicability on FLOSS systems

    Preliminary Comparison of Techniques for Dealing with Imbalance in Software Defect Prediction

    Get PDF
    Imbalanced data is a common problem in data mining when dealing with classi cation problems, where samples of a class vastly outnumber other classes. In this situation, many data mining algorithms generate poor models as they try to opti- mize the overall accuracy and perform badly in classes with very few samples. Software Engineering data in general and defect prediction datasets are not an exception and in this paper, we compare different approaches, namely sampling, cost-sensitive, ensemble and hybrid approaches to the prob- lem of defect prediction with different datasets preprocessed differently. We have used the well-known NASA datasets curated by Shepperd et al. There are differences in the re- sults depending on the characteristics of the dataset and the evaluation metrics, especially if duplicates and inconsisten- cies are removed as a preprocessing step.Unión Europea ICEBERG 324356MICYT TIN2007- 68084-C02-02MICYT TIN2013-46928-C3-2-
    corecore