520 research outputs found

    Latent Dirichlet Allocation (LDA) for improving the topic modeling of the official bulletin of the spanish state (BOE)

    Get PDF
    Since Internet was born most people can access fully free to a lot sources of information. Every day a lot of web pages are created and new content is uploaded and shared. Never in the history the humans has been more informed but also uninformed due the huge amount of information that can be access. When we are looking for something in any search engine the results are too many for reading and filtering one by one. Recommended Systems (RS) was created to help us to discriminate and filter these information according to ours preferences. This contribution analyses the RS of the official agency of publications in Spain (BOE), which is known as "Mi BOE'. The way this RS works was analysed, and all the meta-data of the published documents were analysed in order to know the coverage of the system. The results of our analysis show that more than 89% of the documents cannot be recommended, because they are not well described at the documentary level, some of their key meta-data are empty. So, this contribution proposes a method to label documents automatically based on Latent Dirichlet Allocation (LDA). The results are that using this approach the system could recommend (at a theoretical point of view) more than twice of documents that it now does, 11% vs 23% after applied this approach

    A cloud-based tool for sentiment analysis in reviews about restaurants on TripAdvisor

    Get PDF
    The tourism industry has been promoting its products and services based on the reviews that people often write on travel websites like TripAdvisor.com, Booking.com and other platforms like these. These reviews have a profound effect on the decision making process when evaluating which places to visit, such as which restaurants to book, etc. In this contribution is presented a cloud based software tool for the massive analysis of this social media data (TripAdvisor.com). The main characteristics of the tool developed are: i) the ability to aggregate data obtained from social media; ii) the possibility of carrying out combined analyses of both people and comments; iii) the ability to detect the sense (positive, negative or neutral) in which the comments rotate, quantifying the degree to which they are positive or negative, as well as predicting behaviour patterns from this information; and iv) the ease of doing everything in the same application (data downloading, pre-processing, analysis and visualisation). As a test and validation case, more than 33.500 revisions written in English on restaurants in the Province of Granada (Spain) were analyse

    A Review on Information Accessing Systems Based on Fuzzy Linguistic Modelling

    Get PDF
    This paper presents a survey of some fuzzy linguistic information access systems. The review shows information retrieval systems, filtering systems, recommender systems, and web quality evaluation tools, which are based on tools of fuzzy linguistic modelling. The fuzzy linguistic modelling allows us to represent and manage the subjectivity, vagueness and imprecision that is intrinsic and characteristic of the processes of information searching, and, in such a way, the developed systems allow users the access to quality information in a flexible and user-adapted way.European Union (EU) TIN2007-61079 PET2007-0460Ministry of Public Works 90/07Excellence Andalusian Project TIC529

    Double Penalty in Returns to Education: Informality and Educational Mismatch in the Colombian Labour market

    Get PDF
    This paper examines the returns to education taking into consideration the existence of educational mismatches in the formal and informal employment of a developing country. Results show that the returns of surplus, required and deficit years of schooling are different in the two sectors. Moreover, they suggest that these returns vary along the wage distribution, and that the pattern of variation differs for formal and informal workers. In particular, informal workers face not only lower returns to their education, but suffer a second penalty associated with educational mismatches that puts them at a greater disadvantage compare to their formal counterparts

    Informality and Overeducation in the Labor Market of a Developing Country

    Get PDF
    In this paper, we explore the connection between labor market segmentation in two sectors, a modern protected formal sector and a traditionalunprotected-informal sector, and overeducation in a developing country. Informality is thought to have negative consequences, primarily through poorer working conditions, lack of social security, as well as low levels of productivity throughout the economy. This paper considers an aspect that has not been previously addressed, namely the fact that informality might also affect the way workers match their actual education with that required performing their job. We use micro-data from Colombia to test the relationship between overeducation and informality. Empirical results suggest that, once the endogeneity of employment choice has been accounted for, formal male workers are less likely to be overeducated. Interestingly, the propensity of being overeducated among women does not seem to be closely related to the sector choice

    Decision-making model for designing telecom products/services based on customer preferences and non-preferences

    Get PDF
    The design of the packages of products/services to be offered by a telecom company to its clients is a complex decision-making process that must consider different criteria to achieve both customer satisfaction and optimization of the company’s resources. In this process, Intuitionistic Fuzzy Sets (IFSs) can be used to manage uncertainty and better represent both preferences and non-preferences expressed by people who value each proposed alternative. We present a novel approach to design/develop new products/services that combines the Lean Six Sigma methodology with IFSs. Its main contribution comes from considering both preferences and nonpreferences expressed by real clients, whereas existing proposals only consider their preferences. By also considering their non-preferences, it provides an additional capacity to manage the high uncertainty in the selection of the commercial plan that best suits each client’s needs. Thus, client satisfaction is increased while improving the company’s corporate image, which will lead to customer loyalty and increased revenue. To validate the presented proposal, it has been applied to a real case study of the telecom sector, in which 2135 users have participated. The results obtained have been analysed and compared with those obtained with a model that does not consider the non-preferences expressed by users.Spanish Ministry of Science and Innovation (State Research Agency)Junta de Andalucia PID2019-103880RB-I00 PID2019-109644RB-I00 PY20_0067

    Visualization and evolution of the scientific structure of fuzzy sets research in Spain

    Get PDF
    Introduction. Presents the first bibliometric study on the evolution of the fuzzy sets theory field. It is specially focused on the research carried out by the Spanish comunity. Method. The CoPalRed software, for network analysis, and the co-word analysis technique are used. Analysis. Bibliometric maps showing the main associations among the main concepts in the field are provided for the periods 1965-1993, 1994-1998, 1999-2003 and 2004-2008. Results. The bibliometric maps obtained provide insight into the structure of the fuzzy sets theory research in the Spanish community, visualize the research subfields, and show the existing relationships between those subfields. Furthermore, we compare the Spanish community with other countries (the USA and Canada; the UK and Germany; and Japan and Peoples Republic of China). Conclusions. As a result of the analysis, a complete study of the evolution of the Spanish fuzzy sets community and an analysis of its international importance are presented.Introducción. Presenta el primer estudio bibliométrico sobre la evolución del campo de la teoría de conjuntos difusos. Se centra especialmente en la investigación llevada a cabo por la comunidad española. Método. Se utiliza el software de CoPalRed, para el análisis de red, y la técnica de anñlisis de co-palabras. Aná;lisis. Se proporcionan mapas bibliométricos mostrando las principales asociaciones entre los principales conceptos en el campo para los períodos de 1965-1993, 1994-1998, 1999-2003 y 2004-2008. Resultados. Los mapas bibliométricos obtenidos proporcionan conocimientos sobre la estructura de la investigación sobre teoría de conjuntos difusos en la comunidad española, visualizan los subcampos de investigación y muestran las relaciones existentes entre los subcampos. Ademñs, se compara la comunidad española con otros países (Estados Unidos y Canada; Reino Unido y Alemania; y Japón y República Popular China). Conclusiones. Como resultado de los análisis, se presenta un estudio completo de la evolución de la comunidad española en conjuntos difusos y un anñlisis de su importancia internacional.This work has been supported by the Spanish project FUZZY-LING, Cod. TIN2007-61079, granted by the Spanish Agency for Education and Science (Ministerio de Educación y Ciencia)

    Connecting the Dots in Trustworthy Artificial Intelligence: From AI Principles, Ethics, and Key Requirements to Responsible AI Systems and Regulation

    Full text link
    Trustworthy Artificial Intelligence (AI) is based on seven technical requirements sustained over three main pillars that should be met throughout the system's entire life cycle: it should be (1) lawful, (2) ethical, and (3) robust, both from a technical and a social perspective. However, attaining truly trustworthy AI concerns a wider vision that comprises the trustworthiness of all processes and actors that are part of the system's life cycle, and considers previous aspects from different lenses. A more holistic vision contemplates four essential axes: the global principles for ethical use and development of AI-based systems, a philosophical take on AI ethics, a risk-based approach to AI regulation, and the mentioned pillars and requirements. The seven requirements (human agency and oversight; robustness and safety; privacy and data governance; transparency; diversity, non-discrimination and fairness; societal and environmental wellbeing; and accountability) are analyzed from a triple perspective: What each requirement for trustworthy AI is, Why it is needed, and How each requirement can be implemented in practice. On the other hand, a practical approach to implement trustworthy AI systems allows defining the concept of responsibility of AI-based systems facing the law, through a given auditing process. Therefore, a responsible AI system is the resulting notion we introduce in this work, and a concept of utmost necessity that can be realized through auditing processes, subject to the challenges posed by the use of regulatory sandboxes. Our multidisciplinary vision of trustworthy AI culminates in a debate on the diverging views published lately about the future of AI. Our reflections in this matter conclude that regulation is a key for reaching a consensus among these views, and that trustworthy and responsible AI systems will be crucial for the present and future of our society.Comment: 30 pages, 5 figures, under second revie

    Global trends in coronavirus research at the time of Covid-19: A general bibliometric approach and content analysis using SciMAT

    Get PDF
    Covid-19 represents the greatest challenge facing mankind today. In December 2019, several cases of pneumonia of unknown etiology were reported from China. This coronavirus infection subsequently identified as Covid-19 aroused worldwide concern. As a result, the scientific community has focused attention on Covid-19, as revealed by recent research reported in literature based on a holistic approach. In this regard, this study conducts a bibliometric analysis of coronavirus research in the literature with an emphasis on Covid-19 disease, using as a reference the publications in the Web of Science Core Collection from 1970 to 2020. This research analyzes 12,571 publications from 1970 to (April 18) 2020 by applying advanced bibliometric techniques in SciMAT bibliometric analysis software. The current research therefore provides a complete conceptual analysis of the main coronavirus types and strains in the literature by quantifying the main bibliometric performance indicators, identifying the main authors, organizations, countries, sources, and research areas, and evaluating the development of this field. Furthermore, a science map is constructed to understand the corresponding intellectual structure and main research lines (themes). SciMAT thereby offers a complete approach to the field and evaluates the main performance indicators related to coronavirus, with a focus on Covid-19. Finally, this research serves as a framework to strengthen existing research lines and develop new ones, establishing synergistic relationships that were not visible without the maps generated herein

    Geovisualization of green spaces with augmented reality

    Full text link
    The management of green spaces in urban areas requires new techniques and instruments that provide all the information available to people. It is intended to facilitate data-based services to citizens through the use of mobile devices in green spaces, parks and most visited public gardens. The selected area of study has been the Plaza de Oriente (Madrid) and its surroundings. Maps and images from different sources of information have been implemented in this research work.Then we proceeded to develop a database with woodland and existing cultural heritage. Finally, related this information using geolocation and augmented reality system. We have created a system that allows the user to interact with the single orientation of the camera on your mobile device. This application allows efficient use through the integration of mobile devices and the knowledge of the cultural heritage associated with the public space
    corecore