4,672 research outputs found

    Onto.PT: Automatic Construction of a Lexical Ontology for Portuguese

    Get PDF
    This ongoing research presents an alternative to the man- ual creation of lexical resources and proposes an approach towards the automatic construction of a lexical ontology for Portuguese. Tex- tual sources are exploited in order to obtain a lexical network based on terms and, after clustering and mapping, a wordnet-like lexical on- tology is created. At the end of the paper, current results are shown

    A study on contextual influences on automatic playlist continuation

    Get PDF
    Recommender systems still mainly base their reasoning on pairwise interactions or information on individual entities, like item attributes or ratings, without properly evaluating the multiple dimensions of the recommendation problem. However, in many cases, like in music, items are rarely consumed in isolation, thus users rather need a set of items, selected to work well together, serving a specific purpose, while having some cognitive properties as a whole, related to their perception of quality and satisfaction, under given circumstances. In this paper, we introduce the term of playlist concept in order to capture the implicit characteristics of joint music item selections, related to their context, scope and general perception by the users. Although playlist consumptions may be associated with contextual attributes, these may be of various types, differently influencing users' preferences, based on their character and emotional state, therefore differently reflected on their final selections. We highlight on the use of this term in HybA, our hybrid recommender system, to identify clusters of similar playlists able to capture inherit characteristics and semantic properties, not explicitly described in them. The experimental results presented, show that this conceptual clustering results in playlist continuations of improved quality, compared to using explicit contextual parameters, or the commonly used collaborative filtering technique.Peer ReviewedPostprint (published version

    Text Mining Infrastructure in R

    Get PDF
    During the last decade text mining has become a widely used discipline utilizing statistical and machine learning methods. We present the tm package which provides a framework for text mining applications within R. We give a survey on text mining facilities in R and explain how typical application tasks can be carried out using our framework. We present techniques for count-based analysis methods, text clustering, text classification and string kernels.

    Framework for collaborative knowledge management in organizations

    Get PDF
    Nowadays organizations have been pushed to speed up the rate of industrial transformation to high value products and services. The capability to agilely respond to new market demands became a strategic pillar for innovation, and knowledge management could support organizations to achieve that goal. However, current knowledge management approaches tend to be over complex or too academic, with interfaces difficult to manage, even more if cooperative handling is required. Nevertheless, in an ideal framework, both tacit and explicit knowledge management should be addressed to achieve knowledge handling with precise and semantically meaningful definitions. Moreover, with the increase of Internet usage, the amount of available information explodes. It leads to the observed progress in the creation of mechanisms to retrieve useful knowledge from the huge existent amount of information sources. However, a same knowledge representation of a thing could mean differently to different people and applications. Contributing towards this direction, this thesis proposes a framework capable of gathering the knowledge held by domain experts and domain sources through a knowledge management system and transform it into explicit ontologies. This enables to build tools with advanced reasoning capacities with the aim to support enterprises decision-making processes. The author also intends to address the problem of knowledge transference within an among organizations. This will be done through a module (part of the proposed framework) for domain’s lexicon establishment which purpose is to represent and unify the understanding of the domain’s used semantic

    Doctor of Philosophy

    Get PDF
    dissertationThe explosion of structured Web data (e.g., online databases, Wikipedia infoboxes) creates many opportunities for integrating and querying these data that go far beyond the simple search capabilities provided by search engines. Although much work has been devoted to data integration in the database community, the Web brings new challenges: the Web-scale (e.g., the large and growing volume of data) and the heterogeneity in Web data. Because there are so much data, scalable techniques that require little or no manual intervention and that are robust to noisy data are needed. In this dissertation, we propose a new and effective approach for matching Web-form interfaces and for matching multilingual Wikipedia infoboxes. As a further step toward these problems, we propose a general prudent schema-matching framework that matches a large number of schemas effectively. Our comprehensive experiments for Web-form interfaces and Wikipedia infoboxes show that it can enable on-the-fly, automatic integration of large collections of structured Web data. Another problem we address in this dissertation is schema discovery. While existing integration approaches assume that the relevant data sources and their schemas have been identified in advance, schemas are not always available for structured Web data. Approaches exist that exploit information in Wikipedia to discover the entity types and their associate schemas. However, due to inconsistencies, sparseness, and noise from the community contribution, these approaches are error prone and require substantial human intervention. Given the schema heterogeneity in Wikipedia infoboxes, we developed a new approach that uses the structured information available in infoboxes to cluster similar infoboxes and infer the schemata for entity types. Our approach is unsupervised and resilient to the unpredictable skew in the entity class distribution. Our experiments, using over one hundred thousand infoboxes extracted from Wikipedia, indicate that our approach is effective and produces accurate schemata for Wikipedia entities

    A model to improve the Evaluation and Selection of Public Contest´s Candidates (Police Officers) based on AI technologies

    Get PDF
    Dissertation presented as the partial requirement for obtaining a Master's degree in Data Science and Advanced Analytics, specialization in Business AnalyticsThe number of candidates applying to Public Contests is increasing compared to the number of Human Resources employees required for selecting them for Police Forces. This work intends to perceive how those Public Institutions can evaluate and select their candidates efficiently during the different phases of the recruitment process, and for achieving this purpose AI approaches will be studied. This paper presents two research questions and introduces a corresponding systematic literature review, focusing on AI technologies, so the reader is able to understand which are most used and more appropriate to be applied to Police Forces as a complementary recruitment strategy of the National Criminal Investigation Police agency of Portugal – Polícia Judiciária. Design Science Research (DSR) was the methodological approach chosen. The suggestion of a theoretical framework is the main contribution of this study in pair with the segmentation of the candidates (future Criminal Inspectors). It also helped to comprehend the most important facts facing Public Institutions regarding the usage of AI technologies, to make decisions about evaluating and selecting candidates. Following the PRISMA methodology guidelines, a systematic literature review and meta-analyses method was adopted to identify how can the usage and exploitation of transparent AI have a positive impact on the recruitment process of a Public Institution, resulting in an analysis of 34 papers published between 2017 and 2021. The AI-based theoretical framework, applicable within the analysis of literature papers, solves the problem of how the Institutions can gain insights about their candidates while profiling them; how to obtain more accurate information from the interview phase; and how to reach a more rigorous assessment of their emotional intelligence providing a better alignment of moral values. This way, this work aims to advise the improvement of the decision making to be taken by a recruiter of a Police Force Institution, turning it into a more automated and evidence-based decision when it comes to recruiting the adequate candidate for the place

    Knowledge Representation and WordNets

    Get PDF
    Knowledge itself is a representation of “real facts”. Knowledge is a logical model that presents facts from “the real world” witch can be expressed in a formal language. Representation means the construction of a model of some part of reality. Knowledge representation is contingent to both cognitive science and artificial intelligence. In cognitive science it expresses the way people store and process the information. In the AI field the goal is to store knowledge in such way that permits intelligent programs to represent information as nearly as possible to human intelligence. Knowledge Representation is referred to the formal representation of knowledge intended to be processed and stored by computers and to draw conclusions from this knowledge. Examples of applications are expert systems, machine translation systems, computer-aided maintenance systems and information retrieval systems (including database front-ends).knowledge, representation, ai models, databases, cams
    corecore