537 research outputs found

    A customized semantic service retrieval methodology for the digital ecosystems environment

    Get PDF
    With the emergence of the Web and its pervasive intrusion on individuals, organizations, businesses etc., people now realize that they are living in a digital environment analogous to the ecological ecosystem. Consequently, no individual or organization can ignore the huge impact of the Web on social well-being, growth and prosperity, or the changes that it has brought about to the world economy, transforming it from a self-contained, isolated, and static environment to an open, connected, dynamic environment. Recently, the European Union initiated a research vision in relation to this ubiquitous digital environment, known as Digital (Business) Ecosystems. In the Digital Ecosystems environment, there exist ubiquitous and heterogeneous species, and ubiquitous, heterogeneous, context-dependent and dynamic services provided or requested by species. Nevertheless, existing commercial search engines lack sufficient semantic supports, which cannot be employed to disambiguate user queries and cannot provide trustworthy and reliable service retrieval. Furthermore, current semantic service retrieval research focuses on service retrieval in the Web service field, which cannot provide requested service retrieval functions that take into account the features of Digital Ecosystem services. Hence, in this thesis, we propose a customized semantic service retrieval methodology, enabling trustworthy and reliable service retrieval in the Digital Ecosystems environment, by considering the heterogeneous, context-dependent and dynamic nature of services and the heterogeneous and dynamic nature of service providers and service requesters in Digital Ecosystems.The customized semantic service retrieval methodology comprises: 1) a service information discovery, annotation and classification methodology; 2) a service retrieval methodology; 3) a service concept recommendation methodology; 4) a quality of service (QoS) evaluation and service ranking methodology; and 5) a service domain knowledge updating, and service-provider-based Service Description Entity (SDE) metadata publishing, maintenance and classification methodology.The service information discovery, annotation and classification methodology is designed for discovering ubiquitous service information from the Web, annotating the discovered service information with ontology mark-up languages, and classifying the annotated service information by means of specific service domain knowledge, taking into account the heterogeneous and context-dependent nature of Digital Ecosystem services and the heterogeneous nature of service providers. The methodology is realized by the prototype of a Semantic Crawler, the aim of which is to discover service advertisements and service provider profiles from webpages, and annotating the information with service domain ontologies.The service retrieval methodology enables service requesters to precisely retrieve the annotated service information, taking into account the heterogeneous nature of Digital Ecosystem service requesters. The methodology is presented by the prototype of a Service Search Engine. Since service requesters can be divided according to the group which has relevant knowledge with regard to their service requests, and the group which does not have relevant knowledge with regard to their service requests, we respectively provide two different service retrieval modules. The module for the first group enables service requesters to directly retrieve service information by querying its attributes. The module for the second group enables service requesters to interact with the search engine to denote their queries by means of service domain knowledge, and then retrieve service information based on the denoted queries.The service concept recommendation methodology concerns the issue of incomplete or incorrect queries. The methodology enables the search engine to recommend relevant concepts to service requesters, once they find that the service concepts eventually selected cannot be used to denote their service requests. We premise that there is some extent of overlap between the selected concepts and the concepts denoting service requests, as a result of the impact of service requesters’ understandings of service requests on the selected concepts by a series of human-computer interactions. Therefore, a semantic similarity model is designed that seeks semantically similar concepts based on selected concepts.The QoS evaluation and service ranking methodology is proposed to allow service requesters to evaluate the trustworthiness of a service advertisement and rank retrieved service advertisements based on their QoS values, taking into account the contextdependent nature of services in Digital Ecosystems. The core of this methodology is an extended CCCI (Correlation of Interaction, Correlation of Criterion, Clarity of Criterion, and Importance of Criterion) metrics, which allows a service requester to evaluate the performance of a service provider in a service transaction based on QoS evaluation criteria in a specific service domain. The evaluation result is then incorporated with the previous results to produce the eventual QoS value of the service advertisement in a service domain. Service requesters can rank service advertisements by considering their QoS values under each criterion in a service domain.The methodology for service domain knowledge updating, service-provider-based SDE metadata publishing, maintenance, and classification is initiated to allow: 1) knowledge users to update service domain ontologies employed in the service retrieval methodology, taking into account the dynamic nature of services in Digital Ecosystems; and 2) service providers to update their service profiles and manually annotate their published service advertisements by means of service domain knowledge, taking into account the dynamic nature of service providers in Digital Ecosystems. The methodology for service domain knowledge updating is realized by a voting system for any proposals for changes in service domain knowledge, and by assigning different weights to the votes of domain experts and normal users.In order to validate the customized semantic service retrieval methodology, we build a prototype – a Customized Semantic Service Search Engine. Based on the prototype, we test the mathematical algorithms involved in the methodology by a simulation approach and validate the proposed functions of the methodology by a functional testing approach

    Fuzzy concept analysis for semantic knowledge extraction

    Get PDF
    2010 - 2011Availability of controlled vocabularies, ontologies, and so on is enabling feature to provide some added values in terms of knowledge management. Nevertheless, the design, maintenance and construction of domain ontologies are a human intensive and time consuming task. The Knowledge Extraction consists of automatic techniques aimed to identify and to define relevant concepts and relations of the domain of interest by analyzing structured (relational databases, XML) and unstructured (text, documents, images) sources. Specifically, methodology for knowledge extraction defined in this research work is aimed at enabling automatic ontology/taxonomy construction from existing resources in order to obtain useful information. For instance, the experimental results take into account data produced with Web 2.0 tools (e.g., RSS-Feed, Enterprise Wiki, Corporate Blog, etc.), text documents, and so on. Final results of Knowledge Extraction methodology are taxonomies or ontologies represented in a machine oriented manner by means of semantic web technologies, such as: RDFS, OWL and SKOS. The resulting knowledge models have been applied to different goals. On the one hand, the methodology has been applied in order to extract ontologies and taxonomies and to semantically annotate text. On the other hand, the resulting ontologies and taxonomies are exploited in order to enhance information retrieval performance and to categorize incoming data and to provide an easy way to find interesting resources (such as faceted browsing). Specifically, following objectives have been addressed in this research work: Ontology/Taxonomy Extraction: that concerns to automatic extraction of hierarchical conceptualizations (i.e., taxonomies) and relations expressed by means typical description logic constructs (i.e., ontologies). Information Retrieval: definition of a technique to perform concept-based the retrieval of information according to the user queries. Faceted Browsing: in order to automatically provide faceted browsing capabilities according to the categorization of the extracted contents. Semantic Annotation: definition of a text analysis process, aimed to automatically annotate subjects and predicates identified. The experimental results have been obtained in some application domains: e-learning, enterprise human resource management, clinical decision support system. Future challenges go in the following directions: investigate approaches to support ontology alignment and merging applied to knowledge management.X n.s

    Contributions to privacy in web search engines

    Get PDF
    Els motors de cerca d’Internet recullen i emmagatzemen informació sobre els seus usuaris per tal d’oferir-los millors serveis. A canvi de rebre un servei personalitzat, els usuaris perden el control de les seves pròpies dades. Els registres de cerca poden revelar informació sensible de l’usuari, o fins i tot revelar la seva identitat. En aquesta tesis tractem com limitar aquests problemes de privadesa mentre mantenim suficient informació a les dades. La primera part d’aquesta tesis tracta els mètodes per prevenir la recollida d’informació per part dels motores de cerca. Ja que aquesta informació es requerida per oferir un servei precís, l’objectiu es proporcionar registres de cerca que siguin adequats per proporcionar personalització. Amb aquesta finalitat, proposem un protocol que empra una xarxa social per tal d’ofuscar els perfils dels usuaris. La segona part tracta la disseminació de registres de cerca. Proposem tècniques que la permeten, proporcionant k-anonimat i minimitzant la pèrdua d’informació.Web Search Engines collects and stores information about their users in order to tailor their services better to their users' needs. Nevertheless, while receiving a personalized attention, the users lose the control over their own data. Search logs can disclose sensitive information and the identities of the users, creating risks of privacy breaches. In this thesis we discuss the problem of limiting the disclosure risks while minimizing the information loss. The first part of this thesis focuses on the methods to prevent the gathering of information by WSEs. Since search logs are needed in order to receive an accurate service, the aim is to provide logs that are still suitable to provide personalization. We propose a protocol which uses a social network to obfuscate users' profiles. The second part deals with the dissemination of search logs. We propose microaggregation techniques which allow the publication of search logs, providing kk-anonymity while minimizing the information loss

    User Interfaces for Personal Knowledge Management with Semantic Technologies

    Get PDF
    This thesis describes iMapping and QuiKey, two novel user interface concepts for dealing with structured information. iMapping is a visual knowledge mapping technique based on zooming, which combines the advantages of several existing approaches and scales up to very large maps. QuiKey is a text-based tool to interact with graph-structured knowledge bases with very high interaction efficiency. Both tools have been implemented and positively evaluated in user studies

    Semantic enrichment of knowledge sources supported by domain ontologies

    Get PDF
    This thesis introduces a novel conceptual framework to support the creation of knowledge representations based on enriched Semantic Vectors, using the classical vector space model approach extended with ontological support. One of the primary research challenges addressed here relates to the process of formalization and representation of document contents, where most existing approaches are limited and only take into account the explicit, word-based information in the document. This research explores how traditional knowledge representations can be enriched through incorporation of implicit information derived from the complex relationships (semantic associations) modelled by domain ontologies with the addition of information presented in documents. The relevant achievements pursued by this thesis are the following: (i) conceptualization of a model that enables the semantic enrichment of knowledge sources supported by domain experts; (ii) development of a method for extending the traditional vector space, using domain ontologies; (iii) development of a method to support ontology learning, based on the discovery of new ontological relations expressed in non-structured information sources; (iv) development of a process to evaluate the semantic enrichment; (v) implementation of a proof-of-concept, named SENSE (Semantic Enrichment kNowledge SourcEs), which enables to validate the ideas established under the scope of this thesis; (vi) publication of several scientific articles and the support to 4 master dissertations carried out by the department of Electrical and Computer Engineering from FCT/UNL. It is worth mentioning that the work developed under the semantic referential covered by this thesis has reused relevant achievements within the scope of research European projects, in order to address approaches which are considered scientifically sound and coherent and avoid “reinventing the wheel”.European research projects - CoSpaces (IST-5-034245), CRESCENDO (FP7-234344) and MobiS (FP7-318452

    Semantic Interaction in Web-based Retrieval Systems : Adopting Semantic Web Technologies and Social Networking Paradigms for Interacting with Semi-structured Web Data

    Get PDF
    Existing web retrieval models for exploration and interaction with web data do not take into account semantic information, nor do they allow for new forms of interaction by employing meaningful interaction and navigation metaphors in 2D/3D. This thesis researches means for introducing a semantic dimension into the search and exploration process of web content to enable a significantly positive user experience. Therefore, an inherently dynamic view beyond single concepts and models from semantic information processing, information extraction and human-machine interaction is adopted. Essential tasks for semantic interaction such as semantic annotation, semantic mediation and semantic human-computer interaction were identified and elaborated for two general application scenarios in web retrieval: Web-based Question Answering in a knowledge-based dialogue system and semantic exploration of information spaces in 2D/3D

    Methodology development for measuring virtual university social responsibility (VUSR)

    Get PDF
    This thesis addresses the most challenging issues in online education and its social responsibility. A world’s first ontology on virtual university social responsibility (VUSR) along with an ontological-driven approach for measuring the corporate social responsibility (CSR) for virtual universities (VUs) in five dimensions, namely education, research, engagement, ethics and transparency is developed. The impact on social, economic and ethical standards is ascertained by rigorously defining measurement indicators and performance assessment attributes to help assess CSR

    A knowledge discovery approach to urban analysis

    Get PDF
    Enhancing our knowledge of the complexities of cities in order to empower ourselves to make more informed decisions has always been a challenge for urban research. Recent developments in large-scale computing, together with the new techniques and automated tools for data collection and analysis are opening up promising opportunities for addressing this problem. The main motivation that served as the driving force behind this research is how these developments may contribute to urban data analysis. On this basis, the thesis focuses on urban data analysis in order to search for findings that can enhance our knowledge of urban environments, using the generic process of knowledge discovery using data mining. A knowledge discovery process based on data mining is a fully automated or semi-automated process which involves the application of computational tools and techniques to explore the “previously unknown, and potentially useful information” (Witten & Frank, 2005) hidden in large and often complex and multi-dimensional databases. This information can be obtained in the form of correlations amongst variables, data groupings (classes and clusters) or more complex hypotheses (probabilistic rules of co-occurrence, performance vectors of prediction models etc.). This research targets researchers and practitioners working in the field of urban studies who are interested in quantitative/ computational approaches to urban data analysis and specifically aims to engage the interest of architects, urban designers and planners who do not have a background in statistics or in using data mining methods in their work. Accordingly, the overall aim of the thesis is the development of a knowledge discovery approach to urban analysis; a domain-specific adaptation of the generic process of knowledge discovery using data mining enabling the analyst to discover ‘relational urban knowledge’. ‘Relational urban knowledge’ is a term employed in this thesis to refer to the potentially ‘useful’ and/or ‘valuable’ information patterns and relationships that can be discovered in urban databases by applying data mining algorithms. A knowledge discovery approach to urban analysis through data mining can help us to understand site-specific characteristics of urban environments in a more profound and useful way. On a more specific level, the thesis aims towards ‘knowledge discovery’ in traditional thematic maps published in 2008 by the Istanbul Metropolitan Municipality as a basis of the Master Plan for the Beyoğlu Preservation Area. These thematic maps, which represent urban components, namely buildings, streets, neighbourhoods and their various attributes such as floor space use of the buildings, land price, population density or historical importance, do not really extend our knowledge of Beyoğlu Preservation Area beyond documenting its current state and do not contribute to the interventions presented in the master plan. However it is likely that ‘useful’ and ‘valuable’ information patterns discoverable using data mining algorithms are hidden in them. In accordance with the stated aims, three research questions of the thesis concerns (1) the development of a general process model to adapt the generic process of knowledge discovery using data mining for urban data analysis, (2) the investigation of information patterns and relationships that can be extracted from the traditional thematic maps of the Beyoğlu Preservation Area by further developing and implementing this model and (3) the investigation of how could this ‘relational urban knowledge’ support architects, urban designers or urban planners whilst developing intervention proposals for urban regeneration. A Knowledge Discovery Process Model (KDPM) for urban analysis was developed, as an answer to the the first research question. The KDPM for urban analysis is a domain-specific adaptation of the widely accepted process of knowledge discovery in databases defined by Fayyad, Piatetsky-Shapiro, and Smyth (1996b). The model describes a semi-automated process of database formulation, analysis and evaluation for extracting information patterns and relationships from raw data by combining both GIS and data mining functionalities in a complementary way. The KDPM for urban analysis suggests that GIS functionalities can be used to formulate a database, and GIS and data mining can complement each other in analyzing the database and evaluating the outcomes. The model illustrates that the output of a GIS platform can become the input for a data mining platform and vice versa, resulting in an interlinked analytical process which allows for a more sophisticated analysis of urban data. To investigate the second and third research questions, firstly the KDPM for urban analysis was further developed to construct a GIS database of the Beyoğlu Preservation Area from the thematic maps. Then, three implementations were performed using this GIS database; the Beyoğlu Preservation Area Building Features Database consisting of multiple features attributed to the buildings. In Implementation (1), the KDPM for urban analysis was used to investigate a variety of patterns and relationships that can be extracted from the database using three different data mining methods. In Implementations (2) and (3), the KDPM for urban analysis was implemented to test how the knowledge discovery approach through data mining proposed in this thesis can assist in developing draft plans for the regeneration of a run-down neighbourhood in the Beyoğlu Preservation Area (Tarlabaşı). In Implementation (2), the KDPM for urban analysis is implemented in combination with an evolutionary process to apply a regeneration approach developed by the author; a computational process which generates draft plans for ground floor use, user-profile and tenure-type allocation was developed. In Implementation (3), students applied the KDPM for urban analysis during the course of an international workshop. The model enabled them to explore site-specific particularities of Tarlabaşı that would support their urban intervention proposals. Among the outputs of the thesis three of them are considered as utilizable outputs that distinguish this thesis from previous studies: The KDPM for urban analysis. Although there have been other studies which make use of data mining methods and techniques combined with GIS technology, to the best of our knowledge no previous research has implemented a process model to depict this process and used the model to extract ‘knowledge’ from traditional thematic maps. Researchers and practitioners can re-use this process model to analyze other urban environments. The KDPM for urban analysis is, therefore, one of the main utilizable outputs of the thesis and an important scientific contribution of this study. The Beyoğlu Preservation Area Building Features Database. A large and quite comprehensive GIS database which consists of 45 spatial and non-spatial features attributed to the 11,984 buildings located in the Beyoğlu Preservation Area was constructed. This database is one of the original features of this study. To the best of our knowledge, there are no other examples of applications of data mining using such a comprehensive GIS database, constructed from a range of actual micro-scale data representing such a variety of features attributed to the buildings. This database can be re-used by analysts interested in studying the Beyoğlu Preservation Area. The Beyoğlu Preservation Area Building Features Database is therefore one of the main utilizable outputs of the thesis and represents a scientific contribution to the research material on the Beyoğlu Preservation Area. A computational process which generates draft plans for ground floor use, user-profile and tenure-type allocation, using GIS and data mining functionalities with evolutionary computation. This output of the thesis was generated by Implementation (2), which aimed to investigate Research Question (3). The overall process involved the successive application of Naïve Bayes Classification, Association Rule Analysis and an Evolutionary Algorithm to a subset of the Beyoğlu Preservation Area Building Features Database representing the Tarlabaşı neighbourhood. Briefly, the findings of the data mining analysis were used to formulate a set of rules for assigning ground floor use information to the buildings. These rules were then used for fitness measurements of an Evolutionary Algorithm, together with other fitness measurements for assigning user-profile and tenure-type information (defined by the author according to the regeneration approach developed by the author). As a result, the algorithm transformed the existing allocation of the ground floor use in the buildings located in Tarlabaşı in accordance with the given rules and assigned user-profile and tenure type information for each building. This computational process demonstrated one way to use the data mining analysis findings in developing intervention proposals for urban regeneration. A similar computational process can be implemented in other urban contexts by researchers and practitioners. To the best of our knowledge, no prior research has used data mining analysis findings for fitness measurements of an Evolutionary Algorithm in order to produce draft plans for ground floor use, user-profile and tenure-type allocation. This is, therefore, the most original scientific contribution and utilizable output of the thesis. As a result of the research, on the basis of the data that is available in the thematic maps of the Beyoğlu Preservation Area, the potential of a knowledge discovery approach to urban analysis in revealing the relationships between various components of urban environments and their various attributes is demonstrated. It is also demonstrated that these relationships can reveal site-specific characteristics of urban environments and if found ‘valuable’ by the the targeted researchers and practitioners, these can lead to the development of more informed intervention proposals. Thereby the knowledge discovery approach to urban analysis developed in this thesis may help to improve the quality of urban intervention proposals and consequently the quality of built environments. On the other hand, the implementations carried out in the thesis also exposed the major limitation of the knowledge discovery approach to urban analysis through data mining, which is the fact that the findings discoverable by this approach are limited by the relevant data that is collectable and accessible

    A finder and representation system for knowledge carriers based on granular computing

    Get PDF
    In one of his publications Aristotle states ”All human beings by their nature desire to know” [Kraut 1991]. This desire is initiated the day we are born and accompanies us for the rest of our life. While at a young age our parents serve as one of the principle sources for knowledge, this changes over the course of time. Technological advances and particularly the introduction of the Internet, have given us new possibilities to share and access knowledge from almost anywhere at any given time. Being able to access and share large collections of written down knowledge is only one part of the equation. Just as important is the internalization of it, which in many cases can prove to be difficult to accomplish. Hence, being able to request assistance from someone who holds the necessary knowledge is of great importance, as it can positively stimulate the internalization procedure. However, digitalization does not only provide a larger pool of knowledge sources to choose from but also more people that can be potentially activated, in a bid to receive personalized assistance with a given problem statement or question. While this is beneficial, it imposes the issue that it is hard to keep track of who knows what. For this task so-called Expert Finder Systems have been introduced, which are designed to identify and suggest the most suited candidates to provide assistance. Throughout this Ph.D. thesis a novel type of Expert Finder System will be introduced that is capable of capturing the knowledge users within a community hold, from explicit and implicit data sources. This is accomplished with the use of granular computing, natural language processing and a set of metrics that have been introduced to measure and compare the suitability of candidates. Furthermore, are the knowledge requirements of a problem statement or question being assessed, in order to ensure that only the most suited candidates are being recommended to provide assistance
    corecore