22 research outputs found

    Organizational modeling with a semantic wiki: formalization of content and automatic diagram generation

    Get PDF
    A key to maintain Enterprises competitiveness is the ability to describe, standardize, and adapt the way it reacts to certain types of business events, and how it interacts with suppliers, partners, competitors, and customers. In this context the field of organization modeling has emerged with the aim to create models that help to create a state of self-awareness in the organization. This project's context is the use of Semantic Web in the Organizational modeling area. The Semantic Web technology advantages can be used to improve the way of modeling organizations. This was accomplished using a Semantic wiki to model organizations. Our research and implementation had two main purposes: formalization of textual content in semantic wiki pages; and automatic generation of diagrams from organization data stored in the semantic wiki pages.Orientador: Pedro campos e Co-orientador: David Aveir

    Mining Meaning from Wikipedia

    Get PDF
    Wikipedia is a goldmine of information; not just for its many readers, but also for the growing community of researchers who recognize it as a resource of exceptional scale and utility. It represents a vast investment of manual effort and judgment: a huge, constantly evolving tapestry of concepts and relations that is being applied to a host of tasks. This article provides a comprehensive description of this work. It focuses on research that extracts and makes use of the concepts, relations, facts and descriptions found in Wikipedia, and organizes the work into four broad categories: applying Wikipedia to natural language processing; using it to facilitate information retrieval and information extraction; and as a resource for ontology building. The article addresses how Wikipedia is being used as is, how it is being improved and adapted, and how it is being combined with other structures to create entirely new resources. We identify the research groups and individuals involved, and how their work has developed in the last few years. We provide a comprehensive list of the open-source software they have produced.Comment: An extensive survey of re-using information in Wikipedia in natural language processing, information retrieval and extraction and ontology building. Accepted for publication in International Journal of Human-Computer Studie

    Grounding event references in news

    Get PDF
    Events are frequently discussed in natural language, and their accurate identification is central to language understanding. Yet they are diverse and complex in ontology and reference; computational processing hence proves challenging. News provides a shared basis for communication by reporting events. We perform several studies into news event reference. One annotation study characterises each news report in terms of its update and topic events, but finds that topic is better consider through explicit references to background events. In this context, we propose the event linking task which—analogous to named entity linking or disambiguation—models the grounding of references to notable events. It defines the disambiguation of an event reference as a link to the archival article that first reports it. When two references are linked to the same article, they need not be references to the same event. Event linking hopes to provide an intuitive approximation to coreference, erring on the side of over-generation in contrast with the literature. The task is also distinguished in considering event references from multiple perspectives over time. We diagnostically evaluate the task by first linking references to past, newsworthy events in news and opinion pieces to an archive of the Sydney Morning Herald. The intensive annotation results in only a small corpus of 229 distinct links. However, we observe that a number of hyperlinks targeting online news correspond to event links. We thus acquire two large corpora of hyperlinks at very low cost. From these we learn weights for temporal and term overlap features in a retrieval system. These noisy data lead to significant performance gains over a bag-of-words baseline. While our initial system can accurately predict many event links, most will require deep linguistic processing for their disambiguation

    Grounding event references in news

    Get PDF
    Events are frequently discussed in natural language, and their accurate identification is central to language understanding. Yet they are diverse and complex in ontology and reference; computational processing hence proves challenging. News provides a shared basis for communication by reporting events. We perform several studies into news event reference. One annotation study characterises each news report in terms of its update and topic events, but finds that topic is better consider through explicit references to background events. In this context, we propose the event linking task which—analogous to named entity linking or disambiguation—models the grounding of references to notable events. It defines the disambiguation of an event reference as a link to the archival article that first reports it. When two references are linked to the same article, they need not be references to the same event. Event linking hopes to provide an intuitive approximation to coreference, erring on the side of over-generation in contrast with the literature. The task is also distinguished in considering event references from multiple perspectives over time. We diagnostically evaluate the task by first linking references to past, newsworthy events in news and opinion pieces to an archive of the Sydney Morning Herald. The intensive annotation results in only a small corpus of 229 distinct links. However, we observe that a number of hyperlinks targeting online news correspond to event links. We thus acquire two large corpora of hyperlinks at very low cost. From these we learn weights for temporal and term overlap features in a retrieval system. These noisy data lead to significant performance gains over a bag-of-words baseline. While our initial system can accurately predict many event links, most will require deep linguistic processing for their disambiguation

    Design of an E-learning system using semantic information and cloud computing technologies

    Get PDF
    Humanity is currently suffering from many difficult problems that threaten the life and survival of the human race. It is very easy for all mankind to be affected, directly or indirectly, by these problems. Education is a key solution for most of them. In our thesis we tried to make use of current technologies to enhance and ease the learning process. We have designed an e-learning system based on semantic information and cloud computing, in addition to many other technologies that contribute to improving the educational process and raising the level of students. The design was built after much research on useful technology, its types, and examples of actual systems that were previously discussed by other researchers. In addition to the proposed design, an algorithm was implemented to identify topics found in large textual educational resources. It was tested and proved to be efficient against other methods. The algorithm has the ability of extracting the main topics from textual learning resources, linking related resources and generating interactive dynamic knowledge graphs. This algorithm accurately and efficiently accomplishes those tasks even for bigger books. We used Wikipedia Miner, TextRank, and Gensim within our algorithm. Our algorithm‘s accuracy was evaluated against Gensim, largely improving its accuracy. Augmenting the system design with the implemented algorithm will produce many useful services for improving the learning process such as: identifying main topics of big textual learning resources automatically and connecting them to other well defined concepts from Wikipedia, enriching current learning resources with semantic information from external sources, providing student with browsable dynamic interactive knowledge graphs, and making use of learning groups to encourage students to share their learning experiences and feedback with other learners.Programa de Doctorado en Ingeniería Telemática por la Universidad Carlos III de MadridPresidente: Luis Sánchez Fernández.- Secretario: Luis de la Fuente Valentín.- Vocal: Norberto Fernández Garcí

    Knowledge representation and ontologies for lipids and lipidomics

    Get PDF
    Master'sMASTER OF SCIENC

    On cross-domain social semantic learning

    Get PDF
    Approximately 2.4 billion people are now connected to the Internet, generating massive amounts of data through laptops, mobile phones, sensors and other electronic devices or gadgets. Not surprisingly then, ninety percent of the world's digital data was created in the last two years. This massive explosion of data provides tremendous opportunity to study, model and improve conceptual and physical systems from which the data is produced. It also permits scientists to test pre-existing hypotheses in various fields with large scale experimental evidence. Thus, developing computational algorithms that automatically explores this data is the holy grail of the current generation of computer scientists. Making sense of this data algorithmically can be a complex process, specifically due to two reasons. Firstly, the data is generated by different devices, capturing different aspects of information and resides in different web resources/ platforms on the Internet. Therefore, even if two pieces of data bear singular conceptual similarity, their generation, format and domain of existence on the web can make them seem considerably dissimilar. Secondly, since humans are social creatures, the data often possesses inherent but murky correlations, primarily caused by the causal nature of direct or indirect social interactions. This drastically alters what algorithms must now achieve, necessitating intelligent comprehension of the underlying social nature and semantic contexts within the disparate domain data and a quantifiable way of transferring knowledge gained from one domain to another. Finally, the data is often encountered as a stream and not as static pages on the Internet. Therefore, we must learn, and re-learn as the stream propagates. The main objective of this dissertation is to develop learning algorithms that can identify specific patterns in one domain of data which can consequently augment predictive performance in another domain. The research explores existence of specific data domains which can function in synergy with another and more importantly, proposes models to quantify the synergetic information transfer among such domains. We include large-scale data from various domains in our study: social media data from Twitter, multimedia video data from YouTube, video search query data from Bing Videos, Natural Language search queries from the web, Internet resources in form of web logs (blogs) and spatio-temporal social trends from Twitter. Our work presents a series of solutions to address the key challenges in cross-domain learning, particularly in the field of social and semantic data. We propose the concept of bridging media from disparate sources by building a common latent topic space, which represents one of the first attempts toward answering sociological problems using cross-domain (social) media. This allows information transfer between social and non-social domains, fostering real-time socially relevant applications. We also engineer a concept network from the semantic web, called semNet, that can assist in identifying concept relations and modeling information granularity for robust natural language search. Further, by studying spatio-temporal patterns in this data, we can discover categorical concepts that stimulate collective attention within user groups.Includes bibliographical references (pages 210-214)

    Semantically-enhanced advertisement recommender systems in social networks

    Get PDF
    El objetivo principal de la investigación es estudiar y diseñar un entorno de recomendación publicitaria en las redes sociales que puede ser enriquecido mediante tecnologías semánticas. A pesar de que existen muchas aplicaciones y soluciones para los sistemas de recomendación, en este estudio se diseña un framework robusto con un rendimiento adecuado para poder ser implementado en las redes sociales con el objetivo de ampliar los propósitos de negocio. De este objetivo principal se pueden derivar los siguientes objetivos secundarios: 1. Superar las limitaciones iniciales de los métodos clásicos de recomendación. 2. Aumentar la calidad y precisión de las recomendaciones y el rendimiento del sistema de recomendación. 3. Utilizar convenientemente la metodología planteada. 4. Establecer el marco propuesto en una plataforma de software real. 5. Considerar en la solución la portabilidad como un aspecto clave en los sistemas de software. 6. Considerar la fiabilidad del framework. 7. Tener un nivel de seguridad aceptable para el framework. En primer lugar, es necesario superar las limitaciones de los métodos clásicos de recomendación. En el presente trabajo, este objetivo se alcanzará mediante un método híbrido que se componga de los cuatro métodos básicos de recomendación (filtrado colaborativo, basado en contenido, demográfico y basado en conocimiento), y que recoja cada uno de los beneficios individuales de los mismos. En concreto, a pesar de los problemas conocidos de los métodos basados en filtrado colaborativo, a saber, escasez de datos (del inglés ‘data sparsity’), escalabilidad y arranque en frio (del inglés ‘cold start’), sigue siendo fundamental aprovechar las ventajas de esta técnica colaborativa de recomendación. Además, mediante la adición de técnicas semánticas durante el proceso de cálculo de las recomendaciones publicitarias, se aumentará la calidad y precisión de éstas. La tecnología semántica utilizada en el marco ha mejorado el rendimiento del sistema y supone un punto novedoso, siendo ésta una de las principales contribuciones frente al resto de investigaciones similares. En particular, para mejorar la exactitud de las recomendaciones, la semántica tanto de los distintos elementos de información como de los perfiles de clientes se ha tenido en cuenta. Introducir la semántica en el pronóstico proporciona una visión adicional sobre las explicaciones básicas detrás de las cuales un cliente podría permitir el acceso a productos específicos (algo que se entiende y se cubre con estrategias habituales sin consideración semántica). La semántica utilizada en este estudio es entendida en forma de relaciones entre conceptos. Como resultado, es posible extraer un conocimiento extra de los elementos disponibles. Otro de los objetivos de esta tesis es asegurar que se siga una metodología apropiada. Es necesario que la investigación obtenga resultados aceptables mediante la implementación de algoritmos fáciles de usar y un enfoque adecuado. Para alcanzar este objetivo, se diseña un caso de estudio, y posteriormente se implementa una aplicación Web capaz de determinar recomendaciones para los usuarios. El desarrollo de esta aplicación Web tiene sus propias dificultades y complejidades, pero la aplicación es amigable y fácil de usar. Los usuarios pueden navegar fácilmente en línea y trabajar con las aplicaciones instaladas en el sitio Web. La evaluación de la aproximación propuesta se realizará sobre este entorno real. De esta forma, también se establece como objetivo el establecer el framework en una plataforma de software real para probarlo y observar el rendimiento del mismo. Este objetivo es muy importante dado que si no existe la posibilidad de establecer un prototipo (prueba de concepto) para implementar la idea de la investigación, no será posible llegar a una conclusión adecuada y alcanzar los objetivos del estudio. Así, antes de desarrollar la idea de la investigación, se verificó si era posible encontrar una solución de software para obtener resultados reales en el marco implementado que permitiera posteriormente observar el resultado adecuado y, de este modo, asegurase de que los objetivos y requerimientos iniciales de la investigación en forma de resultados finales pueden ser probados. Asegurar la portabilidad y la fiabilidad es otra de las claves perseguidas en este trabajo. En este contexto, la portabilidad hace referencia a la posibilidad de implementar el framework en distintas plataformas disponibles incluyendo hardware, software, tipo de red social y publicidad. En este caso, el diseño del marco es independiente de cualquier plataforma. El framework se ha propuesto en un formato general y es muy fácil ajustarlo a los sistemas de software y hardware disponibles. Incluso es posible establecer el marco en diferentes sistemas operativos y no hay limitación en el número de instancias de instalación. Por otro lado, la fiabilidad, similar a la validez, es un método para evaluar la naturaleza de la estrategia de estimación utilizada para recopilar información en un estudio. En conjunto, para que los resultados de un estudio se consideren sustanciales, el sistema de estimación debe ser sólido. Lo que se persigue con la fiabilidad es que cualquier resultado crítico sea más que un hallazgo irregular y sea, por tanto, repetible. Distintos científicos deben tener la capacidad de realizar la misma investigación, en las mismas condiciones y producir los mismos resultados. Esto fortalecerá los descubrimientos y garantizará que grupos académicos más extensos reconozcan la teoría. La fiabilidad entendida de este modo es, en consecuencia, esencial para que una teoría se acumule como una verdad experimental reconocida. En esta tesis doctoral se realizan sobre la aplicación Web un total de 73 experimentos, resultando en un nivel prometedor de fiabilidad. Por último, la seguridad es uno de los retos fundamentales en las aplicaciones de la Web social y constituye un requisito básico del marco de trabajo propuesto en esta tesis. La seguridad es, en realidad, una de las principales preocupaciones de todas las aplicaciones software y la implementación del marco en una plataforma segura es, por tanto, muy importante. Para ello se consideró el componente de seguridad como uno de los elementos del marco, el cual se compone de diferentes niveles: (i) autenticación, y (ii) comprobación de identidad a partir del comportamiento. La autenticación única (‘SSO’ del inglés, Single Sign-On) permite a los usuarios loguearse en el sistema. Por otro lado, se mantiene un registro del comportamiento del usuario en las interacciones con la aplicación Web y se compara éste con el histórico. Este segundo nivel de seguridad previene el acceso de atacantes a contenidos no autorizados.The composition of Semantic Web advances with Web 2.0 application plan designs has risen to the social semantic Web, additionally introduced as Web 3.0. In accordance with this thought, a software platform will be displayed that effectively joins both Web 2.0 ideas and Semantic Web advancements. The structure of this study joins a progression of semantic-based application modules in a completely fledged social application with the goal of catching semantics in the purpose of information retrieval. Once the establishments and principle ideas of the alluded framework are brought up and its architecture was explained, a comprehensive model of the system will be demonstrated. Finally, the result of a case study will be validated using the standard metrics. It will be spoken to how the system can help in obtaining semantically-improved financially related data from the clients of the social applications and giving valuable proposals to advertisement recommender. The ability of knowledge contribution nowadays is unmatched ever. At no other time have such a large number of inventive and proficient individuals been associated by such a productive, all-inclusive system. The expenses of social occasion and registering over their commitments have come down to the point where new organizations with extremely humble spending plans give imaginative new administrations to a great number of online members. Collective intelligence is an amazing insight which can have numerous constructive outcomes on social networks. The outcome nowadays is amazing broadness of data and variety of point of view, and a society of mass investment that supports a wellspring of freely accessible substance. The Social Web (containing services, for example, MySpace, Flickr, last.fm, and WordPress) has caught the consideration of a large number of clients and in addition billions of dollars in venture and procurement. Social sites, advancing around the associations amongst individuals and their entities of interest, are experiencing limits in the territories of information integration, dispersal, reuse, compactness, searchability, automation and requesting undertakings like questioning. The Semantic Web is a perfect tool for interlinking and performing operations on various individual and item related information accessible from the Social Web, and has delivered an assortment of ways to deal with beat the limits being knowledgeable about Social Web application ranges. Recommendation is a compelling approach to diminish the expense for discovering data furthermore a capable approach to draw in clients. It has been broadly utilized as a part of numerous e-commerce applications, e.g., Amazon.com, CDNOW.com, eBay.com, Reel.com, et cetera. As of late, numerous techniques have been proposed for suggestion, for instance, Content-based Filtering, Collaborative Filtering, Clustering Model, Classification Model, Graph Model, and Association Rule approach. The proposed approaches have been connected to the conventional Web applications, which as a rule need suggest one and only sort of data (e.g., Amazon prescribes books, news.baidu.com prescribes news, and movielens.com prescribes films). So as to defeat data over-burden, recommender frameworks have turned into a key apparatus for giving clients customized suggestions on things, for example, films, music, books, news, and web pages. Captivated by numerous viable applications, analysts have created calculations and frameworks in the course of the most recent decade. Some of them have been popularized by online merchants, for example, Amazon.com, Netflix.com, and IMDb.com. These frameworks foresee user preferences (frequently spoke to as numeric evaluations) for new items in light of the client's past appraisals on different items. There are regularly two sorts of calculations for recommender frameworks - content-based techniques and collaborative filtering. Content-based techniques measure the likeness of the prescribed item (target item) to the ones that an objective user (i.e., user who gets recommendations) likes or aversions in light of item properties. Then again, collaborative filtering discovers users with tastes that are like the objective users depends on their ratings in the past. Collaborative filtering will then make recommendations to the objective user in light of the feelings of those comparative users. In spite of these endeavors, recommender frameworks still face numerous testing issues. These problems will make many limitations on the operation of recommendation systems. The change in the expectation precision can build client fulfillment, which thusly prompts higher benefits for those e-trade sites. Second, calculations for recommender frameworks experience the side effects of numerous problems. For instance, keeping in mind the end goal to gauge thing closeness, Content-based strategies depend with respect to express thing depictions. Be that as it may, such depictions might be hard to acquire for things like thoughts or feelings. As opposed to the tremendous number of things in recommender frameworks, every client regularly just rates a couple. In this way, the user/thing rating matrix is commonly extremely scanty. It is troublesome for recommender frameworks to precisely quantify client likenesses from those predetermined number of audits. A related issue is the Cold-start issue. Notwithstanding for a framework that is not especially meager, when a client at first joins, the framework has none or maybe just a couple audits from this client. In this manner, the framework can't precisely translate this current client's inclination. To handle those issues, two methodologies have been proposed. The main methodology is to gather the user/item rating matrix through dimensionality lessening systems, for example, Singular Value Decomposition (SVD). By grouping clients or things as per their idle structure, unrepresentative clients or things can be disposed of, and in this way the user/item grid gets to be denser. Nonetheless, these strategies don't essentially enhance the execution of recommender frameworks, and now and again aggravate the execution even. For using this approach, a methodology of kNN has been utilized for the framework to cluster users to two groups of neighbors and the other. So, the framework considers only those neighbor users which have more relative and similar data to the current user. The second approach is to "improve" the user/item rating matrix by 1) presenting default evaluations or verifiable client ratings, e.g., the time spent on perusing articles; 2) utilizing silly evaluating expectations from content-based techniques; or 3) abusing transitive relationship among clients through their past exchanges and feedback. These techniques enhance the execution of recommender frameworks to some degree. Specifically, another worldview of recommender frameworks is proposed by using data in social networks, particularly that of social impact. Customary recommender frameworks do not think about unequivocal social relations among clients, yet the significance of social impact in item advertising has for quite some time been perceived. Instinctively, when we need to purchase an item that is not commonplace, we frequently counsel with our companions who have as of now had involvement with the item, since they are those that we can go after quick exhortation. At the point when companions prescribe an item to us, we additionally have a tendency to acknowledge the suggestion in light of the fact that their inputs are dependable. This is one reason that collaborative filtering has been used as one of the components of the recommender system. Furthermore, the combination of social networks can hypothetically enhance the execution of current recommender frameworks. To start with, as far as the forecast precision, the extra data about clients and their companions acquired from social networks enhances the comprehension of client practices and appraisals. In this manner, we can demonstrate and translate client inclinations all the more absolutely, and accordingly enhance the forecast precision. Second, with companion data in social networks, it is no more important to discover comparable clients by measuring their rating comparability, in light of the fact that the way that two individuals are companions as of now demonstrates that they have things in like manner. In this manner, the information Sparsity issue can be reduced. At long last, for the Cold-start issue, regardless of the possibility that a client has no past audits, recommender framework still can make proposals to the client in view of the inclinations of his/her companions on the off chance that it coordinates with social networks. These instincts and perceptions rouse us to plan another worldview of recommender frameworks that can exploit data in social networks. The late rise of online social networks (OSNs) gives us a chance to examine the part of social impact in recommender frameworks. With the expanding ubiquity of Web 2.0, numerous OSNs, for example, Myspace.com, Facebook.com, and Linkedin.com have risen. Individuals in those systems have their own customized space where they not just distribute their life stories, leisure activities, interests, online journals, and so forth., additionally list their companions. Companions or guests can visit these individual spaces and leave remarks. OSNs give stages where individuals can put themselves on show and keep up associations with companions. As OSNs keep on gaining more fame, the phenomenal measure of individual data and social relations enhance sociology research where it was once constrained by an absence of information. As an exploration, the part of unequivocal social relations in recommender frameworks is as an important part of the research, for example, how client inclinations or evaluations are connected with those of neighbors, and how to utilize such relationships to outline a superior recommender framework. Specifically, a calculation structure is planned which makes suggestions taking into account client's own particular inclinations, the general acknowledgment of the objective thing, and the assessments from social networks. A genuine online social network data from last.fm has been crawled as a contextual investigation, and perform broad examination on this dataset. Additionally, the dataset is utilized, accumulated from the social network, to assess the execution of the proposed framework on the scalability, data sparsity, and cold start. The exploratory aftereffects of our framework show critical change against customary community oriented sifting in those perspectives. For instance, the computed precision in the wake of running the contextual analysis has enhanced by 0.7498 contrasted with conventional shared separating. Moreover, it is proposed to utilize the semantics of client connections by their similitudes and better grained client appraisals to enhance the expectation exactness

    Extracting ontological structures from collaborative tagging systems

    Get PDF
    corecore