786 research outputs found

    Transforming the study of organisms: Phenomic data models and knowledge bases

    Get PDF
    The rapidly decreasing cost of gene sequencing has resulted in a deluge of genomic data from across the tree of life; however, outside a few model organism databases, genomic data are limited in their scientific impact because they are not accompanied by computable phenomic data. The majority of phenomic data are contained in countless small, heterogeneous phenotypic data sets that are very difficult or impossible to integrate at scale because of variable formats, lack of digitization, and linguistic problems. One powerful solution is to represent phenotypic data using data models with precise, computable semantics, but adoption of semantic standards for representing phenotypic data has been slow, especially in biodiversity and ecology. Some phenotypic and trait data are available in a semantic language from knowledge bases, but these are often not interoperable. In this review, we will compare and contrast existing ontology and data models, focusing on nonhuman phenotypes and traits. We discuss barriers to integration of phenotypic data and make recommendations for developing an operationally useful, semantically interoperable phenotypic data ecosystem

    Ontologias para Manutenção Preditiva com Dados sensíveis ao tempo

    Get PDF
    As empresas de fabrico industrial devem assegurar um processo produtivo contínuo para serem competitivas e fornecer os produtos fabricados no prazo e com a qualidade exigida pelos clientes. A quebra da cadeia de fabrico pode ter desfechos graves, resultando numa redução da produção e na interrupção da cadeia de abastecimento. Estes processos são compostos por cadeias de máquinas que executam tarefas em etapas. Cada máquina tem uma tarefa específica a executar, e o resultado de cada etapa é fornecido à próxima etapa. Uma falha imprevista numa das máquinas tende a interromper toda a cadeia produtiva. A manutenção preventiva agendada tem como objetivo evitar a ocorrência de falhas, tendo como base o tempo médio antes da falha (MTBF), que representa a expectativa média de vida de componentes individuais com base em dados históricos. As tarefas de manutenção podem implicar um período de paralisação e a interrupção da produção. Esta manutenção é executada rotineiramente e a substituição de componentes não considera a necessidade premente da sua substituição, sendo os mesmos substituídos com base no ciclo do agendamento. É aqui que a manutenção preditiva é aplicável. Efetuando a recolha de dados de sensores dos equipamentos, é possível detetar irregularidades nos dados recolhidos, através da aplicação de processos de raciocínio e inferência, conduzindo à atempada previsão e deteção de falhas. Levando este cenário à otimização do tempo de manutenção, evitando falhas inesperadas, à redução de custos e ao aumento da produtividade em comparação com a manutenção preventiva. Os dados fornecidos pelos sensores são sensíveis ao tempo, variações e flutuações ocorrem ao longo do tempo e devem ser analisados em relação ao período em que ocorrem. Esta dissertação tem como objetivo o desenvolvimento de uma ontologia para a manutenção preditiva que descreva a sua abrangência e o campo da sua aplicação. A aplicabilidade da ontologia será demonstrada com uma ferramenta, igualmente desenvolvida, que transforma dados sensíveis ao tempo recolhidos em tempo real a partir de sensores de máquinas industriais, fornecidos por WebServices, em indivíduos dessa mesma ontologia, considerando a representação do fator temporal dos dados.Manufacturing companies must ensure a continuous production process to be competitive and supply the manufactured goods in time and with the desired quality the customers expect. Any disruption in the manufacturing chain may have disastrous consequences, representing a shortage of production and the interruption of the supply chain. The manufacturing processes are composed of a chain of industrial machines operating in stages. Each machine has a specific task to complete, and the result of each stage is forwarded to the next stage. An unpredicted malfunction of one of the machines tends to interrupt the whole production chain. Scheduled Preventive maintenance intends to avoid causes leading to faults, but relies on parameters such as Mean Time Before Failure (MTBF), which represents the average expected life span of individual components based on statistical data. A maintenance task may lead to a period of downtime and consequently to a production halt. Being the maintenance scheduled and executed routinely, the replacement of components, does not consider the effective need of its replacement, they are replaced based on the scheduling cycle. This is where predictive maintenance is applicable. By collecting sensor data of industrial equipment, anomalies can be determined through reasoning and inference processes applied to the data, leading to an early fault and time to failure prediction. This scenario leads to maintenance timing optimization, avoidance of unexpected failures, cost savings and improved productivity when compared to preventive maintenance. Data supplied by sensors is timesensitive, as variations and fluctuations occur over periods of time and must be analysed concerning the period they occur. This dissertation aims to develop an ontology for predictive maintenance that describes the scope and field of application. The applicability of the ontology will be demonstrated with a tool, also to be developed, that transforms time-sensitive data collected in real time from sensors of industrial machines, provided by a WebServices, into individuals of the same ontology, considering the representation of the temporal factor of the data

    A framework for integrating syntax, semantics and pragmatics for computer-aided professional practice: With application of costing in construction industry

    Get PDF
    Producing a bill of quantity is a knowledge-based, dynamic and collaborative process, and evolves with variances and current evidence. However, within the context of information system practice in BIM, knowledge of cost estimation has not been represented, nor has it been integrated into the processes based on BIM. This paper intends to establish an innovative means of taking data from the BIM linked to a project, and using it to create the necessary items for a bill of quantity that will enable cost estimation to be undertaken for the project. Our framework is founded upon the belief that three components are necessary to gain a full awareness of the domain which is being computerised; the information type which is to be assessed for compatibility (syntax), the definition for the pricing domain (semantics), and the precise implementation environment for the standards being taken into account (pragmatics). In order to achieve this, a prototype is created that allows a cost item for the bill of quantity to be spontaneously generated, by means of the semantic web ontology and a forward chain algorithm. Within this paper, ‘cost items’ signify the elements included in a bill of quantity, including details of their description, quantity and price. As a means of authenticating the process being developed, the authors of this work effectively implemented it in the production of cost items. In addition, the items created were contrasted with those produced by specialists. For this reason, this innovative framework introduces the possibility of a new means of applying semantic web ontology and forward chain algorithm to construction professional practice resulting in automatic cost estimation. These key outcomes demonstrate that, decoupling the professional practice into three key components of syntax, semantics and pragmatics can provide tangible benefits to domain use

    Fusing Automatically Extracted Annotations for the Semantic Web

    Get PDF
    This research focuses on the problem of semantic data fusion. Although various solutions have been developed in the research communities focusing on databases and formal logic, the choice of an appropriate algorithm is non-trivial because the performance of each algorithm and its optimal configuration parameters depend on the type of data, to which the algorithm is applied. In order to be reusable, the fusion system must be able to select appropriate techniques and use them in combination. Moreover, because of the varying reliability of data sources and algorithms performing fusion subtasks, uncertainty is an inherent feature of semantically annotated data and has to be taken into account by the fusion system. Finally, the issue of schema heterogeneity can have a negative impact on the fusion performance. To address these issues, we propose KnoFuss: an architecture for Semantic Web data integration based on the principles of problem-solving methods. Algorithms dealing with different fusion subtasks are represented as components of a modular architecture, and their capabilities are described formally. This allows the architecture to select appropriate methods and configure them depending on the processed data. In order to handle uncertainty, we propose a novel algorithm based on the Dempster-Shafer belief propagation. KnoFuss employs this algorithm to reason about uncertain data and method results in order to refine the fused knowledge base. Tests show that these solutions lead to improved fusion performance. Finally, we addressed the problem of data fusion in the presence of schema heterogeneity. We extended the KnoFuss framework to exploit results of automatic schema alignment tools and proposed our own schema matching algorithm aimed at facilitating data fusion in the Linked Data environment. We conducted experiments with this approach and obtained a substantial improvement in performance in comparison with public data repositories

    Supervised Typing of Big Graphs using Semantic Embeddings

    Full text link
    We propose a supervised algorithm for generating type embeddings in the same semantic vector space as a given set of entity embeddings. The algorithm is agnostic to the derivation of the underlying entity embeddings. It does not require any manual feature engineering, generalizes well to hundreds of types and achieves near-linear scaling on Big Graphs containing many millions of triples and instances by virtue of an incremental execution. We demonstrate the utility of the embeddings on a type recommendation task, outperforming a non-parametric feature-agnostic baseline while achieving 15x speedup and near-constant memory usage on a full partition of DBpedia. Using state-of-the-art visualization, we illustrate the agreement of our extensionally derived DBpedia type embeddings with the manually curated domain ontology. Finally, we use the embeddings to probabilistically cluster about 4 million DBpedia instances into 415 types in the DBpedia ontology.Comment: 6 pages, to be published in Semantic Big Data Workshop at ACM, SIGMOD 2017; extended version in preparation for Open Journal of Semantic Web (OJSW

    Dwelling on ontology - semantic reasoning over topographic maps

    Get PDF
    The thesis builds upon the hypothesis that the spatial arrangement of topographic features, such as buildings, roads and other land cover parcels, indicates how land is used. The aim is to make this kind of high-level semantic information explicit within topographic data. There is an increasing need to share and use data for a wider range of purposes, and to make data more definitive, intelligent and accessible. Unfortunately, we still encounter a gap between low-level data representations and high-level concepts that typify human qualitative spatial reasoning. The thesis adopts an ontological approach to bridge this gap and to derive functional information by using standard reasoning mechanisms offered by logic-based knowledge representation formalisms. It formulates a framework for the processes involved in interpreting land use information from topographic maps. Land use is a high-level abstract concept, but it is also an observable fact intimately tied to geography. By decomposing this relationship, the thesis correlates a one-to-one mapping between high-level conceptualisations established from human knowledge and real world entities represented in the data. Based on a middle-out approach, it develops a conceptual model that incrementally links different levels of detail, and thereby derives coarser, more meaningful descriptions from more detailed ones. The thesis verifies its proposed ideas by implementing an ontology describing the land use ‘residential area’ in the ontology editor Protégé. By asserting knowledge about high-level concepts such as types of dwellings, urban blocks and residential districts as well as individuals that link directly to topographic features stored in the database, the reasoner successfully infers instances of the defined classes. Despite current technological limitations, ontologies are a promising way forward in the manner we handle and integrate geographic data, especially with respect to how humans conceptualise geographic space

    The role of ontologies in biological and biomedical research: a functional perspective.

    Get PDF
    Ontologies are widely used in biological and biomedical research. Their success lies in their combination of four main features present in almost all ontologies: provision of standard identifiers for classes and relations that represent the phenomena within a domain; provision of a vocabulary for a domain; provision of metadata that describes the intended meaning of the classes and relations in ontologies; and the provision of machine-readable axioms and definitions that enable computational access to some aspects of the meaning of classes and relations. While each of these features enables applications that facilitate data integration, data access and analysis, a great potential lies in the possibility of combining these four features to support integrative analysis and interpretation of multimodal data. Here, we provide a functional perspective on ontologies in biology and biomedicine, focusing on what ontologies can do and describing how they can be used in support of integrative research. We also outline perspectives for using ontologies in data-driven science, in particular their application in structured data mining and machine learning applications.This is the final version of the article. It first appeared from Oxford University Press via http://dx.doi.org/10.1093/bib/bbv01

    On the Role of Assertions for Conceptual Modeling as Enablers of Composable Simulation Solutions

    Get PDF
    This research provides a much needed systematic review of the roles that assertions play in model composability and simulation interoperability. In doing so, this research contributes a partial solution to one of the problems of model composability and simulation interoperability—namely, why do simulation systems fail to achieve the maximum level of interoperability possible? It demonstrates the importance of the assertions that are made during model development and simulation implementation, particularly as they reflect the unique viewpoint of each developer or user. It hypothesizes that it is possible to detect composability conflicts by means of a four-step process developed by the author for capturing and comparing assertions. It demonstrates the process using a well understood example problem—the Falling Body Problem—developing a formal model of assertion, a strategy for assertion comparison, an inventory of forces, and a catalog of significant assertions that might be made for each term in the solution to the problem. Finally, it develops a software application to implement the strategy for comparing sets of assertions. The software successfully detects potential conflicts between ontologies that were otherwise determined to be ontologically consistent, thus proving the hypothesis
    • …
    corecore