49,532 research outputs found
NaviCell: a web-based environment for navigation, curation and maintenance of large molecular interaction maps
Molecular biology knowledge can be systematically represented in a
computer-readable form as a comprehensive map of molecular interactions. There
exist a number of maps of molecular interactions containing detailed
description of various cell mechanisms. It is difficult to explore these large
maps, to comment their content and to maintain them. Though there exist several
tools addressing these problems individually, the scientific community still
lacks an environment that combines these three capabilities together. NaviCell
is a web-based environment for exploiting large maps of molecular interactions,
created in CellDesigner, allowing their easy exploration, curation and
maintenance. NaviCell combines three features: (1) efficient map browsing based
on Google Maps engine; (2) semantic zooming for viewing different levels of
details or of abstraction of the map and (3) integrated web-based blog for
collecting the community feedback. NaviCell can be easily used by experts in
the field of molecular biology for studying molecular entities of their
interest in the context of signaling pathways and cross-talks between pathways
within a global signaling network. NaviCell allows both exploration of detailed
molecular mechanisms represented on the map and a more abstract view of the map
up to a top-level modular representation. NaviCell facilitates curation,
maintenance and updating the comprehensive maps of molecular interactions in an
interactive fashion due to an imbedded blogging system. NaviCell provides an
easy way to explore large-scale maps of molecular interactions, thanks to the
Google Maps and WordPress interfaces, already familiar to many users. Semantic
zooming used for navigating geographical maps is adopted for molecular maps in
NaviCell, making any level of visualization meaningful to the user. In
addition, NaviCell provides a framework for community-based map curation.Comment: 20 pages, 5 figures, submitte
XML in Motion from Genome to Drug
Information technology (IT) has emerged as a central to the solution of contemporary genomics and drug discovery problems. Researchers involved in genomics, proteomics, transcriptional profiling, high throughput structure determination, and in other sub-disciplines of bioinformatics have direct impact on this IT revolution. As the full genome sequences of many species, data from structural genomics, micro-arrays, and proteomics became available, integration of these data to a common platform require sophisticated bioinformatics tools. Organizing these data into knowledgeable databases and developing appropriate software tools for analyzing the same are going to be major challenges. XML (eXtensible Markup Language) forms the backbone of biological data representation and exchange over the internet, enabling researchers to aggregate data from various heterogeneous data resources. The present article covers a comprehensive idea of the integration of XML on particular type of biological databases mainly dealing with sequence-structure-function relationship and its application towards drug discovery. This e-medical science approach should be applied to other scientific domains and the latest trend in semantic web applications is also highlighted
A Molecular Biology Database Digest
Computational Biology or Bioinformatics has been defined as the application of mathematical
and Computer Science methods to solving problems in Molecular Biology that require large scale
data, computation, and analysis [18]. As expected, Molecular Biology databases play an essential
role in Computational Biology research and development. This paper introduces into current
Molecular Biology databases, stressing data modeling, data acquisition, data retrieval, and the
integration of Molecular Biology data from different sources. This paper is primarily intended
for an audience of computer scientists with a limited background in Biology
Semantic Storage: Overview and Assessment
The Semantic Web has a great deal of momentum behind it. The promise of a ‘better web’, where information is given well defined meaning and computers are better able to work with it has captured the imagination of a significant number of people, particularly in academia. Language standards such as RDF and OWL have appeared with remarkable speed, and development continues apace. To back up this development, there is a requirement for ‘semantic databases’, where this data can be conveniently stored, operated upon, and retrieved. These already exist in the form of triple stores, but do not yet fulfil all the requirements that may be made of them, particularly in the area of performing inference using OWL. This paper analyses the current stores along with forthcoming technology, and finds that it is unlikely that a combination of speed, scalability, and complex inferencing will be practical in the immediate future. It concludes by suggesting alternative development routes
Storage Solutions for Big Data Systems: A Qualitative Study and Comparison
Big data systems development is full of challenges in view of the variety of
application areas and domains that this technology promises to serve.
Typically, fundamental design decisions involved in big data systems design
include choosing appropriate storage and computing infrastructures. In this age
of heterogeneous systems that integrate different technologies for optimized
solution to a specific real world problem, big data system are not an exception
to any such rule. As far as the storage aspect of any big data system is
concerned, the primary facet in this regard is a storage infrastructure and
NoSQL seems to be the right technology that fulfills its requirements. However,
every big data application has variable data characteristics and thus, the
corresponding data fits into a different data model. This paper presents
feature and use case analysis and comparison of the four main data models
namely document oriented, key value, graph and wide column. Moreover, a feature
analysis of 80 NoSQL solutions has been provided, elaborating on the criteria
and points that a developer must consider while making a possible choice.
Typically, big data storage needs to communicate with the execution engine and
other processing and visualization technologies to create a comprehensive
solution. This brings forth second facet of big data storage, big data file
formats, into picture. The second half of the research paper compares the
advantages, shortcomings and possible use cases of available big data file
formats for Hadoop, which is the foundation for most big data computing
technologies. Decentralized storage and blockchain are seen as the next
generation of big data storage and its challenges and future prospects have
also been discussed
Knowledge organization systems in mathematics and in libraries
Based on the project activities planned in the context of the Specialized Information Service for Mathematics (TIB Hannover, FAU Erlangen, L3S, SUB Göttingen) we give an overview over the history and interplay of subject cataloguing in libraries, the development of computerized methods for metadata processing and the rise of the Semantic Web. We survey various knowledge organization systems such as the Mathematics Subject Classification, the German Authority File, the clustering International Authority File VIAF, and lexical databases such as WordNet and their potential use for mathematics in education and research. We briefly address the difference between thesauri and ontologies and the relations they typically contain from a linguistic perspective. We will then discuss with the audience how the current efforts to represent and handle mathematical theories as semantic objects can help deflect the decline of semantic resource annotation in libraries that has been predicted by some due to the existence of highly performant retrieval algorithms (based on statistical, neuronal, or other big data methods). We will also explore the potential characteristics of a fruitful symbiosis between carefully cultivated kernels of semantic structure and automated methods in order to scale those structures up to the level that is necessary in order to cope with the amounts of digital data found in libraries and in (mathematical) research (e.g., in simulations) today
A Query Integrator and Manager for the Query Web
We introduce two concepts: the Query Web as a layer of interconnected queries over the document web and the semantic web, and a Query Web Integrator and Manager (QI) that enables the Query Web to evolve. QI permits users to write, save and reuse queries over any web accessible source, including other queries saved in other installations of QI. The saved queries may be in any language (e.g. SPARQL, XQuery); the only condition for interconnection is that the queries return their results in some form of XML. This condition allows queries to chain off each other, and to be written in whatever language is appropriate for the task. We illustrate the potential use of QI for several biomedical use cases, including ontology view generation using a combination of graph-based and logical approaches, value set generation for clinical data management, image annotation using terminology obtained from an ontology web service, ontology-driven brain imaging data integration, small-scale clinical data integration, and wider-scale clinical data integration. Such use cases illustrate the current range of applications of QI and lead us to speculate about the potential evolution from smaller groups of interconnected queries into a larger query network that layers over the document and semantic web. The resulting Query Web could greatly aid researchers and others who now have to manually navigate through multiple information sources in order to answer specific questions
- …