22 research outputs found

    FluxMap: A VANTED add-on for the visual exploration of flux distributions in biological networks

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>The quantification of metabolic fluxes is gaining increasing importance in the analysis of the metabolic behavior of biological systems such as organisms, tissues or cells. Various methodologies (wetlab or drylab) result in sets of fluxes which require an appropriate visualization for interpretation by scientists. The visualization of flux distributions is a necessary prerequisite for intuitive flux data exploration in the context of metabolic networks.</p> <p>Results</p> <p>We present FluxMap, a tool for the advanced visualization and exploration of flux data in the context of metabolic networks. The template-based flux data import assigns flux values and optional quality parameters (e. g. the confidence interval) to biochemical reactions. It supports the discrimination between mass and substance fluxes, such as C- or N-fluxes. After import, flux data mapping and network-based visualization allow the interactive exploration of the dataset. Various visualization options enable the user to adapt layout and network representation according to individual purposes.</p> <p>Conclusions</p> <p>The Vanted add-on FluxMap comprises a comprehensive set of functionalities for visualization and advanced visual exploration of flux distributions in biological networks. It is available as a Java open source tool from http://www.vanted.org/fluxmap.</p

    Mass conserved elementary kinetics is sufficient for the existence of a non-equilibrium steady state concentration

    Get PDF
    Living systems are forced away from thermodynamic equilibrium by exchange of mass and energy with their environment. In order to model a biochemical reaction network in a non-equilibrium state one requires a mathematical formulation to mimic this forcing. We provide a general formulation to force an arbitrary large kinetic model in a manner that is still consistent with the existence of a non-equilibrium steady state. We can guarantee the existence of a non-equilibrium steady state assuming only two conditions; that every reaction is mass balanced and that continuous kinetic reaction rate laws never lead to a negative molecule concentration. These conditions can be verified in polynomial time and are flexible enough to permit one to force a system away from equilibrium. In an expository biochemical example we show how a reversible, mass balanced perpetual reaction, with thermodynamically infeasible kinetic parameters, can be used to perpetually force a kinetic model of anaerobic glycolysis in a manner consistent with the existence of a steady state. Easily testable existence conditions are foundational for efforts to reliably compute non-equilibrium steady states in genome-scale biochemical kinetic models.Comment: 11 pages, 2 figures (v2 is now placed in proper context of the excellent 1962 paper by James Wei entitled "Axiomatic treatment of chemical reaction systems". In addition, section 4, on "Utility of steady state existence theorem" has been expanded.

    Updates in metabolomics tools and resources: 2014-2015

    Get PDF
    Data processing and interpretation represent the most challenging and time-consuming steps in high-throughput metabolomic experiments, regardless of the analytical platforms (MS or NMR spectroscopy based) used for data acquisition. Improved machinery in metabolomics generates increasingly complex datasets that create the need for more and better processing and analysis software and in silico approaches to understand the resulting data. However, a comprehensive source of information describing the utility of the most recently developed and released metabolomics resources—in the form of tools, software, and databases—is currently lacking. Thus, here we provide an overview of freely-available, and open-source, tools, algorithms, and frameworks to make both upcoming and established metabolomics researchers aware of the recent developments in an attempt to advance and facilitate data processing workflows in their metabolomics research. The major topics include tools and researches for data processing, data annotation, and data visualization in MS and NMR-based metabolomics. Most in this review described tools are dedicated to untargeted metabolomics workflows; however, some more specialist tools are described as well. All tools and resources described including their analytical and computational platform dependencies are summarized in an overview Table

    Toward systems biology in brown algae to explore acclimation and adaptation to the shore environment.

    Get PDF
    International audienceBrown algae belong to a phylogenetic lineage distantly related to land plants and animals. They are almost exclusively found in the intertidal zone, a harsh and frequently changing environment where organisms are submitted to marine and terrestrial constraints. In relation with their unique evolutionary history and their habitat, they feature several peculiarities, including at the level of their primary and secondary metabolism. The establishment of Ectocarpus siliculosus as a model organism for brown algae has represented a framework in which several omics techniques have been developed, in particular, to study the response of these organisms to abiotic stresses. With the recent publication of medium to high throughput profiling data, it is now possible to envision integrating observations at the cellular scale to apply systems biology approaches. As a first step, we propose a protocol focusing on integrating heterogeneous knowledge gained on brown algal metabolism. The resulting abstraction of the system will then help understanding how brown algae cope with changes in abiotic parameters within their unique habitat, and to decipher some of the mechanisms underlying their (1) acclimation and (2) adaptation, respectively consequences of (1) the behavior or (2) the topology of the system resulting from the integrative approach

    Robustifying Experimental Tracer Design for13C-Metabolic Flux Analysis

    Get PDF
    13C metabolic flux analysis (MFA) has become an indispensable tool to measure metabolic reaction rates (fluxes) in living organisms, having an increasingly diverse range of applications. Here, the choice of the13C labeled tracer composition makes the difference between an information-rich experiment and an experiment with only limited insights. To improve the chances for an informative labeling experiment, optimal experimental design approaches have been devised for13C-MFA, all relying on some a priori knowledge about the actual fluxes. If such prior knowledge is unavailable, e.g., for research organisms and producer strains, existing methods are left with a chicken-and-egg problem. In this work, we present a general computational method, termed robustified experimental design (R-ED), to guide the decision making about suitable tracer choices when prior knowledge about the fluxes is lacking. Instead of focusing on one mixture, optimal for specific flux values, we pursue a sampling based approach and introduce a new design criterion, which characterizes the extent to which mixtures are informative in view of all possible flux values. The R-ED workflow enables the exploration of suitable tracer mixtures and provides full flexibility to trade off information and cost metrics. The potential of the R-ED workflow is showcased by applying the approach to the industrially relevant antibiotic producer Streptomyces clavuligerus, where we suggest informative, yet economic labeling strategies

    Scientific Workflows for Metabolic Flux Analysis

    Get PDF
    Metabolic engineering is a highly interdisciplinary research domain that interfaces biology, mathematics, computer science, and engineering. Metabolic flux analysis with carbon tracer experiments (13 C-MFA) is a particularly challenging metabolic engineering application that consists of several tightly interwoven building blocks such as modeling, simulation, and experimental design. While several general-purpose workflow solutions have emerged in recent years to support the realization of complex scientific applications, the transferability of these approaches are only partially applicable to 13C-MFA workflows. While problems in other research fields (e.g., bioinformatics) are primarily centered around scientific data processing, 13C-MFA workflows have more in common with business workflows. For instance, many bioinformatics workflows are designed to identify, compare, and annotate genomic sequences by "pipelining" them through standard tools like BLAST. Typically, the next workflow task in the pipeline can be automatically determined by the outcome of the previous step. Five computational challenges have been identified in the endeavor of conducting 13 C-MFA studies: organization of heterogeneous data, standardization of processes and the unification of tools and data, interactive workflow steering, distributed computing, and service orientation. The outcome of this thesis is a scientific workflow framework (SWF) that is custom-tailored for the specific requirements of 13 C-MFA applications. The proposed approach – namely, designing the SWF as a collection of loosely-coupled modules that are glued together with web services – alleviates the realization of 13C-MFA workflows by offering several features. By design, existing tools are integrated into the SWF using web service interfaces and foreign programming language bindings (e.g., Java or Python). Although the attributes "easy-to-use" and "general-purpose" are rarely associated with distributed computing software, the presented use cases show that the proposed Hadoop MapReduce framework eases the deployment of computationally demanding simulations on cloud and cluster computing resources. An important building block for allowing interactive researcher-driven workflows is the ability to track all data that is needed to understand and reproduce a workflow. The standardization of 13 C-MFA studies using a folder structure template and the corresponding services and web interfaces improves the exchange of information for a group of researchers. Finally, several auxiliary tools are developed in the course of this work to complement the SWF modules, i.e., ranging from simple helper scripts to visualization or data conversion programs. This solution distinguishes itself from other scientific workflow approaches by offering a system of loosely-coupled components that are flexibly arranged to match the typical requirements in the metabolic engineering domain. Being a modern and service-oriented software framework, new applications are easily composed by reusing existing components

    Omics visualization and its application to presymptomatic diagnosis of oral cancer

    Get PDF
    Brink B. Omics visualization and its application to presymptomatic diagnosis of oral cancer. Bielefeld: Universität Bielefeld; 2018.About 30 zettabytes (30 · 10^21 bytes) of data are generated worldwide every second — so much that over 90 % of the data in the world today has been created in the last two years alone. Science as well is flooded by an ever increasing amount of data. However, accessing the infor- mation hidden in this massive amount of data is a challenging task and in science often presence a hindrance to knowledge discovery. One way to overcome this is a good visualization, which can greatly support people and scientists in exploring, understanding, and enjoy- ing data. In this thesis, I present three examples for a task oriented visualization in some of the most data-rich disciplines in science: bio- chemistry, healthcare, and biology. The first example is situated in the field of biochemistry. Since the 1980s, natural sciences challenged educational institutions and media to keep the society on an appropriate level of knowledge and un- derstanding. By investigating the potential of infographics, graphical design, and game motivation, I present a mnemonic card game based on creative design to aid the learning of a special group of biomol- ecules, the amino acids. Each amino acid is composed of a number of features. The latter are intuitively encoded into shapes, colors, and textures to assist our abilities in interpreting visual stimuli. Thus, it facilitates recognizing such features, grouping them, noting relation- ships, and ultimately memorizing the structural formulas. The cards translate complex molecular structures into visual formats that are both easier to assess and to understand. The result is a unique teach- ing tool that is not only subject-oriented, fun, and engaging, but also helps students retain relevant information such as properties and for- mulas through perceptual memory. The second example tackles a problem from the field of healthcare. Oral cancer has a major impact worldwide, accounting for 274 000 new cases and 145 000 deaths each year, making it sixth most com- mon cancer. Developing methods for the detection of cancer in its earliest stages can greatly increase the chances for a successful treat- ment. Many cancers (including oral cancer) are known to develop- ment through multiple steps, which are caused by certain mutations to the genome. A recently published protocol by Hughesman et al. (2016) describes means for high-throughput detection of these mu- tations using droplet digital PCR. However, methods for automated analysis and visualization of this data are unavailable. In this the- sis, I present ddPCRclust, an R package for automated analysis of droplet digital PCR data. It can automatically analyze and visualize data from droplet digital PCR experiments with up to four targets per reaction in a non-orthogonal layout. Results are on a par with manual analysis, but only take minutes to compute instead of hours. The ac- companying Shiny application ddPCRvis provides easy access to the functionalities of ddPCRclust through a web-browser based graphical user interface, enabling the user to interactively filter data and change parameters, as well as view and modify results. The third example involves some of the most data-rich disciplines in biology - transcriptomics, proteomics, and metabolomics. Omics Fusion is a web based platform for the integrative analysis of omics data. It provides a collection of new and established tools and visual- ization methods to support researchers in exploring omics data, vali- dating results, or understanding how to adjust experiments in order to make new discoveries. It is easily extendible and new visualization methods are added continuously. I present an example for a task- oriented visualization of functional annotated omics data based on the established Clusters of Orthologous Groups (COG) database and gene ontology (GO) terms

    NOVEL APPLICATIONS OF MACHINE LEARNING IN BIOINFORMATICS

    Get PDF
    Technological advances in next-generation sequencing and biomedical imaging have led to a rapid increase in biomedical data dimension and acquisition rate, which is challenging the conventional data analysis strategies. Modern machine learning techniques promise to leverage large data sets for finding hidden patterns within them, and for making accurate predictions. This dissertation aims to design novel machine learning-based models to transform biomedical big data into valuable biological insights. The research presented in this dissertation focuses on three bioinformatics domains: splice junction classification, gene regulatory network reconstruction, and lesion detection in mammograms. A critical step in defining gene structures and mRNA transcript variants is to accurately identify splice junctions. In the first work, we built the first deep learning-based splice junction classifier, DeepSplice. It outperforms the state-of-the-art classification tools in terms of both classification accuracy and computational efficiency. To uncover transcription factors governing metabolic reprogramming in non-small-cell lung cancer patients, we developed TFmeta, a machine learning approach to reconstruct relationships between transcription factors and their target genes in the second work. Our approach achieves the best performance on benchmark data sets. In the third work, we designed deep learning-based architectures to perform lesion detection in both 2D and 3D whole mammogram images
    corecore