67 research outputs found
Monthly hydrometeorological ensemble prediction of streamflow droughts and corresponding drought indices
Streamflow droughts, characterized by low runoff as consequence of a drought event, affect numerous aspects of life. Economic sectors that are impacted by low streamflow are, e.g., power production, agriculture, tourism, water quality management and shipping. Those sectors could potentially benefit from forecasts of streamflow drought events, even of short events on the monthly time scales or below. Numerical hydrometeorological models have increasingly been used to forecast low streamflow and have become the focus of recent research. Here, we consider daily ensemble runoff forecasts for the river Thur, which has its source in the Swiss Alps. We focus on the evaluation of low streamflow and of the derived indices as duration, severity and magnitude, characterizing streamflow droughts up to a lead time of one month. <br><br> The ECMWF VarEPS 5-member ensemble reforecast, which covers 18 yr, is used as forcing for the hydrological model PREVAH. A thorough verification reveals that, compared to probabilistic peak-flow forecasts, which show skill up to a lead time of two weeks, forecasts of streamflow droughts are skilful over the entire forecast range of one month. For forecasts at the lower end of the runoff regime, the quality of the initial state seems to be crucial to achieve a good forecast quality in the longer range. It is shown that the states used in this study to initialize forecasts satisfy this requirement. The produced forecasts of streamflow drought indices, derived from the ensemble forecasts, could be beneficially included in a decision-making process. This is valid for probabilistic forecasts of streamflow drought events falling below a daily varying threshold, based on a quantile derived from a runoff climatology. Although the forecasts have a tendency to overpredict streamflow droughts, it is shown that the relative economic value of the ensemble forecasts reaches up to 60%, in case a forecast user is able to take preventive action based on the forecast
Homogenisation of a gridded snow water equivalent climatology for Alpine terrain: methodology and applications
Gridded snow water equivalent (SWE) data sets are valuable for estimating the
snow water resources and verify different model systems, e.g. hydrological,
land surface or atmospheric models. However, changing data availability
represents a considerable challenge when trying to derive consistent time
series for SWE products. In an attempt to improve the product consistency, we
first evaluated the differences between two climatologies of SWE grids that
were calculated on the basis of data from 110 and 203 stations, respectively.
The "shorter" climatology (2001–2009) was produced using 203 stations
(map203) and the "longer" one (1971–2009) 110 stations
(map110). Relative to map203, map110
underestimated SWE, especially at higher elevations and at the end of the
winter season. We tested the potential of quantile mapping to compensate for
mapping errors in map110 relative to map203. During a
9 yr calibration period from 2001 to 2009, for which both map203
and map110 were available, the method could successfully refine the
spatial and temporal SWE representation in map110 by making
seasonal, regional and altitude-related distinctions. Expanding the
calibration to the full 39 yr showed that the general underestimation of
map110 with respect to map203 could be removed for the
whole winter. The calibrated SWE maps fitted the reference (map203)
well when averaged over regions and time periods, where the mean error is
approximately zero. However, deviations between the calibrated maps and
map203 were observed at single grid cells and years. When we looked
at three different regions in more detail, we found that the calibration had
the largest effect in the region with the highest proportion of catchment
areas above 2000 m a.s.l. and that the general underestimation of
map110 compared to map203 could be removed for the entire
snow season. The added value of the calibrated SWE climatology is illustrated
with practical examples: the verification of a hydrological model, the
estimation of snow resource anomalies and the predictability of runoff
through SWE
"EDML1": a chronology for the EPICA deep ice core from Dronning Maud Land, Antarctica, over the last 150 000 years.
A chronology called EDML1 has been developed for the EPICA ice core from Dronning Maud Land (EDML). EDML1 is closely interlinked with EDC3, the new chronology for the EPICA ice core from Dome-C (EDC) through a stratigraphic match between EDML and EDC that consists of 322 volcanic match points over the last 128 ka. The EDC3 chronology comprises a glaciological model at EDC, which is constrained and later selectively tuned using primary dating information from EDC as well as from EDML, the latter being transferred using the tight stratigraphic link between the two cores. Finally, EDML1 was built by exporting EDC3 to EDML. For ages younger than 41 ka BP the new synchronized time scale EDML1/EDC3 is based on dated volcanic events and on a match to the Greenlandic ice core chronology GICC05 via <sup>10</sup>Be and methane. The internal consistency between EDML1 and EDC3 is estimated to be typically ~6 years and always less than 450 years over the last 128 ka (always less than 130 years over the last 60 ka), which reflects an unprecedented synchrony of time scales. EDML1 ends at 150 ka BP (2417 m depth) because the match between EDML and EDC becomes ambiguous further down. This hints at a complex ice flow history for the deepest 350 m of the EDML ice core
Recommended from our members
How do I know if my forecasts are better? Using benchmarks in Hydrological ensemble prediction
The skill of a forecast can be assessed by comparing the relative proximity of both the forecast and a benchmark to the observations. Example benchmarks include climatology or a naïve forecast. Hydrological ensemble prediction systems (HEPS) are currently transforming the hydrological forecasting environment but in this new field there is little information to guide researchers and operational forecasters on how benchmarks can be best used to evaluate their probabilistic forecasts. In this study, it is identified that the forecast skill calculated can vary depending on the benchmark selected and that the selection of a benchmark for determining forecasting system skill is sensitive to a number of hydrological and system factors. A benchmark intercomparison experiment is then undertaken using the continuous ranked probability score (CRPS), a reference forecasting system and a suite of 23 different methods to derive benchmarks. The benchmarks are assessed within the operational set-up of the European Flood Awareness System (EFAS) to determine those that are ‘toughest to beat’ and so give the most robust discrimination of forecast skill, particularly for the spatial average fields that EFAS relies upon.
Evaluating against an observed discharge proxy the benchmark that has most utility for EFAS and avoids the most naïve skill across different hydrological situations is found to be meteorological persistency. This benchmark uses the latest meteorological observations of precipitation and temperature to drive the hydrological model. Hydrological long term average benchmarks, which are currently used in EFAS, are very easily beaten by the forecasting system and the use of these produces much naïve skill. When decomposed into seasons, the advanced meteorological benchmarks, which make use of meteorological observations from the past 20 years at the same calendar date, have the most skill discrimination. They are also good at discriminating skill in low flows and for all catchment sizes. Simpler meteorological benchmarks are particularly useful for high flows. Recommendations for EFAS are to move to routine use of meteorological persistency, an advanced meteorological benchmark and a simple meteorological benchmark in order to provide a robust evaluation of forecast skill. This work provides the first comprehensive evidence on how benchmarks can be used in evaluation of skill in probabilistic hydrological forecasts and which benchmarks are most useful for skill discrimination and avoidance of naïve skill in a large scale HEPS. It is recommended that all HEPS use the evidence and methodology provided here to evaluate which benchmarks to employ; so forecasters can have trust in their skill evaluation and will have confidence that their forecasts are indeed better
A graph-search framework for associating gene identifiers with documents
BACKGROUND: One step in the model organism database curation process is to find, for each article, the identifier of every gene discussed in the article. We consider a relaxation of this problem suitable for semi-automated systems, in which each article is associated with a ranked list of possible gene identifiers, and experimentally compare methods for solving this geneId ranking problem. In addition to baseline approaches based on combining named entity recognition (NER) systems with a "soft dictionary" of gene synonyms, we evaluate a graph-based method which combines the outputs of multiple NER systems, as well as other sources of information, and a learning method for reranking the output of the graph-based method. RESULTS: We show that named entity recognition (NER) systems with similar F-measure performance can have significantly different performance when used with a soft dictionary for geneId-ranking. The graph-based approach can outperform any of its component NER systems, even without learning, and learning can further improve the performance of the graph-based ranking approach. CONCLUSION: The utility of a named entity recognition (NER) system for geneId-finding may not be accurately predicted by its entity-level F1 performance, the most common performance measure. GeneId-ranking systems are best implemented by combining several NER systems. With appropriate combination methods, usefully accurate geneId-ranking systems can be constructed based on easily-available resources, without resorting to problem-specific, engineered components
Extracting causal relations on HIV drug resistance from literature
<p>Abstract</p> <p>Background</p> <p>In HIV treatment it is critical to have up-to-date resistance data of applicable drugs since HIV has a very high rate of mutation. These data are made available through scientific publications and must be extracted manually by experts in order to be used by virologists and medical doctors. Therefore there is an urgent need for a tool that partially automates this process and is able to retrieve relations between drugs and virus mutations from literature.</p> <p>Results</p> <p>In this work we present a novel method to extract and combine relationships between HIV drugs and mutations in viral genomes. Our extraction method is based on natural language processing (NLP) which produces grammatical relations and applies a set of rules to these relations. We applied our method to a relevant set of PubMed abstracts and obtained 2,434 extracted relations with an estimated performance of 84% for F-score. We then combined the extracted relations using logistic regression to generate resistance values for each <drug, mutation> pair. The results of this relation combination show more than 85% agreement with the Stanford HIVDB for the ten most frequently occurring mutations. The system is used in 5 hospitals from the Virolab project <url>http://www.virolab.org</url> to preselect the most relevant novel resistance data from literature and present those to virologists and medical doctors for further evaluation.</p> <p>Conclusions</p> <p>The proposed relation extraction and combination method has a good performance on extracting HIV drug resistance data. It can be used in large-scale relation extraction experiments. The developed methods can also be applied to extract other type of relations such as gene-protein, gene-disease, and disease-mutation.</p
OSIRISv1.2: A named entity recognition system for sequence variants of genes in biomedical literature
<p>Abstract</p> <p>Background</p> <p>Single Nucleotide Polymorphisms, among other type of sequence variants, constitute key elements in genetic epidemiology and pharmacogenomics. While sequence data about genetic variation is found at databases such as dbSNP, clues about the functional and phenotypic consequences of the variations are generally found in biomedical literature. The identification of the relevant documents and the extraction of the information from them are hampered by the large size of literature databases and the lack of widely accepted standard notation for biomedical entities. Thus, automatic systems for the identification of citations of allelic variants of genes in biomedical texts are required.</p> <p>Results</p> <p>Our group has previously reported the development of OSIRIS, a system aimed at the retrieval of literature about allelic variants of genes <url>http://ibi.imim.es/osirisform.html</url>. Here we describe the development of a new version of OSIRIS (OSIRISv1.2, <url>http://ibi.imim.es/OSIRISv1.2.html</url>) which incorporates a new entity recognition module and is built on top of a local mirror of the MEDLINE collection and HgenetInfoDB: a database that collects data on human gene sequence variations. The new entity recognition module is based on a pattern-based search algorithm for the identification of variation terms in the texts and their mapping to dbSNP identifiers. The performance of OSIRISv1.2 was evaluated on a manually annotated corpus, resulting in 99% precision, 82% recall, and an F-score of 0.89. As an example, the application of the system for collecting literature citations for the allelic variants of genes related to the diseases intracranial aneurysm and breast cancer is presented.</p> <p>Conclusion</p> <p>OSIRISv1.2 can be used to link literature references to dbSNP database entries with high accuracy, and therefore is suitable for collecting current knowledge on gene sequence variations and supporting the functional annotation of variation databases. The application of OSIRISv1.2 in combination with controlled vocabularies like MeSH provides a way to identify associations of biomedical interest, such as those that relate SNPs with diseases.</p
Snow Chemistry Across Antarctica
An updated compilation of published and new data of major-ion (Ca, Cl, K, Mg, Na, NO3, SO4) and methylsulfonate (MS) concentrations in snow from 520 Antarctic sites is provided by the national ITASE (International Trans-Antarctic Scientific Expedition) programmes of Australia, Brazil, China, Germany, Italy, Japan, Korea, New Zealand, Norway, the United Kingdom, the United States and the national Antarctic programme of Finland. The comparison shows that snow chemistry concentrations vary by up to four orders of magnitude across Antarctica and exhibit distinct geographical patterns. The Antarctic-wide comparison of glaciochemical records provides a unique opportunity to improve our understanding of the fundamental factors that ultimately control the chemistry of snow or ice samples. This paper aims to initiate data compilation and administration in order to provide a framework for facilitation of Antarctic-wide snow chemistry discussions across all ITASE nations and other contributing groups. The data are made available through the ITASE web page (http:// www2.umaine.edu/itase/content/syngroups/snowchem.html) and will be updated with new data as they are provided. In addition, recommendations for future research efforts are summarized
BioCreative III interactive task: an overview
The BioCreative challenge evaluation is a community-wide effort for evaluating text mining and information extraction systems applied to the biological domain. The biocurator community, as an active user of biomedical literature, provides a diverse and engaged end user group for text mining tools. Earlier BioCreative challenges involved many text mining teams in developing basic capabilities relevant to biological curation, but they did not address the issues of system usage, insertion into the workflow and adoption by curators. Thus in BioCreative III (BC-III), the InterActive Task (IAT) was introduced to address the utility and usability of text mining tools for real-life biocuration tasks. To support the aims of the IAT in BC-III, involvement of both developers and end users was solicited, and the development of a user interface to address the tasks interactively was requested
- …