280,822 research outputs found

    Valuing Biodiversity from an Economic Perspective: AUnified Economic, Ecological and Genetic Approach

    Get PDF
    We develop a conceptual framework for valuing biodiversity from an economic perspective. We consider biodiversity important because of a number of characteristics or services that it provides or enhances. We argue for a dynamic economic welfare measure of biodiversity that complements the existing literature on benefit-cost approaches and genetic distance/phylogenic tree approaches, which to date have been more static. Using a unified model of optimal economic management of an ecosystem under ecological and genetic constraints, we identify gains realized by management policies leading to a more diverse system, using the Bellman state valuation function of the problem. We show that a more diverse system could attain a higher value even though the genetic distance of the species in the more diverse system could be almost zero. We relate this endogenous measure of the biodiversity value to ecologically/biologically oriented biodiversity metrics (species richness, Shannon or Simpson indices).

    Review of high-contrast imaging systems for current and future ground- and space-based telescopes I. Coronagraph design methods and optical performance metrics

    Full text link
    The Optimal Optical Coronagraph (OOC) Workshop at the Lorentz Center in September 2017 in Leiden, the Netherlands gathered a diverse group of 25 researchers working on exoplanet instrumentation to stimulate the emergence and sharing of new ideas. In this first installment of a series of three papers summarizing the outcomes of the OOC workshop, we present an overview of design methods and optical performance metrics developed for coronagraph instruments. The design and optimization of coronagraphs for future telescopes has progressed rapidly over the past several years in the context of space mission studies for Exo-C, WFIRST, HabEx, and LUVOIR as well as ground-based telescopes. Design tools have been developed at several institutions to optimize a variety of coronagraph mask types. We aim to give a broad overview of the approaches used, examples of their utility, and provide the optimization tools to the community. Though it is clear that the basic function of coronagraphs is to suppress starlight while maintaining light from off-axis sources, our community lacks a general set of standard performance metrics that apply to both detecting and characterizing exoplanets. The attendees of the OOC workshop agreed that it would benefit our community to clearly define quantities for comparing the performance of coronagraph designs and systems. Therefore, we also present a set of metrics that may be applied to theoretical designs, testbeds, and deployed instruments. We show how these quantities may be used to easily relate the basic properties of the optical instrument to the detection significance of the given point source in the presence of realistic noise.Comment: To appear in Proceedings of the SPIE, vol. 1069

    Fr\'echet ChemNet Distance: A metric for generative models for molecules in drug discovery

    Full text link
    The new wave of successful generative models in machine learning has increased the interest in deep learning driven de novo drug design. However, assessing the performance of such generative models is notoriously difficult. Metrics that are typically used to assess the performance of such generative models are the percentage of chemically valid molecules or the similarity to real molecules in terms of particular descriptors, such as the partition coefficient (logP) or druglikeness. However, method comparison is difficult because of the inconsistent use of evaluation metrics, the necessity for multiple metrics, and the fact that some of these measures can easily be tricked by simple rule-based systems. We propose a novel distance measure between two sets of molecules, called Fr\'echet ChemNet distance (FCD), that can be used as an evaluation metric for generative models. The FCD is similar to a recently established performance metric for comparing image generation methods, the Fr\'echet Inception Distance (FID). Whereas the FID uses one of the hidden layers of InceptionNet, the FCD utilizes the penultimate layer of a deep neural network called ChemNet, which was trained to predict drug activities. Thus, the FCD metric takes into account chemically and biologically relevant information about molecules, and also measures the diversity of the set via the distribution of generated molecules. The FCD's advantage over previous metrics is that it can detect if generated molecules are a) diverse and have similar b) chemical and c) biological properties as real molecules. We further provide an easy-to-use implementation that only requires the SMILES representation of the generated molecules as input to calculate the FCD. Implementations are available at: https://www.github.com/bioinf-jku/FCDComment: Implementations are available at: https://www.github.com/bioinf-jku/FC

    Text Style Transfer Evaluation Using Large Language Models

    Full text link
    Evaluating Text Style Transfer (TST) is a complex task due to its multifaceted nature. The quality of the generated text is measured based on challenging factors, such as style transfer accuracy, content preservation, and overall fluency. While human evaluation is considered to be the gold standard in TST assessment, it is costly and often hard to reproduce. Therefore, automated metrics are prevalent in these domains. Nevertheless, it remains unclear whether these automated metrics correlate with human evaluations. Recent strides in Large Language Models (LLMs) have showcased their capacity to match and even exceed average human performance across diverse, unseen tasks. This suggests that LLMs could be a feasible alternative to human evaluation and other automated metrics in TST evaluation. We compare the results of different LLMs in TST using multiple input prompts. Our findings highlight a strong correlation between (even zero-shot) prompting and human evaluation, showing that LLMs often outperform traditional automated metrics. Furthermore, we introduce the concept of prompt ensembling, demonstrating its ability to enhance the robustness of TST evaluation. This research contributes to the ongoing evaluation of LLMs in diverse tasks, offering insights into successful outcomes and areas of limitation
    corecore