127 research outputs found

    Bayesian methods outperform parsimony but at the expense of precision in the estimation of phylogeny from discrete morphological data

    Get PDF
    Different analytical methods can yield competing interpretations of evolutionary history and, currently, there is no definitive method for phylogenetic reconstruction using morphological data. Parsimony has been the primary method for analysing morphological data, but there has been a resurgence of interest in the likelihood-based Mk-model. Here, we test the performance of the Bayesian implementation of the Mk-model relative to both equal and implied-weight implementations of parsimony. Using simulated morphological data, we demonstrate that the Mk-model outperforms equal-weights parsimony in terms of topological accuracy, and implied-weights performs the most poorly. However, the Mk-model produces phylogenies that have less resolution than parsimony methods. This difference in the accuracy and precision of parsimony and Bayesian approaches to topology estimation needs to be considered when selecting a method for phylogeny reconstruction

    Uncertain-tree:Discriminating among competing approaches to the phylogenetic analysis of phenotype data

    Get PDF
    Morphological data provide the only means of classifying the majority of life's history, but the choice between competing phylogenetic methods for the analysis of morphology is unclear. Traditionally, parsimony methods have been favoured but recent studies have shown that these approaches are less accurate than the Bayesian implementation of the Mk model. Here we expand on these findings in several ways: we assess the impact of tree shape and maximum-likelihood estimation using the Mk model, as well as analysing data composed of both binary and multistate characters. We find that all methods struggle to correctly resolve deep clades within asymmetric trees, and when analysing small character matrices. The Bayesian Mk model is the most accurate method for estimating topology, but with lower resolution than other methods. Equal weights parsimony is more accurate than implied weights parsimony, and maximum-likelihood estimation using the Mk model is the least accurate method. We conclude that the Bayesian implementation of the Mk model should be the default method for phylogenetic estimation from phenotype datasets, and we explore the implications of our simulations in reanalysing several empirical morphological character matrices. A consequence of our finding is that high levels of resolution or the ability to classify species or groups with much confidence should not be expected when using small datasets. It is now necessary to depart from the traditional parsimony paradigms of constructing character matrices, towards datasets constructed explicitly for Bayesian methods.This research was funded by NERC (NE/L501554/1 to J.E.O.R. and L.A.P.; NE/K500823/1 to M.N.P.; NE/L002434/1 to J.F.; NE/N003438/1 to P.C.J.D.), BBSRC (BB/N000919/1 to P.C.J.D.), the University of Bristol (STaR scholarship to A.R.T.), Royal Society Wolfson Research Merit Award (P.C.J.D.) and the John Templeton Foundation (43915 to D.P. and L.H.).N

    Views on social media and its linkage to longitudinal data from two generations of a UK cohort study

    Get PDF
    Background: Cohort studies gather huge volumes of information about a range of phenotypes but new sources of information such as social media data are yet to be integrated. Participant’s long-term engagement with cohort studies, as well as the potential for their social media data to be linked to other longitudinal data, could provide novel advances but may also give participants a unique perspective on the acceptability of this growing research area. Methods: Two focus groups explored participant views towards the acceptability and best practice for the collection of social media data for research purposes. Participants were drawn from the Avon Longitudinal Study of Parents and Children cohort; individuals from the index cohort of young people (N=9) and from the parent generation (N=5) took part in two separate 90-minute focus groups. The discussions were audio recorded and subjected to qualitative analysis. Results: Participants were generally supportive of the collection of social media data to facilitate health and social research. They felt that their trust in the cohort study would encourage them to do so. Concern was expressed about the collection of data from friends or connections who had not consented. In terms of best practice for collecting the data, participants generally preferred the use of anonymous data derived from social media to be shared with researchers. Conclusion: Cohort studies have trusting relationships with their participants; for this relationship to extend to linking their social media data with longitudinal information, procedural safeguards are needed. Participants understand the goals and potential of research integrating social media data into cohort studies, but further research is required on the acquisition of their friend’s data. The views gathered from participants provide important guidance for future work seeking to integrate social media in cohort studies

    Participant acceptability of digital footprint data collection strategies:an exemplar approach to participant engagement and involvement in the ALSPAC birth cohort study

    Get PDF
    INTRODUCTION: Digital footprint records – the tracks and traces amassed by individuals as a result of their interactions with the internet, digital devices and services – can provide ecologically valid data on individual behaviours. These could enhance longitudinal population study databanks; but few UK longitudinal studies are attempting this. When using novel sources of data, study managers must engage with participants in order to develop ethical data processing frameworks that facilitate data sharing whilst safeguarding participant interests. OBJECTIVES: This paper aims to summarise the participant involvement approach used by the ALSPAC birth cohort study to inform the development of a framework for using linked participant digital footprint data, and provide an exemplar for other data linkage infrastructures. METHODS: The paper synthesises five qualitative forms of inquiry. Thematic analysis was used to code transcripts for common themes in relation to conditions associated with the acceptability of sharing digital footprint data for longitudinal research. RESULTS: We identified six themes: participant understanding; sensitivity of location data; concerns for third parties; clarity on data granularity; mechanisms of data sharing and consent; and trustworthiness of the organisation. For cohort members to consider the sharing of digital footprint data acceptable, they require information about the value, validity and risks; control over sharing elements of the data they consider sensitive; appropriate mechanisms to authorise or object to their records being used; and trust in the organisation. CONCLUSION: Realising the potential for using digital footprint records within longitudinal research will be subject to ensuring that this use of personal data is acceptable; and that rigorously controlled population data science benefiting the public good is distinguishable from the misuse and lack of personal control of similar data within other settings. Participant co-development informs the ethical-governance framework for these novel linkages in a manner which is acceptable and does not undermine the role of the trusted data custodian

    Epicosm -a framework for linking online social media in epidemiological cohorts

    Get PDF
    Motivation Social media represent an unrivalled opportunity for epidemiological cohorts to collect large amounts of high-resolution time course data on mental health. Equally, the high-quality data held by epidemiological cohorts could greatly benefit social media research as a source of ground truth for validating digital phenotyping algorithms. However, there is currently a lack of software for doing this in a secure and acceptable manner. We worked with cohort leaders and participants to co-design an open-source, robust and expandable software framework for gathering social media data in epidemiological cohorts. Implementation Epicosm is implemented as a Python framework that is straightforward to deploy and run inside a cohort’s data safe haven. General features The software regularly gathers Tweets from a list of accounts and stores them in a database for linking to existing cohort data. Availability This open-source software is freely available at [https://dynamicgenetics.github.io/Epicosm/]

    Soft-Bodied Fossils Are Not Simply Rotten Carcasses – Toward a Holistic Understanding of Exceptional Fossil Preservation:Exceptional Fossil Preservation Is Complex and Involves the Interplay of Numerous Biological and Geological Processes

    Get PDF
    Exceptionally preserved fossils are the product of complex interplays of biological and geological processes including burial, autolysis and microbial decay, authigenic mineralization, diagenesis, metamorphism, and finally weathering and exhumation. Determining which tissues are preserved and how biases affect their preservation pathways is important for interpreting fossils in phylogenetic, ecological, and evolutionary frameworks. Although laboratory decay experiments reveal important aspects of fossilization, applying the results directly to the interpretation of exceptionally preserved fossils may overlook the impact of other key processes that remove or preserve morphological information. Investigations of fossils preserving non-biomineralized tissues suggest that certain structures that are decay resistant (e.g., the notochord) are rarely preserved (even where carbonaceous components survive), and decay-prone structures (e.g., nervous systems) can fossilize, albeit rarely. As we review here, decay resistance is an imperfect indicator of fossilization potential, and a suite of biological and geological processes account for the features preserved in exceptional fossils.</p
    • …
    corecore