1,627 research outputs found

    Understanding Surprise: Can Less Likely Events Be Less Surprising?

    Get PDF
    Surprise is often thought of as an experience that is elicited following an unexpected event. However, it may also be the case that surprise stems from an event that is simply difficult to explain. In this paper, we investigate the latter view. Specifically, we question why the provision of an enabling factor can mitigate perceived surprise for an unexpected event despite lowering the overall probability of that event. One possibility is that surprise occurs when a person cannot rationalise an outcome event in the context of the scenario representation. A second possibility is that people can generate plausible explanations for unexpected events but that surprise is experienced when those explanations are uncertain. We explored these hypotheses in an experiment where a first group of participants rated surprise for a number of scenario outcomes and a second group rated surprise after generating a plausible explanation for those outcomes. Finally, a third group of participants rated surprise for the both the original outcomes and the reasons generated for those outcomes by the second group. Our results suggest that people can come up with plausible explanations for unexpected events but that surprise results when these explanations are uncertain

    Understanding Surprise: Can Less Likely Events Be Less Surprising?

    Get PDF
    Surprise is often thought of as an experience that is elicited following an unexpected event. However, it may also be the case that surprise stems from an event that is simply difficult to explain. In this paper, we investigate the latter view. Specifically, we question why the provision of an enabling factor can mitigate perceived surprise for an unexpected event despite lowering the overall probability of that event. One possibility is that surprise occurs when a person cannot rationalise an outcome event in the context of the scenario representation. A second possibility is that people can generate plausible explanations for unexpected events but that surprise is experienced when those explanations are uncertain. We explored these hypotheses in an experiment where a first group of participants rated surprise for a number of scenario outcomes and a second group rated surprise after generating a plausible explanation for those outcomes. Finally, a third group of participants rated surprise for the both the original outcomes and the reasons generated for those outcomes by the second group. Our results suggest that people can come up with plausible explanations for unexpected events but that surprise results when these explanations are uncertain

    MultiPhyl: a high-throughput phylogenomics webserver using distributed computing

    Get PDF
    With the number of fully sequenced genomes increasing steadily, there is greater interest in performing large-scale phylogenomic analyses from large numbers of individual gene families. Maximum likelihood (ML) has been shown repeatedly to be one of the most accurate methods for phylogenetic construction. Recently, there have been a number of algorithmic improvements in maximum-likelihood-based tree search methods. However, it can still take a long time to analyse the evolutionary history of many gene families using a single computer. Distributed computing refers to a method of combining the computing power of multiple computers in order to perform some larger overall calculation. In this article, we present the first high-throughput implementation of a distributed phylogenetics platform, MultiPhyl, capable of using the idle computational resources of many heterogeneous non-dedicated machines to form a phylogenetics supercomputer. MultiPhyl allows a user to upload hundreds or thousands of amino acid or nucleotide alignments simultaneously and perform computationally intensive tasks such as model selection, tree searching and bootstrapping of each of the alignments using many desktop machines. The program implements a set of 88 amino acid models and 56 nucleotide maximum likelihood models and a variety of statistical methods for choosing between alternative models. A MultiPhyl webserver is available for public use at: http://www.cs.nuim.ie/distributed/multiphyl.php

    Assessment of methods for amino acid matrix selection and their use on empirical data shows that ad hoc assumptions for choice of matrix are not justified

    Get PDF
    BACKGROUND: In recent years, model based approaches such as maximum likelihood have become the methods of choice for constructing phylogenies. A number of authors have shown the importance of using adequate substitution models in order to produce accurate phylogenies. In the past, many empirical models of amino acid substitution have been derived using a variety of different methods and protein datasets. These matrices are normally used as surrogates, rather than deriving the maximum likelihood model from the dataset being examined. With few exceptions, selection between alternative matrices has been carried out in an ad hoc manner. RESULTS: We start by highlighting the potential dangers of arbitrarily choosing protein models by demonstrating an empirical example where a single alignment can produce two topologically different and strongly supported phylogenies using two different arbitrarily-chosen amino acid substitution models. We demonstrate that in simple simulations, statistical methods of model selection are indeed robust and likely to be useful for protein model selection. We have investigated patterns of amino acid substitution among homologous sequences from the three Domains of life and our results show that no single amino acid matrix is optimal for any of the datasets. Perhaps most interestingly, we demonstrate that for two large datasets derived from the proteobacteria and archaea, one of the most favored models in both datasets is a model that was originally derived from retroviral Pol proteins. CONCLUSION: This demonstrates that choosing protein models based on their source or method of construction may not be appropriate

    Identification of structural variation in mouse genomes.

    Get PDF
    Structural variation is variation in structure of DNA regions affecting DNA sequence length and/or orientation. It generally includes deletions, insertions, copy-number gains, inversions, and transposable elements. Traditionally, the identification of structural variation in genomes has been challenging. However, with the recent advances in high-throughput DNA sequencing and paired-end mapping (PEM) methods, the ability to identify structural variation and their respective association to human diseases has improved considerably. In this review, we describe our current knowledge of structural variation in the mouse, one of the prime model systems for studying human diseases and mammalian biology. We further present the evolutionary implications of structural variation on transposable elements. We conclude with future directions on the study of structural variation in mouse genomes that will increase our understanding of molecular architecture and functional consequences of structural variation

    Genome Report: Whole Genome Sequence of Two Wild-Derived Mus musculus domesticus Inbred Strains, LEWES/EiJ and ZALENDE/EiJ, with Different Diploid Numbers

    Get PDF
    Wild-derived mouse inbred strains are becoming increasingly popular for complex traits analysis, evolutionary studies, and systems genetics. Here, we report the whole-genome sequencing of two wild-derived mouse inbred strains, LEWES/EiJ and ZALENDE/EiJ, of Mus musculus domesticus origin. These two inbred strains were selected based on their geographic origin, karyotype, and use in ongoing research. We generated 14× and 18× coverage sequence, respectively, and discovered over 1.1 million novel variants, most of which are private to one of these strains. This report expands the number of wild-derived inbred genomes in the Mus genus from six to eight. The sequence variation can be accessed via an online query tool; variant calls (VCF format) and alignments (BAM format) are available for download from a dedicated ftp site. Finally, the sequencing data have also been stored in a lossless, compressed, and indexed format using the multi-string Burrows-Wheeler transform. All data can be used without restriction

    Using reference-free compressed data structures to analyze sequencing reads from thousands of human genomes.

    Get PDF
    We are rapidly approaching the point where we have sequenced millions of human genomes. There is a pressing need for new data structures to store raw sequencing data and efficient algorithms for population scale analysis. Current reference-based data formats do not fully exploit the redundancy in population sequencing nor take advantage of shared genetic variation. In recent years, the Burrows-Wheeler transform (BWT) and FM-index have been widely employed as a full-text searchable index for read alignment and de novo assembly. We introduce the concept of a population BWT and use it to store and index the sequencing reads of 2705 samples from the 1000 Genomes Project. A key feature is that, as more genomes are added, identical read sequences are increasingly observed, and compression becomes more efficient. We assess the support in the 1000 Genomes read data for every base position of two human reference assembly versions, identifying that 3.2 Mbp with population support was lost in the transition from GRCh37 with 13.7 Mbp added to GRCh38. We show that the vast majority of variant alleles can be uniquely described by overlapping 31-mers and show how rapid and accurate SNP and indel genotyping can be carried out across the genomes in the population BWT. We use the population BWT to carry out nonreference queries to search for the presence of all known viral genomes and discover human T-lymphotropic virus 1 integrations in six samples in a recognized epidemiological distribution

    A dinuclear ruthenium(II) phototherapeutic that targets duplex and quadruplex DNA

    Get PDF
    With the aim of developing a sensitizer for photodynamic therapy, a previously reported luminescent dinuclear complex that functions as a DNA probe in live cells was modified to produce a new isostructural derivative containing RuII(TAP)2 fragments (TAP = 1,4,5,8- tetraazaphenanthrene). The structure of the new complex has been confirmed by a variety of techniques including single crystal X-ray analysis. Unlike its parent, the new complex displays RuL-based 3MLCT emission in both MeCN and water. Results from electrochemical studies and emission quenching experiments involving guanosine monophosphate are consistent with an excited state located on a TAP moiety. This hypothesis is further supported by detailed DFT calculations, which take into account solvent effects on excited state dynamics. Cell-free steady-state and time-resolved optical studies on the interaction of the new complex with duplex and quadruplex DNA show that the complex binds with high affinity to both structures and indicate that its photoexcited state is also quenched by DNA, a process that is accompanied by the generation of the guanine radical cation sites as photo-oxidization products. Like the parent complex, this new compound is taken up by live cells where it primarily localizes within the nucleus and displays low cytotoxicity in the absence of light. However, in complete contrast to [{RuII(phen)2}2(tpphz)]4+, the new complex is therapeutically activated by light to become highly phototoxic toward malignant human melanoma cell line showing that it is a promising lead for the treatment of this recalcitrant cancer.EPSRC grant EP/M015572/1 Unviersity of Sheffield/EPSRC Doctoral Fellowship Prize EPSRC Capital Equipment Award ERASMUS
    corecore