3,894 research outputs found

    How the other half lives: CRISPR-Cas's influence on bacteriophages

    Full text link
    CRISPR-Cas is a genetic adaptive immune system unique to prokaryotic cells used to combat phage and plasmid threats. The host cell adapts by incorporating DNA sequences from invading phages or plasmids into its CRISPR locus as spacers. These spacers are expressed as mobile surveillance RNAs that direct CRISPR-associated (Cas) proteins to protect against subsequent attack by the same phages or plasmids. The threat from mobile genetic elements inevitably shapes the CRISPR loci of archaea and bacteria, and simultaneously the CRISPR-Cas immune system drives evolution of these invaders. Here we highlight our recent work, as well as that of others, that seeks to understand phage mechanisms of CRISPR-Cas evasion and conditions for population coexistence of phages with CRISPR-protected prokaryotes.Comment: 24 pages, 8 figure

    METGAN: Generative Tumour Inpainting and Modality Synthesis in Light Sheet Microscopy

    Full text link
    Novel multimodal imaging methods are capable of generating extensive, super high resolution datasets for preclinical research. Yet, a massive lack of annotations prevents the broad use of deep learning to analyze such data. So far, existing generative models fail to mitigate this problem because of frequent labeling errors. In this paper, we introduce a novel generative method which leverages real anatomical information to generate realistic image-label pairs of tumours. We construct a dual-pathway generator, for the anatomical image and label, trained in a cycle-consistent setup, constrained by an independent, pretrained segmentor. The generated images yield significant quantitative improvement compared to existing methods. To validate the quality of synthesis, we train segmentation networks on a dataset augmented with the synthetic data, substantially improving the segmentation over baseline

    Structural basis for CRISPR RNA-guided DNA recognition by Cascade

    Get PDF
    The CRISPR (clustered regularly interspaced short palindromic repeats) immune system in prokaryotes uses small guide RNAs to neutralize invading viruses and plasmids. In Escherichia coli, immunity depends on a ribonucleoprotein complex called Cascade. Here we present the composition and low-resolution structure of Cascade and show how it recognizes double-stranded DNA (dsDNA) targets in a sequence-specific manner. Cascade is a 405-kDa complex comprising five functionally essential CRISPR-associated (Cas) proteins (CasA1B2C6D1E1) and a 61-nucleotide CRISPR RNA (crRNA) with 5′-hydroxyl and 2′,3′-cyclic phosphate termini. The crRNA guides Cascade to dsDNA target sequences by forming base pairs with the complementary DNA strand while displacing the noncomplementary strand to form an R-loop. Cascade recognizes target DNA without consuming ATP, which suggests that continuous invader DNA surveillance takes place without energy investment. The structure of Cascade shows an unusual seahorse shape that undergoes conformational changes when it binds target DNA.

    Intercomparison of spectroradiometers and Sun photometers for the determination of the aerosol optical depth during the VELETA-2002 field campaign

    Get PDF
    [ 1] In July 2002 the VELETA-2002 field campaign was held in Sierra Nevada ( Granada) in the south of Spain. The main objectives of this field campaign were the study of the influence of elevation and atmospheric aerosols on measured UV radiation. In the first stage of the field campaign, a common calibration and intercomparison between Licor-1800 spectroradiometers and Cimel-318 Sun photometers was performed in order to assess the quality of the measurements from the whole campaign. The intercomparison of the Licor spectroradiometers showed, for both direct and global irradiances, that when the comparisons were restricted to the visible part of the spectrum the deviations were within the instruments' nominal accuracies which allows us to rely on these instruments for measuring physical properties of aerosols at the different measurement stations. A simultaneous calibration on AOD data was performed for the Cimel-318 Sun photometers. When a common calibration and methodology was applied, the deviation was lowered to much less than 0.01 for AOD. At the same time an intercomparison has been made between the AOD values given by the spectroradiometers and the Sun photometers, with deviations obtained from 0.01 to 0.03 for the AOD in the visible range, depending on the channel. In the UVA range, the AOD uncertainty was estimated to be around 0.02 and 0.05 for Cimel and Licor respectively. In general the experimental differences were in agreement with this uncertainty estimation. In the UVB range the AOD measurements should not be used due to maximum instrumental uncertainties

    Integrated genomics and proteomics define huntingtin CAG length-dependent networks in mice.

    Get PDF
    To gain insight into how mutant huntingtin (mHtt) CAG repeat length modifies Huntington's disease (HD) pathogenesis, we profiled mRNA in over 600 brain and peripheral tissue samples from HD knock-in mice with increasing CAG repeat lengths. We found repeat length-dependent transcriptional signatures to be prominent in the striatum, less so in cortex, and minimal in the liver. Coexpression network analyses revealed 13 striatal and 5 cortical modules that correlated highly with CAG length and age, and that were preserved in HD models and sometimes in patients. Top striatal modules implicated mHtt CAG length and age in graded impairment in the expression of identity genes for striatal medium spiny neurons and in dysregulation of cyclic AMP signaling, cell death and protocadherin genes. We used proteomics to confirm 790 genes and 5 striatal modules with CAG length-dependent dysregulation at the protein level, and validated 22 striatal module genes as modifiers of mHtt toxicities in vivo

    SNPLims: a data management system for genome wide association studies

    Get PDF
    <p>Abstract</p> <p>Background</p> <p>Recent progresses in genotyping technologies allow the generation high-density genetic maps using hundreds of thousands of genetic markers for each DNA sample. The availability of this large amount of genotypic data facilitates the whole genome search for genetic basis of diseases.</p> <p>We need a suitable information management system to efficiently manage the data flow produced by whole genome genotyping and to make it available for further analyses.</p> <p>Results</p> <p>We have developed an information system mainly devoted to the storage and management of SNP genotype data produced by the Illumina platform from the raw outputs of genotyping into a relational database.</p> <p>The relational database can be accessed in order to import any existing data and export user-defined formats compatible with many different genetic analysis programs.</p> <p>After calculating family-based or case-control association study data, the results can be imported in SNPLims. One of the main features is to allow the user to rapidly identify and annotate statistically relevant polymorphisms from the large volume of data analyzed. Results can be easily visualized either graphically or creating ASCII comma separated format output files, which can be used as input to further analyses.</p> <p>Conclusions</p> <p>The proposed infrastructure allows to manage a relatively large amount of genotypes for each sample and an arbitrary number of samples and phenotypes. Moreover, it enables the users to control the quality of the data and to perform the most common screening analyses and identify genes that become “candidate” for the disease under consideration.</p

    blob loss: instance imbalance aware loss functions for semantic segmentation

    Full text link
    Deep convolutional neural networks have proven to be remarkably effective in semantic segmentation tasks. Most popular loss functions were introduced targeting improved volumetric scores, such as the Sorensen Dice coefficient. By design, DSC can tackle class imbalance; however, it does not recognize instance imbalance within a class. As a result, a large foreground instance can dominate minor instances and still produce a satisfactory Sorensen Dice coefficient. Nevertheless, missing out on instances will lead to poor detection performance. This represents a critical issue in applications such as disease progression monitoring. For example, it is imperative to locate and surveil small-scale lesions in the follow-up of multiple sclerosis patients. We propose a novel family of loss functions, nicknamed blob loss, primarily aimed at maximizing instance-level detection metrics, such as F1 score and sensitivity. Blob loss is designed for semantic segmentation problems in which the instances are the connected components within a class. We extensively evaluate a DSC-based blob loss in five complex 3D semantic segmentation tasks featuring pronounced instance heterogeneity in terms of texture and morphology. Compared to soft Dice loss, we achieve 5 percent improvement for MS lesions, 3 percent improvement for liver tumor, and an average 2 percent improvement for Microscopy segmentation tasks considering F1 score.Comment: 23 pages, 7 figures // corrected one mistake where it said beta instead of alpha in the tex

    Inferring learning from big data:The importance of a transdisciplinary and multidimensional approach

    Get PDF
    The use of big data in higher education has evolved rapidly with a focus on the practical application of new tools and methods for supporting learning. In this paper, we depart from the core emphasis on application and delve into a mostly neglected aspect of the big data conversation in higher education. Drawing on developments in cognate disciplines, we analyse the inherent difficulties in inferring the complex phenomenon that is learning from big datasets. This forms the basis of a discussion about the possibilities for systematic collaboration across different paradigms and disciplinary backgrounds in interpreting big data for enhancing learning. The aim of this paper is to provide the foundation for a research agenda, where differing conceptualisations of learning become a strength in interpreting patterns in big datasets, rather than a point of contention
    • …
    corecore