2,578 research outputs found

    PESSCARA: An Example Infrastructure for Big Data Research

    Get PDF
    Big data requires a flexible system for data management and curation which has to be intuitive, and it should also be able to execute non-linear analysis pipelines suitable to handle with the nature of big data. This is certainly true for medical images where the amount of data grows exponentially every year and the nature of images rapidly changes with technological advances and rapid genomic advances. In this chapter, we describe a system that provides flexible management for medical images plus a wide array of associated metadata, including clinical data, genomic data, and clinical trial information. The system consists of open-source Content Management System (CMS) that has a highly configurable workflow; has a single interface that can store, manage, enable curation, and retrieve imaging-based studies; and can handle the requirement for data auditing and project management. Furthermore, the system can be extended to interact with all the modern big data analysis technologies

    A proposal for a coordinated effort for the determination of brainwide neuroanatomical connectivity in model organisms at a mesoscopic scale

    Get PDF
    In this era of complete genomes, our knowledge of neuroanatomical circuitry remains surprisingly sparse. Such knowledge is however critical both for basic and clinical research into brain function. Here we advocate for a concerted effort to fill this gap, through systematic, experimental mapping of neural circuits at a mesoscopic scale of resolution suitable for comprehensive, brain-wide coverage, using injections of tracers or viral vectors. We detail the scientific and medical rationale and briefly review existing knowledge and experimental techniques. We define a set of desiderata, including brain-wide coverage; validated and extensible experimental techniques suitable for standardization and automation; centralized, open access data repository; compatibility with existing resources, and tractability with current informatics technology. We discuss a hypothetical but tractable plan for mouse, additional efforts for the macaque, and technique development for human. We estimate that the mouse connectivity project could be completed within five years with a comparatively modest budget.Comment: 41 page

    The Open-Access European Prevention of Alzheimer's Dementia (EPAD) MRI dataset and processing workflow

    Get PDF
    The European Prevention of Alzheimer Dementia (EPAD) is a multi-center study that aims to characterize the preclinical and prodromal stages of Alzheimer's Disease. The EPAD imaging dataset includes core (3D T1w, 3D FLAIR) and advanced (ASL, diffusion MRI, and resting-state fMRI) MRI sequences. Here, we give an overview of the semi-automatic multimodal and multisite pipeline that we developed to curate, preprocess, quality control (QC), and compute image-derived phenotypes (IDPs) from the EPAD MRI dataset. This pipeline harmonizes DICOM data structure across sites and performs standardized MRI preprocessing steps. A semi-automated MRI QC procedure was implemented to visualize and flag MRI images next to site-specific distributions of QC features - i.e. metrics that represent image quality. The value of each of these QC features was evaluated through comparison with visual assessment and step-wise parameter selection based on logistic regression. IDPs were computed from 5 different MRI modalities and their sanity and potential clinical relevance were ascertained by assessing their relationship with biological markers of aging and dementia. The EPAD v1500.0 data release encompassed core structural scans from 1356 participants 842 fMRI, 831 dMRI, and 858 ASL scans. From 1356 3D T1w images, we identified 17 images with poor quality and 61 with moderate quality. Five QC features - Signal to Noise Ratio (SNR), Contrast to Noise Ratio (CNR), Coefficient of Joint Variation (CJV), Foreground-Background energy Ratio (FBER), and Image Quality Rate (IQR) - were selected as the most informative on image quality by comparison with visual assessment. The multimodal IDPs showed greater impairment in associations with age and dementia biomarkers, demonstrating the potential of the dataset for future clinical analyses

    Data preparation for artificial intelligence in medical imaging: A comprehensive guide to open-access platforms and tools

    Get PDF
    The vast amount of data produced by today's medical imaging systems has led medical professionals to turn to novel technologies in order to efficiently handle their data and exploit the rich information present in them. In this context, artificial intelligence (AI) is emerging as one of the most prominent solutions, promising to revolutionise every day clinical practice and medical research. The pillar supporting the development of reliable and robust AI algorithms is the appropriate preparation of the medical images to be used by the AI-driven solutions. Here, we provide a comprehensive guide for the necessary steps to prepare medical images prior to developing or applying AI algorithms. The main steps involved in a typical medical image preparation pipeline include: (i) image acquisition at clinical sites, (ii) image de-identification to remove personal information and protect patient privacy, (iii) data curation to control for image and associated information quality, (iv) image storage, and (v) image annotation. There exists a plethora of open access tools to perform each of the aforementioned tasks and are hereby reviewed. Furthermore, we detail medical image repositories covering different organs and diseases. Such repositories are constantly increasing and enriched with the advent of big data. Lastly, we offer directions for future work in this rapidly evolving field

    Improving fairness in machine learning systems: What do industry practitioners need?

    Full text link
    The potential for machine learning (ML) systems to amplify social inequities and unfairness is receiving increasing popular and academic attention. A surge of recent work has focused on the development of algorithmic tools to assess and mitigate such unfairness. If these tools are to have a positive impact on industry practice, however, it is crucial that their design be informed by an understanding of real-world needs. Through 35 semi-structured interviews and an anonymous survey of 267 ML practitioners, we conduct the first systematic investigation of commercial product teams' challenges and needs for support in developing fairer ML systems. We identify areas of alignment and disconnect between the challenges faced by industry practitioners and solutions proposed in the fair ML research literature. Based on these findings, we highlight directions for future ML and HCI research that will better address industry practitioners' needs.Comment: To appear in the 2019 ACM CHI Conference on Human Factors in Computing Systems (CHI 2019
    • …
    corecore