383 research outputs found

    Synthia's Melody: A Benchmark Framework for Unsupervised Domain Adaptation in Audio

    Full text link
    Despite significant advancements in deep learning for vision and natural language, unsupervised domain adaptation in audio remains relatively unexplored. We, in part, attribute this to the lack of an appropriate benchmark dataset. To address this gap, we present Synthia's melody, a novel audio data generation framework capable of simulating an infinite variety of 4-second melodies with user-specified confounding structures characterised by musical keys, timbre, and loudness. Unlike existing datasets collected under observational settings, Synthia's melody is free of unobserved biases, ensuring the reproducibility and comparability of experiments. To showcase its utility, we generate two types of distribution shifts-domain shift and sample selection bias-and evaluate the performance of acoustic deep learning models under these shifts. Our evaluations reveal that Synthia's melody provides a robust testbed for examining the susceptibility of these models to varying levels of distribution shift

    Grand Military Waltz

    Get PDF

    How Do UX Practitioners Communicate AI as a Design Material? Artifacts, Conceptions, and Propositions

    Full text link
    UX practitioners (UXPs) face novel challenges when working with and communicating artificial intelligence (AI) as a design material. We explore how UXPs communicate AI concepts when given hands-on experience training and experimenting with AI models. To do so, we conducted a task-based design study with 27 UXPs in which they prototyped and created a design presentation for a AI-enabled interface while having access to a simple AI model training tool. Through analyzing UXPs' design presentations and post-activity interviews, we found that although UXPs struggled to clearly communicate some AI concepts, tinkering with AI broadened common ground when communicating with technical stakeholders. UXPs also identified key risks and benefits of AI in their designs, and proposed concrete next steps for both UX and AI work. We conclude with a sensitizing concept and recommendations for design and AI tools to enhance multi-stakeholder communication and collaboration when crafting human-centered AI experiences

    Metastatic model of HPV+ oropharyngeal squamous cell carcinoma demonstrates heterogeneity in tumor metastasis

    Get PDF
    Human papillomavirus induced (HPV+) cancer incidence is rapidly rising, comprising 60–80% of oropharyngeal squamous cell carcinomas (OPSCCs); while rare, recurrent/metastatic disease accounts for nearly all related deaths. An in vivo pre-clinical model for these invasive cancers is necessary for testing new therapies. We characterize an immune competent recurrent/metastatic HPV+ murine model of OPSSC which consists of four lung metastatic (MLM) cell lines isolated from an animal with HPV+ OPSCC that failed cisplatin/radiation treatment. These individual metastatic clonal cell lines were tested to verify their origin (parental transgene expression and define their physiological properties: proliferation, metastatic potential, heterogeneity and sensitivity/resistance to cisplatin and radiation. All MLMs retain expression of parental HPV16 E6 and E7 and degrade P53 yet are heterogeneous from one another and from the parental cell line as defined by Illumina expression microarray. Consistent with this, reverse phase protein array defines differences in protein expression/activation between MLMs as well as the parental line. While in vitro growth rates of MLMs are slower than the parental line, in vivo growth of MLM clones is greatly enhanced. Moreover, in vivo resistance to standard therapies is dramatically increased in 3 of the 4 MLMs. Lymphatic and/or lung metastasis occurs 100% of the time in one MLM line. This recurrent/metastatic model of HPV+ OPSCC retains the characteristics evident in refractory human disease (heterogeneity, resistance to therapy, metastasis in lymph nodes/lungs) thus serving as an ideal translational system to test novel therapeutics. Moreover, this system may provide insights into the molecular mechanisms of metastasis

    Detecting COVID-19 from breathing and coughing sounds using deep neural networks

    Get PDF

    Ecology & computer audition: applications of audio technology to monitor organisms and environment

    Get PDF
    Among the 17 Sustainable Development Goals (SDGs) proposed within the 2030 Agenda and adopted by all the United Nations member states, the 13th SDG is a call for action to combat climate change. Moreover, SDGs 14 and 15 claim the protection and conservation of life below water and life on land, respectively. In this work, we provide a literature-founded overview of application areas, in which computer audition – a powerful but in this context so far hardly considered technology, combining audio signal processing and machine intelligence – is employed to monitor our ecosystem with the potential to identify ecologically critical processes or states. We distinguish between applications related to organisms, such as species richness analysis and plant health monitoring, and applications related to the environment, such as melting ice monitoring or wildfire detection. This work positions computer audition in relation to alternative approaches by discussing methodological strengths and limitations, as well as ethical aspects. We conclude with an urgent call to action to the research community for a greater involvement of audio intelligence methodology in future ecosystem monitoring approaches

    A large-scale and PCR-referenced vocal audio dataset for COVID-19

    Get PDF
    The UK COVID-19 Vocal Audio Dataset is designed for the training and evaluation of machine learning models that classify SARS-CoV-2 infection status or associated respiratory symptoms using vocal audio. The UK Health Security Agency recruited voluntary participants through the national Test and Trace programme and the REACT-1 survey in England from March 2021 to March 2022, during dominant transmission of the Alpha and Delta SARS-CoV-2 variants and some Omicron variant sublineages. Audio recordings of volitional coughs, exhalations, and speech were collected in the 'Speak up to help beat coronavirus' digital survey alongside demographic, self-reported symptom and respiratory condition data, and linked to SARS-CoV-2 test results. The UK COVID-19 Vocal Audio Dataset represents the largest collection of SARS-CoV-2 PCR-referenced audio recordings to date. PCR results were linked to 70,794 of 72,999 participants and 24,155 of 25,776 positive cases. Respiratory symptoms were reported by 45.62% of participants. This dataset has additional potential uses for bioacoustics research, with 11.30% participants reporting asthma, and 27.20% with linked influenza PCR test results.Comment: 37 pages, 4 figure
    corecore