387 research outputs found
How Do UX Practitioners Communicate AI as a Design Material? Artifacts, Conceptions, and Propositions
UX practitioners (UXPs) face novel challenges when working with and
communicating artificial intelligence (AI) as a design material. We explore how
UXPs communicate AI concepts when given hands-on experience training and
experimenting with AI models. To do so, we conducted a task-based design study
with 27 UXPs in which they prototyped and created a design presentation for a
AI-enabled interface while having access to a simple AI model training tool.
Through analyzing UXPs' design presentations and post-activity interviews, we
found that although UXPs struggled to clearly communicate some AI concepts,
tinkering with AI broadened common ground when communicating with technical
stakeholders. UXPs also identified key risks and benefits of AI in their
designs, and proposed concrete next steps for both UX and AI work. We conclude
with a sensitizing concept and recommendations for design and AI tools to
enhance multi-stakeholder communication and collaboration when crafting
human-centered AI experiences
Synthia's Melody: A Benchmark Framework for Unsupervised Domain Adaptation in Audio
Despite significant advancements in deep learning for vision and natural
language, unsupervised domain adaptation in audio remains relatively
unexplored. We, in part, attribute this to the lack of an appropriate benchmark
dataset. To address this gap, we present Synthia's melody, a novel audio data
generation framework capable of simulating an infinite variety of 4-second
melodies with user-specified confounding structures characterised by musical
keys, timbre, and loudness. Unlike existing datasets collected under
observational settings, Synthia's melody is free of unobserved biases, ensuring
the reproducibility and comparability of experiments. To showcase its utility,
we generate two types of distribution shifts-domain shift and sample selection
bias-and evaluate the performance of acoustic deep learning models under these
shifts. Our evaluations reveal that Synthia's melody provides a robust testbed
for examining the susceptibility of these models to varying levels of
distribution shift
Metastatic model of HPV+ oropharyngeal squamous cell carcinoma demonstrates heterogeneity in tumor metastasis
Human papillomavirus induced (HPV+) cancer incidence is rapidly rising, comprising 60–80% of oropharyngeal squamous cell carcinomas (OPSCCs); while rare, recurrent/metastatic disease accounts for nearly all related deaths. An in vivo pre-clinical model for these invasive cancers is necessary for testing new therapies. We characterize an immune competent recurrent/metastatic HPV+ murine model of OPSSC which consists of four lung metastatic (MLM) cell lines isolated from an animal with HPV+ OPSCC that failed cisplatin/radiation treatment. These individual metastatic clonal cell lines were tested to verify their origin (parental transgene expression and define their physiological properties: proliferation, metastatic potential, heterogeneity and sensitivity/resistance to cisplatin and radiation. All MLMs retain expression of parental HPV16 E6 and E7 and degrade P53 yet are heterogeneous from one another and from the parental cell line as defined by Illumina expression microarray. Consistent with this, reverse phase protein array defines differences in protein expression/activation between MLMs as well as the parental line. While in vitro growth rates of MLMs are slower than the parental line, in vivo growth of MLM clones is greatly enhanced. Moreover, in vivo resistance to standard therapies is dramatically increased in 3 of the 4 MLMs. Lymphatic and/or lung metastasis occurs 100% of the time in one MLM line. This recurrent/metastatic model of HPV+ OPSCC retains the characteristics evident in refractory human disease (heterogeneity, resistance to therapy, metastasis in lymph nodes/lungs) thus serving as an ideal translational system to test novel therapeutics. Moreover, this system may provide insights into the molecular mechanisms of metastasis
Ecology & computer audition: applications of audio technology to monitor organisms and environment
Among the 17 Sustainable Development Goals (SDGs) proposed within the 2030 Agenda and adopted by all the United Nations member states, the 13th SDG is a call for action to combat climate change. Moreover, SDGs 14 and 15 claim the protection and conservation of life below water and life on land, respectively. In this work, we provide a literature-founded overview of application areas, in which computer audition – a powerful but in this context so far hardly considered technology, combining audio signal processing and machine intelligence – is employed to monitor our ecosystem with the potential to identify ecologically critical processes or states. We distinguish between applications related to organisms, such as species richness analysis and plant health monitoring, and applications related to the environment, such as melting ice monitoring or wildfire detection. This work positions computer audition in relation to alternative approaches by discussing methodological strengths and limitations, as well as ethical aspects. We conclude with an urgent call to action to the research community for a greater involvement of audio intelligence methodology in future ecosystem monitoring approaches
A large-scale and PCR-referenced vocal audio dataset for COVID-19
The UK COVID-19 Vocal Audio Dataset is designed for the training and
evaluation of machine learning models that classify SARS-CoV-2 infection status
or associated respiratory symptoms using vocal audio. The UK Health Security
Agency recruited voluntary participants through the national Test and Trace
programme and the REACT-1 survey in England from March 2021 to March 2022,
during dominant transmission of the Alpha and Delta SARS-CoV-2 variants and
some Omicron variant sublineages. Audio recordings of volitional coughs,
exhalations, and speech were collected in the 'Speak up to help beat
coronavirus' digital survey alongside demographic, self-reported symptom and
respiratory condition data, and linked to SARS-CoV-2 test results. The UK
COVID-19 Vocal Audio Dataset represents the largest collection of SARS-CoV-2
PCR-referenced audio recordings to date. PCR results were linked to 70,794 of
72,999 participants and 24,155 of 25,776 positive cases. Respiratory symptoms
were reported by 45.62% of participants. This dataset has additional potential
uses for bioacoustics research, with 11.30% participants reporting asthma, and
27.20% with linked influenza PCR test results.Comment: 37 pages, 4 figure
- …