4,859 research outputs found
Recommended from our members
The role of HG in the analysis of temporal iteration and interaural correlation
Edge-centric Optimization of Multi-modal ML-driven eHealth Applications
Smart eHealth applications deliver personalized and preventive digital
healthcare services to clients through remote sensing, continuous monitoring,
and data analytics. Smart eHealth applications sense input data from multiple
modalities, transmit the data to edge and/or cloud nodes, and process the data
with compute intensive machine learning (ML) algorithms. Run-time variations
with continuous stream of noisy input data, unreliable network connection,
computational requirements of ML algorithms, and choice of compute placement
among sensor-edge-cloud layers affect the efficiency of ML-driven eHealth
applications. In this chapter, we present edge-centric techniques for optimized
compute placement, exploration of accuracy-performance trade-offs, and
cross-layered sense-compute co-optimization for ML-driven eHealth applications.
We demonstrate the practical use cases of smart eHealth applications in
everyday settings, through a sensor-edge-cloud framework for an objective pain
assessment case study
Exparel Compared to Standard Bupivacaine for Postoperative Analgesia Following Lumbar Spine Fusion
Inadequate postoperative pain management following spinal surgery contributes to delayed mobilization and chronic pain. The current standard of care following spinal surgery consists of opiates and anesthetics to target multiple pain pathways. However, opiates are limited by their significant adverse effects and anesthetics are limited by their short duration of delivery. One approach that may overcome these limitations is liposomal or lipid-encapsulated drug formulations, which have been shown to extend the duration of drug delivery to target tissues. In this study, we will compare adjunctive opioid consumption in a randomized controlled trial of patients undergoing elective posterior lumbar spinal surgery that receive liposomal and conventional anesthetic versus conventional anesthetic. We hypothesize that liposomal anesthetic will improve pain management by reducing the total adjunctive opioid consumption required. This study will address a key limitation of conventional anesthetics and may provide evidence for the utility of liposomal anesthetic in postoperative pain management
A Comprehensive Study on Pain Assessment from Multimodal Sensor Data
Pain assessment is a critical aspect of healthcare, influencing timely interventions and patient well-being. Traditional pain evaluation methods often rely on subjective patient reports, leading to inaccuracies and disparities in treatment, especially for patients who present difficulties to communicate due to cognitive impairments. Our contributions are three-fold. Firstly, we analyze the correlations of the data extracted from biomedical sensors. Then, we use state-of-the-art computer vision techniques to analyze videos focusing on the facial expressions of the patients, both per-frame and using the temporal context. We compare them and provide a baseline for pain assessment methods using two popular benchmarks: UNBC-McMaster Shoulder Pain Expression Archive Database and BioVid Heat Pain Database. We achieved an accuracy of over 96% and over 94% for the F1 Score, recall and precision metrics in pain estimation using single frames with the UNBC-McMaster dataset, employing state-of-the-art computer vision techniques such as Transformer-based architectures for vision tasks. In addition, from the conclusions drawn from the study, future lines of work in this area are discussed
Multimodaalsel emotsioonide tuvastamisel põhineva inimese-roboti suhtluse arendamine
Väitekirja elektrooniline versioon ei sisalda publikatsiooneÜks afektiivse arvutiteaduse peamistest huviobjektidest on mitmemodaalne emotsioonituvastus, mis leiab rakendust peamiselt inimese-arvuti interaktsioonis. Emotsiooni äratundmiseks uuritakse nendes süsteemides nii inimese näoilmeid kui kakõnet. Käesolevas töös uuritakse inimese emotsioonide ja nende avaldumise visuaalseid ja akustilisi tunnuseid, et töötada välja automaatne multimodaalne emotsioonituvastussüsteem. Kõnest arvutatakse mel-sageduse kepstri kordajad, helisignaali erinevate komponentide energiad ja prosoodilised näitajad. Näoilmeteanalüüsimiseks kasutatakse kahte erinevat strateegiat. Esiteks arvutatakse inimesenäo tähtsamate punktide vahelised erinevad geomeetrilised suhted. Teiseks võetakse emotsionaalse sisuga video kokku vähendatud hulgaks põhikaadriteks, misantakse sisendiks konvolutsioonilisele tehisnärvivõrgule emotsioonide visuaalsekseristamiseks. Kolme klassifitseerija väljunditest (1 akustiline, 2 visuaalset) koostatakse uus kogum tunnuseid, mida kasutatakse õppimiseks süsteemi viimasesetapis. Loodud süsteemi katsetati SAVEE, Poola ja Serbia emotsionaalse kõneandmebaaside, eNTERFACE’05 ja RML andmebaaside peal. Saadud tulemusednäitavad, et võrreldes olemasolevatega võimaldab käesoleva töö raames loodudsüsteem suuremat täpsust emotsioonide äratundmisel. Lisaks anname käesolevastöös ülevaate kirjanduses väljapakutud süsteemidest, millel on võimekus tunda äraemotsiooniga seotud ̆zeste. Selle ülevaate eesmärgiks on hõlbustada uute uurimissuundade leidmist, mis aitaksid lisada töö raames loodud süsteemile ̆zestipõhiseemotsioonituvastuse võimekuse, et veelgi enam tõsta süsteemi emotsioonide äratundmise täpsust.Automatic multimodal emotion recognition is a fundamental subject of interest in affective computing. Its main applications are in human-computer interaction. The systems developed for the foregoing purpose consider combinations of different modalities, based on vocal and visual cues. This thesis takes the foregoing modalities into account, in order to develop an automatic multimodal emotion recognition system. More specifically, it takes advantage of the information extracted from speech and face signals. From speech signals, Mel-frequency cepstral coefficients, filter-bank energies and prosodic features are extracted. Moreover, two different strategies are considered for analyzing the facial data. First, facial landmarks' geometric relations, i.e. distances and angles, are computed. Second, we summarize each emotional video into a reduced set of key-frames. Then they are taught to visually discriminate between the emotions. In order to do so, a convolutional neural network is applied to the key-frames summarizing the videos. Afterward, the output confidence values of all the classifiers from both of the modalities are used to define a new feature space. Lastly, the latter values are learned for the final emotion label prediction, in a late fusion. The experiments are conducted on the SAVEE, Polish, Serbian, eNTERFACE'05 and RML datasets. The results show significant performance improvements by the proposed system in comparison to the existing alternatives, defining the current state-of-the-art on all the datasets. Additionally, we provide a review of emotional body gesture recognition systems proposed in the literature. The aim of the foregoing part is to help figure out possible future research directions for enhancing the performance of the proposed system. More clearly, we imply that incorporating data representing gestures, which constitute another major component of the visual modality, can result in a more efficient framework
Computational Approaches for Monitoring of Health Parameters and Their Evaluation for Application in Clinical Setting.
The algorithms and mathematical methods developed in this work focus on using computational approaches for low cost solution of health care problems for better patient outcome. Furthermore, evaluation of those approaches for clinical application considering the risk and benefit in a clinical setting is studied. Those risks and benefits are discussed in terms of sensitivity, specificity and area under the receiver operating characteristics curve. With a rising cost of health care and increasing number of aging population, there is a need for innovative and low cost solutions for health care problems. In this work, algorithms, mathematical techniques for the solutions of the problems related to physiological parameter monitoring have been explored and their evaluation approaches for application in a clinical setting have been studied. The physiological parameters include affective state, pain level, heart rate, oxygen saturation, hemoglobin level and blood pressure. For the mathematical basis development for different data intensive problems, eigenvalue based methods along with others have been used in designing innovative solutions for health care problems, developing new algorithms for smart monitoring of patients; from home monitoring to combat casualty situations. Eigenvalue based methods already have wide applications in many areas such as analysis of stability in control systems, search algorithms (Google Page Rank), Eigenface methods for face recognition, principal component analysis for data compression and pattern recognition. Here, the research work in 1) multi-parameter monitoring of affective state, 2) creating a smart phone based pain detection tool from facial images, 3) early detection of hemorrhage from arterial blood pressure data, 4) noninvasive measurement of physiological signals including hemoglobin level and 5) evaluation of the results for clinical application are presented
Audio-Visual Fusion for Emotion Recognition in the Valence-Arousal Space Using Joint Cross-Attention
Automatic emotion recognition (ER) has recently gained lot of interest due to
its potential in many real-world applications. In this context, multimodal
approaches have been shown to improve performance (over unimodal approaches) by
combining diverse and complementary sources of information, providing some
robustness to noisy and missing modalities. In this paper, we focus on
dimensional ER based on the fusion of facial and vocal modalities extracted
from videos, where complementary audio-visual (A-V) relationships are explored
to predict an individual's emotional states in valence-arousal space. Most
state-of-the-art fusion techniques rely on recurrent networks or conventional
attention mechanisms that do not effectively leverage the complementary nature
of A-V modalities. To address this problem, we introduce a joint
cross-attentional model for A-V fusion that extracts the salient features
across A-V modalities, that allows to effectively leverage the inter-modal
relationships, while retaining the intra-modal relationships. In particular, it
computes the cross-attention weights based on correlation between the joint
feature representation and that of the individual modalities. By deploying the
joint A-V feature representation into the cross-attention module, it helps to
simultaneously leverage both the intra and inter modal relationships, thereby
significantly improving the performance of the system over the vanilla
cross-attention module. The effectiveness of our proposed approach is validated
experimentally on challenging videos from the RECOLA and AffWild2 datasets.
Results indicate that our joint cross-attentional A-V fusion model provides a
cost-effective solution that can outperform state-of-the-art approaches, even
when the modalities are noisy or absent.Comment: arXiv admin note: substantial text overlap with arXiv:2203.14779,
arXiv:2111.0522
- …