22 research outputs found
Design and Prototyping of a Bio-inspired Kinematic Sensing Suit for the Shoulder Joint: Precursor to a Multi-DoF Shoulder Exosuit
Soft wearable robots are a promising new design paradigm for rehabilitation
and active assistance applications. Their compliant nature makes them ideal for
complex joints like the shoulder, but intuitive control of these robots require
robust and compliant sensing mechanisms. In this work, we introduce the sensing
framework for a multi-DoF shoulder exosuit capable of sensing the kinematics of
the shoulder joint. The proposed tendon-based sensing system is inspired by the
concept of muscle synergies, the body's sense of proprioception, and finds its
basis in the organization of the muscles responsible for shoulder movements. A
motion-capture-based evaluation of the developed sensing system showed
conformance to the behaviour exhibited by the muscles that inspired its routing
and validates the hypothesis of the tendon-routing to be extended to the
actuation framework of the exosuit in the future. The mapping from multi-sensor
space to joint space is a multivariate multiple regression problem and was
derived using an Artificial Neural Network (ANN). The sensing framework was
tested with a motion-tracking system and achieved performance with root mean
square error (RMSE) of approximately 5.43 degrees and 3.65 degrees for the
azimuth and elevation joint angles, respectively, measured over 29000 frames
(4+ minutes) of motion-capture data.Comment: 8 pages, 7 figures, 1 tabl
Nonlinearity Compensation in a Multi-DoF Shoulder Sensing Exosuit for Real-Time Teleoperation
The compliant nature of soft wearable robots makes them ideal for complex
multiple degrees of freedom (DoF) joints, but also introduce additional
structural nonlinearities. Intuitive control of these wearable robots requires
robust sensing to overcome the inherent nonlinearities. This paper presents a
joint kinematics estimator for a bio-inspired multi-DoF shoulder exosuit
capable of compensating the encountered nonlinearities. To overcome the
nonlinearities and hysteresis inherent to the soft and compliant nature of the
suit, we developed a deep learning-based method to map the sensor data to the
joint space. The experimental results show that the new learning-based
framework outperforms recent state-of-the-art methods by a large margin while
achieving 12ms inference time using only a GPU-based edge-computing device. The
effectiveness of our combined exosuit and learning framework is demonstrated
through real-time teleoperation with a simulated NAO humanoid robot.Comment: 8 pages, 7 figures, 3 tables. Accepted to be published in IEEE
RoboSoft 202
Egocentric Image Captioning for Privacy-Preserved Passive Dietary Intake Monitoring
Camera-based passive dietary intake monitoring is able to continuously
capture the eating episodes of a subject, recording rich visual information,
such as the type and volume of food being consumed, as well as the eating
behaviours of the subject. However, there currently is no method that is able
to incorporate these visual clues and provide a comprehensive context of
dietary intake from passive recording (e.g., is the subject sharing food with
others, what food the subject is eating, and how much food is left in the
bowl). On the other hand, privacy is a major concern while egocentric wearable
cameras are used for capturing. In this paper, we propose a privacy-preserved
secure solution (i.e., egocentric image captioning) for dietary assessment with
passive monitoring, which unifies food recognition, volume estimation, and
scene understanding. By converting images into rich text descriptions,
nutritionists can assess individual dietary intake based on the captions
instead of the original images, reducing the risk of privacy leakage from
images. To this end, an egocentric dietary image captioning dataset has been
built, which consists of in-the-wild images captured by head-worn and
chest-worn cameras in field studies in Ghana. A novel transformer-based
architecture is designed to caption egocentric dietary images. Comprehensive
experiments have been conducted to evaluate the effectiveness and to justify
the design of the proposed architecture for egocentric dietary image
captioning. To the best of our knowledge, this is the first work that applies
image captioning to dietary intake assessment in real life settings
Egocentric image captioning for privacy-preserved passive dietary intake monitoring
Camera-based passive dietary intake monitoring is able to continuously capture the eating episodes of a subject, recording rich visual information, such as the type and volume of food being consumed, as well as the eating behaviors of the subject. However, there currently is no method that is able to incorporate these visual clues and provide a comprehensive context of dietary intake from passive recording (e.g., is the subject sharing food with others, what food the subject is eating, and how much food is left in the bowl). On the other hand, privacy is a major concern while egocentric wearable cameras are used for capturing. In this article, we propose a privacy-preserved secure solution (i.e., egocentric image captioning) for dietary assessment with passive monitoring, which unifies food recognition, volume estimation, and scene understanding. By converting images into rich text descriptions, nutritionists can assess individual dietary intake based on the captions instead of the original images, reducing the risk of privacy leakage from images. To this end, an egocentric dietary image captioning dataset has been built, which consists of in-the-wild images captured by head-worn and chest-worn cameras in field studies in Ghana. A novel transformer-based architecture is designed to caption egocentric dietary images. Comprehensive experiments have been conducted to evaluate the effectiveness and to justify the design of the proposed architecture for egocentric dietary image captioning. To the best of our knowledge, this is the first work that applies image captioning for dietary intake assessment in real-life settings
The global burden of cancer attributable to risk factors, 2010-19 : a systematic analysis for the Global Burden of Disease Study 2019
Background Understanding the magnitude of cancer burden attributable to potentially modifiable risk factors is crucial for development of effective prevention and mitigation strategies. We analysed results from the Global Burden of Diseases, Injuries, and Risk Factors Study (GBD) 2019 to inform cancer control planning efforts globally. Methods The GBD 2019 comparative risk assessment framework was used to estimate cancer burden attributable to behavioural, environmental and occupational, and metabolic risk factors. A total of 82 risk-outcome pairs were included on the basis of the World Cancer Research Fund criteria. Estimated cancer deaths and disability-adjusted life-years (DALYs) in 2019 and change in these measures between 2010 and 2019 are presented. Findings Globally, in 2019, the risk factors included in this analysis accounted for 4.45 million (95% uncertainty interval 4.01-4.94) deaths and 105 million (95.0-116) DALYs for both sexes combined, representing 44.4% (41.3-48.4) of all cancer deaths and 42.0% (39.1-45.6) of all DALYs. There were 2.88 million (2.60-3.18) risk-attributable cancer deaths in males (50.6% [47.8-54.1] of all male cancer deaths) and 1.58 million (1.36-1.84) risk-attributable cancer deaths in females (36.3% [32.5-41.3] of all female cancer deaths). The leading risk factors at the most detailed level globally for risk-attributable cancer deaths and DALYs in 2019 for both sexes combined were smoking, followed by alcohol use and high BMI. Risk-attributable cancer burden varied by world region and Socio-demographic Index (SDI), with smoking, unsafe sex, and alcohol use being the three leading risk factors for risk-attributable cancer DALYs in low SDI locations in 2019, whereas DALYs in high SDI locations mirrored the top three global risk factor rankings. From 2010 to 2019, global risk-attributable cancer deaths increased by 20.4% (12.6-28.4) and DALYs by 16.8% (8.8-25.0), with the greatest percentage increase in metabolic risks (34.7% [27.9-42.8] and 33.3% [25.8-42.0]). Interpretation The leading risk factors contributing to global cancer burden in 2019 were behavioural, whereas metabolic risk factors saw the largest increases between 2010 and 2019. Reducing exposure to these modifiable risk factors would decrease cancer mortality and DALY rates worldwide, and policies should be tailored appropriately to local cancer risk factor burden. Copyright (C) 2022 The Author(s). Published by Elsevier Ltd. This is an Open Access article under the CC BY 4.0 license.Peer reviewe
Abstract. Neuro-Fuzzy Shadow Filter
In video sequence processing, shadow remains a major source of error for object segmentation. Traditional methods of shadow removal are mainly based on colour difference thresholding between the background and current images. The application of colour filters on MPEG or MJPEG images, however, is often erroneous as the chrominance information is significantly reduced due to compression. In addition, as the colour attributes of shadows and objects are often very similar, discrete thresholding cannot always provide reliable results. This paper presents a novel approach for adaptive shadow removal by incorporating four different filters in a neuro-fuzzy framework. The neurofuzzy classifier has the ability of real-time self-adaptation and training, and its performance has been quantitatively assessed with both indoor and outdoor video sequences