451 research outputs found
Design for novel enhanced weightless neural network and multi-classifier.
Weightless neural systems have often struggles in terms of speed, performances, and memory issues. There is also lack of sufficient interfacing of weightless neural systems to others systems. Addressing these issues motivates and forms the aims and objectives of this thesis. In addressing these issues, algorithms are formulated, classifiers, and multi-classifiers are designed, and hardware design of classifier are also reported. Specifically, the purpose of this thesis is to report on the algorithms and designs of weightless neural systems.
A background material for the research is a weightless neural network known as Probabilistic Convergent Network (PCN). By introducing two new and different interfacing method, the word "Enhanced" is added to PCN thereby giving it the name Enhanced Probabilistic Convergent Network (EPCN). To solve the problem of speed and performances when large-class databases are employed in data analysis, multi-classifiers are designed whose composition vary depending on problem complexity. It also leads to the introduction of a novel gating function with application of EPCN as an intelligent combiner. For databases which are not very large, single classifiers suffices. Speed and ease of application in adverse condition were considered as improvement which has led to the design of EPCN in hardware. A novel hashing function is implemented and tested on hardware-based EPCN.
Results obtained have indicated the utility of employing weightless neural systems. The results obtained also indicate significant new possible areas of application of weightless neural systems
Audio-coupled video content understanding of unconstrained video sequences
Unconstrained video understanding is a difficult task. The main aim of this thesis is to
recognise the nature of objects, activities and environment in a given video clip using
both audio and video information. Traditionally, audio and video information has not
been applied together for solving such complex task, and for the first time we propose,
develop, implement and test a new framework of multi-modal (audio and video) data
analysis for context understanding and labelling of unconstrained videos.
The framework relies on feature selection techniques and introduces a novel algorithm
(PCFS) that is faster than the well-established SFFS algorithm. We use the framework for
studying the benefits of combining audio and video information in a number of different
problems. We begin by developing two independent content recognition modules. The
first one is based on image sequence analysis alone, and uses a range of colour, shape,
texture and statistical features from image regions with a trained classifier to recognise
the identity of objects, activities and environment present. The second module uses audio
information only, and recognises activities and environment. Both of these approaches
are preceded by detailed pre-processing to ensure that correct video segments containing
both audio and video content are present, and that the developed system can be made
robust to changes in camera movement, illumination, random object behaviour etc. For
both audio and video analysis, we use a hierarchical approach of multi-stage
classification such that difficult classification tasks can be decomposed into simpler and
smaller tasks.
When combining both modalities, we compare fusion techniques at different levels of
integration and propose a novel algorithm that combines advantages of both feature and
decision-level fusion. The analysis is evaluated on a large amount of test data comprising
unconstrained videos collected for this work. We finally, propose a decision correction
algorithm which shows that further steps towards combining multi-modal classification
information effectively with semantic knowledge generates the best possible results
Bio-mimetic Spiking Neural Networks for unsupervised clustering of spatio-temporal data
Spiking neural networks aspire to mimic the brain more closely than traditional artificial neural networks. They are characterised by a spike-like activation function inspired by the shape of an action potential in biological neurons. Spiking networks remain a niche area of research, perform worse than the traditional artificial networks, and their real-world applications are limited. We hypothesised that neuroscience-inspired spiking neural networks with spike-timing-dependent plasticity demonstrate useful learning capabilities. Our objective was to identify features which play a vital role in information processing in the brain but are not commonly used in artificial networks, implement them in spiking networks without copying constraints that apply to living organisms, and to characterise their effect on data processing. The networks we created are not brain models; our approach can be labelled as artificial life. We performed a literature review and selected features such as local weight updates, neuronal sub-types, modularity, homeostasis and structural plasticity. We used the review as a guide for developing the consecutive iterations of the network, and eventually a whole evolutionary developmental system. We analysed the modelâs performance on clustering of spatio-temporal data. Our results show that combining evolution and unsupervised learning leads to a faster convergence on the optimal solutions, better stability of fit solutions than each approach separately. The choice of fitness definition affects the networkâs performance on fitness-related and unrelated tasks. We found that neuron type-specific weight homeostasis can be used to stabilise the networks, thus enabling longer training. We also demonstrated that networks with a rudimentary architecture can evolve developmental rules which improve their fitness. This interdisciplinary work provides contributions to three fields: it proposes novel artificial intelligence approaches, tests the possible role of the selected biological phenomena in information processing in the brain, and explores the evolution of learning in an artificial life system
Fear Classification using Affective Computing with Physiological Information and Smart-Wearables
MenciĂłn Internacional en el tĂtulo de doctorAmong the 17 Sustainable Development Goals proposed within the 2030 Agenda
and adopted by all of the United Nations member states, the fifth SDG is a call
for action to effectively turn gender equality into a fundamental human right and
an essential foundation for a better world. It includes the eradication of all types
of violence against women. Focusing on the technological perspective, the range of
available solutions intended to prevent this social problem is very limited. Moreover,
most of the solutions are based on a panic button approach, leaving aside
the usage and integration of current state-of-the-art technologies, such as the Internet
of Things (IoT), affective computing, cyber-physical systems, and smart-sensors.
Thus, the main purpose of this research is to provide new insight into the design and
development of tools to prevent and combat Gender-based Violence risky situations
and, even, aggressions, from a technological perspective, but without leaving aside
the different sociological considerations directly related to the problem. To achieve
such an objective, we rely on the application of affective computing from a realist
point of view, i.e. targeting the generation of systems and tools capable of being implemented
and used nowadays or within an achievable time-frame. This pragmatic
vision is channelled through: 1) an exhaustive study of the existing technological
tools and mechanisms oriented to the fight Gender-based Violence, 2) the proposal
of a new smart-wearable system intended to deal with some of the current technological
encountered limitations, 3) a novel fear-related emotion classification approach
to disentangle the relation between emotions and physiology, and 4) the definition
and release of a new multi-modal dataset for emotion recognition in women.
Firstly, different fear classification systems using a reduced set of physiological signals are explored and designed. This is done by employing open datasets together
with the combination of time, frequency and non-linear domain techniques. This
design process is encompassed by trade-offs between both physiological considerations
and embedded capabilities. The latter is of paramount importance due to
the edge-computing focus of this research. Two results are highlighted in this first
task, the designed fear classification system that employed the DEAP dataset data
and achieved an AUC of 81.60% and a Gmean of 81.55% on average for a subjectindependent
approach, and only two physiological signals; and the designed fear
classification system that employed the MAHNOB dataset data achieving an AUC
of 86.00% and a Gmean of 73.78% on average for a subject-independent approach,
only three physiological signals, and a Leave-One-Subject-Out configuration. A detailed
comparison with other emotion recognition systems proposed in the literature
is presented, which proves that the obtained metrics are in line with the state-ofthe-
art.
Secondly, Bindi is presented. This is an end-to-end autonomous multimodal system
leveraging affective IoT throughout auditory and physiological commercial off-theshelf
smart-sensors, hierarchical multisensorial fusion, and secured server architecture
to combat Gender-based Violence by automatically detecting risky situations
based on a multimodal intelligence engine and then triggering a protection protocol.
Specifically, this research is focused onto the hardware and software design of one of
the two edge-computing devices within Bindi. This is a bracelet integrating three
physiological sensors, actuators, power monitoring integrated chips, and a System-
On-Chip with wireless capabilities. Within this context, different embedded design
space explorations are presented: embedded filtering evaluation, online physiological
signal quality assessment, feature extraction, and power consumption analysis.
The reported results in all these processes are successfully validated and, for some
of them, even compared against physiological standard measurement equipment.
Amongst the different obtained results regarding the embedded design and implementation
within the bracelet of Bindi, it should be highlighted that its low power
consumption provides a battery life to be approximately 40 hours when using a 500
mAh battery.
Finally, the particularities of our use case and the scarcity of open multimodal datasets dealing with emotional immersive technology, labelling methodology considering
the gender perspective, balanced stimuli distribution regarding the target
emotions, and recovery processes based on the physiological signals of the volunteers
to quantify and isolate the emotional activation between stimuli, led us to the definition
and elaboration of Women and Emotion Multi-modal Affective Computing
(WEMAC) dataset. This is a multimodal dataset in which 104 women who never
experienced Gender-based Violence that performed different emotion-related stimuli
visualisations in a laboratory environment. The previous fear binary classification
systems were improved and applied to this novel multimodal dataset. For instance,
the proposed multimodal fear recognition system using this dataset reports up to
60.20% and 67.59% for ACC and F1-score, respectively. These values represent a
competitive result in comparison with the state-of-the-art that deal with similar
multi-modal use cases.
In general, this PhD thesis has opened a new research line within the research group
under which it has been developed. Moreover, this work has established a solid base
from which to expand knowledge and continue research targeting the generation of
both mechanisms to help vulnerable groups and socially oriented technology.Programa de Doctorado en IngenierĂa ElĂ©ctrica, ElectrĂłnica y AutomĂĄtica por la Universidad Carlos III de MadridPresidente: David Atienza Alonso.- Secretaria: Susana PatĂłn Ălvarez.- Vocal: Eduardo de la Torre Arnan
The Development of an assistive chair for elderly with sit to stand problems
A thesis submitted to the University of Bedfordshire in partial fulfilment of the requirements for the degree of Doctor of PhilosophyStanding up from a seated position, known as sit-to-stand (STS) movement, is one of the most frequently performed activities of daily living (ADLs). However, the aging generation are often encountered with STS issues owning to their declined motor functions and sensory capacity for postural control. The motivated is rooted from the contemporary market available STS assistive devices that are lack of genuine interaction with elderly users. Prior to the software implementation, the robot chair platform with integrated sensing footmat is developed with STS biomechanical concerns for the elderly.
The work has its main emphasis on recognising the personalised behavioural patterns from the elderly usersâ STS movements, namely the STS intentions and personalised STS feature prediction. The former is known as intention recognition while the latter is defined as assistance prediction, both achieved by innovative machine learning techniques. The proposed intention recognition performs well in multiple subjects scenarios with different postures involved thanks to its competence of handling these uncertainties. To the provision of providing the assistance needed by the elderly user, a time series prediction model is presented, aiming to configure the personalised ground reaction force (GRF) curve over time which suggests successful movement. This enables the computation of deficits between the predicted oncoming GRF curve and the personalised one. A multiple steps ahead prediction into the future is also implemented so that the completion time of actuation in reality is taken into account
- âŠ