109 research outputs found
Objective Classes for Micro-Facial Expression Recognition
Micro-expressions are brief spontaneous facial expressions that appear on a
face when a person conceals an emotion, making them different to normal facial
expressions in subtlety and duration. Currently, emotion classes within the
CASME II dataset are based on Action Units and self-reports, creating conflicts
during machine learning training. We will show that classifying expressions
using Action Units, instead of predicted emotion, removes the potential bias of
human reporting. The proposed classes are tested using LBP-TOP, HOOF and HOG 3D
feature descriptors. The experiments are evaluated on two benchmark FACS coded
datasets: CASME II and SAMM. The best result achieves 86.35\% accuracy when
classifying the proposed 5 classes on CASME II using HOG 3D, outperforming the
result of the state-of-the-art 5-class emotional-based classification in CASME
II. Results indicate that classification based on Action Units provides an
objective method to improve micro-expression recognition.Comment: 11 pages, 4 figures and 5 tables. This paper will be submitted for
journal revie
Enhanced algorithms for lesion detection and recognition in ultrasound breast images
Mammography is the gold standard for breast cancer detection. However, it has very
high false positive rates and is based on ionizing radiation. This has led to interest in
using multi-modal approaches. One modality is diagnostic ultrasound, which is based
on non-ionizing radiation and picks up many of the cancers that are generally missed
by mammography. However, the presence of speckle noise in ultrasound images has a
negative effect on image interpretation. Noise reduction, inconsistencies in capture
and segmentation of lesions still remain challenging open research problems in
ultrasound images.
The target of the proposed research is to enhance the state-of-art computer vision
algorithms used in ultrasound imaging and to investigate the role of computer
processed images in human diagnostic performance. [Continues.
Grid-enabled mammographic auditing and training system
Effective use of new technologies to support healthcare initiatives is important and current research is moving towards
implementing secure grid-enabled healthcare provision. In the UK, a large-scale collaborative research project (GIMI:
Generic Infrastructures for Medical Informatics), which is concerned with the development of a secure IT infrastructure
to support very widespread medical research across the country, is underway. In the UK, there are some 109 breast
screening centers and a growing number of individuals (circa 650) nationally performing approximately 1.5 million
screening examinations per year. At the same, there is a serious, and ongoing, national workforce issue in screening
which has seen a loss of consultant mammographers and a growth in specially trained technologists and other nonradiologists.
Thus there is a need to offer effective and efficient mammographic training so as to maintain high levels of
screening skills. Consequently, a grid based system has been proposed which has the benefit of offering very large
volumes of training cases that the mammographers can access anytime and anywhere. A database, spread geographically
across three university systems, of screening cases is used as a test set of known cases. The GIMI mammography
training system first audits these cases to ensure that they are appropriately described and annotated. Subsequently, the
cases are utilized for training in a grid-based system which has been developed. This paper briefly reviews the
background to the project and then details the ongoing research. In conclusion, we discuss the contributions, limitations,
and future plans of such a grid based approach
3D-CNN for Facial Micro- and Macro-expression Spotting on Long Video Sequences using Temporal Oriented Reference Frame
Facial expression spotting is the preliminary step for micro- and
macro-expression analysis. The task of reliably spotting such expressions in
video sequences is currently unsolved. The current best systems depend upon
optical flow methods to extract regional motion features, before categorisation
of that motion into a specific class of facial movement. Optical flow is
susceptible to drift error, which introduces a serious problem for motions with
long-term dependencies, such as high frame-rate macro-expression. We propose a
purely deep learning solution which, rather than track frame differential
motion, compares via a convolutional model, each frame with two temporally
local reference frames. Reference frames are sampled according to calculated
micro- and macro-expression durations. We show that our solution achieves
state-of-the-art performance (F1-score of 0.126) in a dataset of high
frame-rate (200 fps) long video sequences (SAMM-LV) and is competitive in a low
frame-rate (30 fps) dataset (CAS(ME)2). In this paper, we document our deep
learning model and parameters, including how we use local contrast
normalisation, which we show is critical for optimal results. We surpass a
limitation in existing methods, and advance the state of deep learning in the
domain of facial expression spotting
- …