200,528 research outputs found
Detection of bimanual gestures everywhere: why it matters, what we need and what is missing
Bimanual gestures are of the utmost importance for the study of motor
coordination in humans and in everyday activities. A reliable detection of
bimanual gestures in unconstrained environments is fundamental for their
clinical study and to assess common activities of daily living. This paper
investigates techniques for a reliable, unconstrained detection and
classification of bimanual gestures. It assumes the availability of inertial
data originating from the two hands/arms, builds upon a previously developed
technique for gesture modelling based on Gaussian Mixture Modelling (GMM) and
Gaussian Mixture Regression (GMR), and compares different modelling and
classification techniques, which are based on a number of assumptions inspired
by literature about how bimanual gestures are represented and modelled in the
brain. Experiments show results related to 5 everyday bimanual activities,
which have been selected on the basis of three main parameters: (not)
constraining the two hands by a physical tool, (not) requiring a specific
sequence of single-hand gestures, being recursive (or not). In the best
performing combination of modeling approach and classification technique, five
out of five activities are recognized up to an accuracy of 97%, a precision of
82% and a level of recall of 100%.Comment: Submitted to Robotics and Autonomous Systems (Elsevier
General-purpose and special-purpose visual systems
The information that eyes supply supports a wide variety of functions, from the guidance systems that enable an animal to navigate successfully around the environment, to the detection and identification of predators, prey, and conspecifics. The eyes with which we are most familiar the single-chambered eyes of vertebrates and cephalopod molluscs, and the compound eyes of insects and higher crustaceans allow these animals to perform the full range of visual tasks. These eyes have evidently evolved in conjunction with brains that are capable of subjecting the raw visual information to many different kinds of analysis, depending on the nature of the task that the animal is engaged in. However, not all eyes evolved to provide such comprehensive information. For example, in bivalve molluscs we find eyes of very varied design (pinholes, concave mirrors, and apposition compound eyes) whose only function is to detect approaching predators and thereby allow the animal to protect itself by closing its shell. Thus, there are special-purpose eyes as well as eyes with multiple functions
Interoperable services based on activity monitoring in ambient assisted living environments
Ambient Assisted Living (AAL) is considered as the main technological solution that will enable the aged and people in recovery to maintain their independence and a consequent high quality of life for a longer period of time than would otherwise be the case. This goal is achieved by monitoring human’s activities and deploying the appropriate collection of services to set environmental features and satisfy user preferences in a given context. However, both human monitoring and services deployment are particularly hard to accomplish due to the uncertainty and ambiguity characterising human actions, and heterogeneity of hardware devices composed in an AAL system. This research addresses both the aforementioned challenges by introducing 1) an innovative system, based on Self Organising Feature Map (SOFM), for automatically classifying the resting location of a moving object in an indoor environment and 2) a strategy able to generate context-aware based Fuzzy Markup Language (FML) services in order to maximize the users’ comfort and hardware interoperability level. The overall system runs on a distributed embedded platform with a specialised ceiling- mounted video sensor for intelligent activity monitoring. The system has the ability to learn resting locations, to measure overall activity levels, to detect specific events such as potential falls and to deploy the right sequence of fuzzy services modelled through FML for supporting people in that particular context. Experimental results show less than 20% classification error in monitoring human activities and providing the right set of services, showing the robustness of our approach over others in literature with minimal power consumption
Structured evaluation of virtual environments for special-needs education
This paper describes the development of a structured approach to evaluate experiential and communication virtual learning environments (VLEs) designed specifically for use in the education of children with severe learning difficulties at the Shepherd special needs school in Nottingham, UK. Constructivist learning theory was used as a basis for the production of an evaluation framework, used to evaluate the design of three VLEs and how they were used by students with respect to this learning theory. From an observational field study of student-teacher pairs using the VLEs, 18 behaviour categories were identified as relevant to five of the seven constructivist principles defined by Jonassen (1994). Analysis of student-teacher behaviour was used to provide support for, or against, the constructivist principles. The results show that the three VLEs meet the constructivist principles in very different ways and recommendations for design modifications are put forward
Real-time human ambulation, activity, and physiological monitoring:taxonomy of issues, techniques, applications, challenges and limitations
Automated methods of real-time, unobtrusive, human ambulation, activity, and wellness monitoring and data analysis using various algorithmic techniques have been subjects of intense research. The general aim is to devise effective means of addressing the demands of assisted living, rehabilitation, and clinical observation and assessment through sensor-based monitoring. The research studies have resulted in a large amount of literature. This paper presents a holistic articulation of the research studies and offers comprehensive insights along four main axes: distribution of existing studies; monitoring device framework and sensor types; data collection, processing and analysis; and applications, limitations and challenges. The aim is to present a systematic and most complete study of literature in the area in order to identify research gaps and prioritize future research directions
Design and semantics of form and movement (DeSForM 2006)
Design and Semantics of Form and Movement (DeSForM) grew from applied research exploring emerging design methods and practices to support new generation product and interface design. The products and interfaces are concerned with: the context of ubiquitous computing and ambient technologies and the need for greater empathy in the pre-programmed behaviour of the ‘machines’ that populate our lives. Such explorative research in the CfDR has been led by Young, supported by Kyffin, Visiting Professor from Philips Design and sponsored by Philips Design over a period of four years (research funding £87k). DeSForM1 was the first of a series of three conferences that enable the presentation and debate of international work within this field: • 1st European conference on Design and Semantics of Form and Movement (DeSForM1), Baltic, Gateshead, 2005, Feijs L., Kyffin S. & Young R.A. eds. • 2nd European conference on Design and Semantics of Form and Movement (DeSForM2), Evoluon, Eindhoven, 2006, Feijs L., Kyffin S. & Young R.A. eds. • 3rd European conference on Design and Semantics of Form and Movement (DeSForM3), New Design School Building, Newcastle, 2007, Feijs L., Kyffin S. & Young R.A. eds. Philips sponsorship of practice-based enquiry led to research by three teams of research students over three years and on-going sponsorship of research through the Northumbria University Design and Innovation Laboratory (nuDIL). Young has been invited on the steering panel of the UK Thinking Digital Conference concerning the latest developments in digital and media technologies. Informed by this research is the work of PhD student Yukie Nakano who examines new technologies in relation to eco-design textiles
Recommended from our members
Behavioural pattern identification and prediction in intelligent environments
In this paper, the application of soft computing techniques in prediction of an occupant's behaviour in an inhabited intelligent environment is addressed. In this research, daily activities of elderly people who live in their own homes suffering from dementia are studied. Occupancy sensors are used to extract the movement patterns of the occupant. The occupancy data is then converted into temporal sequences of activities which are eventually used to predict the occupant behaviour. To build the prediction model, different dynamic recurrent neural networks are investigated. Recurrent neural networks have shown a great ability in finding the temporal relationships of input patterns. The experimental results show that non-linear autoregressive network with exogenous inputs model correctly extracts the long term prediction patterns of the occupant and outperformed the Elman network. The results presented here are validated using data generated from a simulator and real environments
Assistive technology design and development for acceptable robotics companions for ageing years
© 2013 Farshid Amirabdollahian et al., licensee Versita Sp. z o. o. This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivs license, which means that the text may be used for non-commercial purposes, provided credit is given to the author.A new stream of research and development responds to changes in life expectancy across the world. It includes technologies which enhance well-being of individuals, specifically for older people. The ACCOMPANY project focuses on home companion technologies and issues surrounding technology development for assistive purposes. The project responds to some overlooked aspects of technology design, divided into multiple areas such as empathic and social human-robot interaction, robot learning and memory visualisation, and monitoring persons’ activities at home. To bring these aspects together, a dedicated task is identified to ensure technological integration of these multiple approaches on an existing robotic platform, Care-O-Bot®3 in the context of a smart-home environment utilising a multitude of sensor arrays. Formative and summative evaluation cycles are then used to assess the emerging prototype towards identifying acceptable behaviours and roles for the robot, for example role as a butler or a trainer, while also comparing user requirements to achieved progress. In a novel approach, the project considers ethical concerns and by highlighting principles such as autonomy, independence, enablement, safety and privacy, it embarks on providing a discussion medium where user views on these principles and the existing tension between some of these principles, for example tension between privacy and autonomy over safety, can be captured and considered in design cycles and throughout project developmentsPeer reviewe
Use Cases for Abnormal Behaviour Detection in Smart Homes
While people have many ideas about how a smart home should react to particular behaviours from their inhabitant, there seems to have been relatively little attempt to organise this systematically. In this paper, we attempt to rectify this in consideration of context awareness and novelty detection for a smart home that monitors its inhabitant for illness and unexpected behaviour. We do this through the concept of the Use Case, which is used in software engineering to specify the behaviour of a system. We describe a set of scenarios and the possible outputs that the smart home could give and introduce the SHMUC Repository of Smart Home Use Cases. Based on this, we can consider how probabilistic and logic-based reasoning systems would produce different capabilities
- …