16,515 research outputs found
Multimodal Classification of Urban Micro-Events
In this paper we seek methods to effectively detect urban micro-events. Urban
micro-events are events which occur in cities, have limited geographical
coverage and typically affect only a small group of citizens. Because of their
scale these are difficult to identify in most data sources. However, by using
citizen sensing to gather data, detecting them becomes feasible. The data
gathered by citizen sensing is often multimodal and, as a consequence, the
information required to detect urban micro-events is distributed over multiple
modalities. This makes it essential to have a classifier capable of combining
them. In this paper we explore several methods of creating such a classifier,
including early, late, hybrid fusion and representation learning using
multimodal graphs. We evaluate performance on a real world dataset obtained
from a live citizen reporting system. We show that a multimodal approach yields
higher performance than unimodal alternatives. Furthermore, we demonstrate that
our hybrid combination of early and late fusion with multimodal embeddings
performs best in classification of urban micro-events
Multimedia search without visual analysis: the value of linguistic and contextual information
This paper addresses the focus of this special issue by analyzing the potential contribution of linguistic content and other non-image aspects to the processing of audiovisual data. It summarizes the various ways in which linguistic content analysis contributes to enhancing the semantic annotation of multimedia content, and, as a consequence, to improving the effectiveness of conceptual media access tools. A number of techniques are presented, including the time-alignment of textual resources, audio and speech processing, content reduction and reasoning tools, and the exploitation of surface features
How will the Internet of Things enable Augmented Personalized Health?
Internet-of-Things (IoT) is profoundly redefining the way we create, consume,
and share information. Health aficionados and citizens are increasingly using
IoT technologies to track their sleep, food intake, activity, vital body
signals, and other physiological observations. This is complemented by IoT
systems that continuously collect health-related data from the environment and
inside the living quarters. Together, these have created an opportunity for a
new generation of healthcare solutions. However, interpreting data to
understand an individual's health is challenging. It is usually necessary to
look at that individual's clinical record and behavioral information, as well
as social and environmental information affecting that individual. Interpreting
how well a patient is doing also requires looking at his adherence to
respective health objectives, application of relevant clinical knowledge and
the desired outcomes.
We resort to the vision of Augmented Personalized Healthcare (APH) to exploit
the extensive variety of relevant data and medical knowledge using Artificial
Intelligence (AI) techniques to extend and enhance human health to presents
various stages of augmented health management strategies: self-monitoring,
self-appraisal, self-management, intervention, and disease progress tracking
and prediction. kHealth technology, a specific incarnation of APH, and its
application to Asthma and other diseases are used to provide illustrations and
discuss alternatives for technology-assisted health management. Several
prominent efforts involving IoT and patient-generated health data (PGHD) with
respect converting multimodal data into actionable information (big data to
smart data) are also identified. Roles of three components in an evidence-based
semantic perception approach- Contextualization, Abstraction, and
Personalization are discussed
Anticipatory Mobile Computing: A Survey of the State of the Art and Research Challenges
Today's mobile phones are far from mere communication devices they were ten
years ago. Equipped with sophisticated sensors and advanced computing hardware,
phones can be used to infer users' location, activity, social setting and more.
As devices become increasingly intelligent, their capabilities evolve beyond
inferring context to predicting it, and then reasoning and acting upon the
predicted context. This article provides an overview of the current state of
the art in mobile sensing and context prediction paving the way for
full-fledged anticipatory mobile computing. We present a survey of phenomena
that mobile phones can infer and predict, and offer a description of machine
learning techniques used for such predictions. We then discuss proactive
decision making and decision delivery via the user-device feedback loop.
Finally, we discuss the challenges and opportunities of anticipatory mobile
computing.Comment: 29 pages, 5 figure
Applying semantic web technologies to knowledge sharing in aerospace engineering
This paper details an integrated methodology to optimise Knowledge reuse and sharing, illustrated with a use case in the aeronautics domain. It uses Ontologies as a central modelling strategy for the Capture of Knowledge from legacy docu-ments via automated means, or directly in systems interfacing with Knowledge workers, via user-defined, web-based forms. The domain ontologies used for Knowledge Capture also guide the retrieval of the Knowledge extracted from the data using a Semantic Search System that provides support for multiple modalities during search. This approach has been applied and evaluated successfully within the aerospace domain, and is currently being extended for use in other domains on an increasingly large scale
- âŠ