14,251 research outputs found
A study in the cognition of individualsâ identity: Solving the problem of singular cognition in object and agent tracking
This article compares the ability to track individuals lacking mental states with the ability to track intentional agents. It explains why reference to individuals raises the problem of explaining how cognitive agents track unique individuals and in what sense reference is based on procedures of perceptual-motor and epistemic tracking. We suggest applying the notion of singular-files from theories in perception and semantics to the problem of tracking intentional agents. In order to elucidate the nature of agent-files, three views of the relation between object- and agent-tracking are distinguished: the Independence, Deflationary and Organism-Dependence Views. The correct view is argued to be the latter, which states that perceptual and epistemic tracking of a unique human organism requires tracking both its spatio-temporal object-properties and its agent-properties
Enhanced tracking and recognition of moving objects by reasoning about spatio-temporal continuity.
A framework for the logical and statistical analysis and annotation of dynamic scenes containing occlusion and other uncertainties is presented. This framework consists
of three elements; an object tracker module, an object recognition/classification module and a logical consistency, ambiguity and error reasoning engine. The principle behind the object tracker and object recognition modules is to reduce error by increasing ambiguity (by merging objects in close proximity and presenting multiple
hypotheses). The reasoning engine deals with error, ambiguity and occlusion in a unified framework to produce a hypothesis that satisfies fundamental constraints
on the spatio-temporal continuity of objects. Our algorithm finds a globally consistent model of an extended video sequence that is maximally supported by a voting function based on the output of a statistical classifier. The system results
in an annotation that is significantly more accurate than what would be obtained
by frame-by-frame evaluation of the classifier output. The framework has been implemented
and applied successfully to the analysis of team sports with a single
camera.
Key words: Visua
Geospatial Narratives and their Spatio-Temporal Dynamics: Commonsense Reasoning for High-level Analyses in Geographic Information Systems
The modelling, analysis, and visualisation of dynamic geospatial phenomena
has been identified as a key developmental challenge for next-generation
Geographic Information Systems (GIS). In this context, the envisaged
paradigmatic extensions to contemporary foundational GIS technology raises
fundamental questions concerning the ontological, formal representational, and
(analytical) computational methods that would underlie their spatial
information theoretic underpinnings.
We present the conceptual overview and architecture for the development of
high-level semantic and qualitative analytical capabilities for dynamic
geospatial domains. Building on formal methods in the areas of commonsense
reasoning, qualitative reasoning, spatial and temporal representation and
reasoning, reasoning about actions and change, and computational models of
narrative, we identify concrete theoretical and practical challenges that
accrue in the context of formal reasoning about `space, events, actions, and
change'. With this as a basis, and within the backdrop of an illustrated
scenario involving the spatio-temporal dynamics of urban narratives, we address
specific problems and solutions techniques chiefly involving `qualitative
abstraction', `data integration and spatial consistency', and `practical
geospatial abduction'. From a broad topical viewpoint, we propose that
next-generation dynamic GIS technology demands a transdisciplinary scientific
perspective that brings together Geography, Artificial Intelligence, and
Cognitive Science.
Keywords: artificial intelligence; cognitive systems; human-computer
interaction; geographic information systems; spatio-temporal dynamics;
computational models of narrative; geospatial analysis; geospatial modelling;
ontology; qualitative spatial modelling and reasoning; spatial assistance
systemsComment: ISPRS International Journal of Geo-Information (ISSN 2220-9964);
Special Issue on: Geospatial Monitoring and Modelling of Environmental
Change}. IJGI. Editor: Duccio Rocchini. (pre-print of article in press
Progressive Transient Photon Beams
In this work we introduce a novel algorithm for transient rendering in
participating media. Our method is consistent, robust, and is able to generate
animations of time-resolved light transport featuring complex caustic light
paths in media. We base our method on the observation that the spatial
continuity provides an increased coverage of the temporal domain, and
generalize photon beams to transient-state. We extend the beam steady-state
radiance estimates to include the temporal domain. Then, we develop a
progressive version of spatio-temporal density estimations, that converges to
the correct solution with finite memory requirements by iteratively averaging
several realizations of independent renders with a progressively reduced kernel
bandwidth. We derive the optimal convergence rates accounting for space and
time kernels, and demonstrate our method against previous consistent transient
rendering methods for participating media
Recommended from our members
Time as a strand of the dance medium
Time and space are at the core of our aesthetic experiences of dance performances, yet dance has been frequently categorised as a space-based art. In this paper I revise the choreological perspective developed by Preston-Dunlop and SĂĄnchez-Colberg that conceives dance as an embodied performative art articulated in a multistranded medium (performer, movement, sound, space). I argue that time should be allowed a distinct place in the choreological discourse since its presence is key to the expressivity of a dance piece. I conceptualise the meaning of the time strand and expose how different substrands emerge, connect with others and become expressive in dance performances. My investigation considers in particular the aesthetics of time in live performances in the theatre compared to dances created for the camera, focusing specifically on instances of contemporary transpositions from one context to the other
Video Time: Properties, Encoders and Evaluation
Time-aware encoding of frame sequences in a video is a fundamental problem in
video understanding. While many attempted to model time in videos, an explicit
study on quantifying video time is missing. To fill this lacuna, we aim to
evaluate video time explicitly. We describe three properties of video time,
namely a) temporal asymmetry, b)temporal continuity and c) temporal causality.
Based on each we formulate a task able to quantify the associated property.
This allows assessing the effectiveness of modern video encoders, like C3D and
LSTM, in their ability to model time. Our analysis provides insights about
existing encoders while also leading us to propose a new video time encoder,
which is better suited for the video time recognition tasks than C3D and LSTM.
We believe the proposed meta-analysis can provide a reasonable baseline to
assess video time encoders on equal grounds on a set of temporal-aware tasks.Comment: 14 pages, BMVC 201
Learning Human Motion Models for Long-term Predictions
We propose a new architecture for the learning of predictive spatio-temporal
motion models from data alone. Our approach, dubbed the Dropout Autoencoder
LSTM, is capable of synthesizing natural looking motion sequences over long
time horizons without catastrophic drift or motion degradation. The model
consists of two components, a 3-layer recurrent neural network to model
temporal aspects and a novel auto-encoder that is trained to implicitly recover
the spatial structure of the human skeleton via randomly removing information
about joints during training time. This Dropout Autoencoder (D-AE) is then used
to filter each predicted pose of the LSTM, reducing accumulation of error and
hence drift over time. Furthermore, we propose new evaluation protocols to
assess the quality of synthetic motion sequences even for which no ground truth
data exists. The proposed protocols can be used to assess generated sequences
of arbitrary length. Finally, we evaluate our proposed method on two of the
largest motion-capture datasets available to date and show that our model
outperforms the state-of-the-art on a variety of actions, including cyclic and
acyclic motion, and that it can produce natural looking sequences over longer
time horizons than previous methods
Non-linear Synthesis of Complex Laser Waveforms at Remote Distances
Strong deformation of ultrashort laser pulse shapes is unavoidable when
delivering high intensities at remote distances due to non-linear effects
taking place while propagating. Relying on the reversibility of laser
filamentation, we propose to explicitly design laser pulse shapes so that
propagation serves as a non-linear field synthesizer at a remote target
location. Such an approach allows, for instance, coherent control of molecules
at a remote distance, in the context of standoff detection of pathogens or
explosives.Comment: 17 pages, 6 figure
On human motion prediction using recurrent neural networks
Human motion modelling is a classical problem at the intersection of graphics
and computer vision, with applications spanning human-computer interaction,
motion synthesis, and motion prediction for virtual and augmented reality.
Following the success of deep learning methods in several computer vision
tasks, recent work has focused on using deep recurrent neural networks (RNNs)
to model human motion, with the goal of learning time-dependent representations
that perform tasks such as short-term motion prediction and long-term human
motion synthesis. We examine recent work, with a focus on the evaluation
methodologies commonly used in the literature, and show that, surprisingly,
state-of-the-art performance can be achieved by a simple baseline that does not
attempt to model motion at all. We investigate this result, and analyze recent
RNN methods by looking at the architectures, loss functions, and training
procedures used in state-of-the-art approaches. We propose three changes to the
standard RNN models typically used for human motion, which result in a simple
and scalable RNN architecture that obtains state-of-the-art performance on
human motion prediction.Comment: Accepted at CVPR 1
- âŚ