9,755 research outputs found
Dynamic similarity promotes interpersonal coordination in joint-action
Human movement has been studied for decades and dynamic laws of motion that
are common to all humans have been derived. Yet, every individual moves
differently from everyone else (faster/slower, harder/smoother etc). We propose
here an index of such variability, namely an individual motor signature (IMS)
able to capture the subtle differences in the way each of us moves. We show
that the IMS of a person is time-invariant and that it significantly differs
from those of other individuals. This allows us to quantify the dynamic
similarity, a measure of rapport between dynamics of different individuals'
movements, and demonstrate that it facilitates coordination during interaction.
We use our measure to confirm a key prediction of the theory of similarity that
coordination between two individuals performing a joint-action task is higher
if their motions share similar dynamic features. Furthermore, we use a virtual
avatar driven by an interactive cognitive architecture based on feedback
control theory to explore the effects of different kinematic features of the
avatar motion on the coordination with human players
Jointly Tracking and Separating Speech Sources Using Multiple Features and the generalized labeled multi-Bernoulli Framework
This paper proposes a novel joint multi-speaker tracking-and-separation
method based on the generalized labeled multi-Bernoulli (GLMB) multi-target
tracking filter, using sound mixtures recorded by microphones. Standard
multi-speaker tracking algorithms usually only track speaker locations, and
ambiguity occurs when speakers are spatially close. The proposed multi-feature
GLMB tracking filter treats the set of vectors of associated speaker features
(location, pitch and sound) as the multi-target multi-feature observation,
characterizes transitioning features with corresponding transition models and
overall likelihood function, thus jointly tracks and separates each
multi-feature speaker, and addresses the spatial ambiguity problem. Numerical
evaluation verifies that the proposed method can correctly track locations of
multiple speakers and meanwhile separate speech signals
A biologically plausible system for detecting saliency in video
Neuroscientists and cognitive scientists credit the dorsal and ventral pathways for the capability of detecting both still salient and motion salient objects. In this work, a framework is developed to explore potential models of still and motion saliency and is an extension of the original VENUS system. The early visual pathway is modeled by using Independent Component Analysis to learn a set of Gabor-like receptive fields similar to those found in the mammalian visual pathway. These spatial receptive fields form a set of 2D basis feature matrices, which are used to decompose complex visual stimuli into their spatial components. A still saliency map is formed by combining the outputs of convoluting the learned spatial receptive fields with the input stimuli. The dorsal pathway is primarily focused on motion-based information. In this framework, the model uses simple motion segmentation and tracking algorithms to create a statistical model of the motion and color-related information in video streams. A key feature of the human visual system is the ability to detect novelty. This framework uses a set of Gaussian distributions to model color and motion. When a unique event is detected, Gaussian distributions are created and the event is declared novel. The next time a similar event is detected the framework is able to determine that the event is not novel based on the previously created distributions. A forgetting term is also included that allows events that have not been detected for a long period of time to be forgotten
A single-chip FPGA implementation of real-time adaptive background model
This paper demonstrates the use of a single-chip
FPGA for the extraction of highly accurate background
models in real-time. The models are based
on 24-bit RGB values and 8-bit grayscale intensity
values. Three background models are presented, all
using a camcorder, single FPGA chip, four blocks
of RAM and a display unit. The architectures have
been implemented and tested using a Panasonic NVDS60B
digital video camera connected to a Celoxica
RC300 Prototyping Platform with a Xilinx Virtex
II XC2v6000 FPGA and 4 banks of onboard RAM.
The novel FPGA architecture presented has the advantages
of minimizing latency and the movement of
large datasets, by conducting time critical processes
on BlockRAM. The systems operate at clock rates
ranging from 57MHz to 65MHz and are capable
of performing pre-processing functions like temporal
low-pass filtering on standard frame size of 640X480
pixels at up to 210 frames per second
CVABS: Moving Object Segmentation with Common Vector Approach for Videos
Background modelling is a fundamental step for several real-time computer
vision applications that requires security systems and monitoring. An accurate
background model helps detecting activity of moving objects in the video. In
this work, we have developed a new subspace based background modelling
algorithm using the concept of Common Vector Approach with Gram-Schmidt
orthogonalization. Once the background model that involves the common
characteristic of different views corresponding to the same scene is acquired,
a smart foreground detection and background updating procedure is applied based
on dynamic control parameters. A variety of experiments is conducted on
different problem types related to dynamic backgrounds. Several types of
metrics are utilized as objective measures and the obtained visual results are
judged subjectively. It was observed that the proposed method stands
successfully for all problem types reported on CDNet2014 dataset by updating
the background frames with a self-learning feedback mechanism.Comment: 12 Pages, 4 Figures, 1 Tabl
- …