967 research outputs found
A study on Analysis and Utilization of Crowd-sourced Spatio-temporal Contexts from Social Media
兵庫県立大学大学院201
A study on Analysis and Utilization of Crowd-sourced Spatio-temporal Contexts from Social Media
兵庫県立大学大学院201
Video interaction using pen-based technology
Dissertação para obtenção do Grau de Doutor em
InformáticaVideo can be considered one of the most complete and complex media and its manipulating
is still a difficult and tedious task. This research applies pen-based technology to
video manipulation, with the goal to improve this interaction. Even though the human
familiarity with pen-based devices, how they can be used on video interaction, in order
to improve it, making it more natural and at the same time fostering the user’s creativity
is an open question.
Two types of interaction with video were considered in this work: video annotation
and video editing. Each interaction type allows the study of one of the interaction modes
of using pen-based technology: indirectly, through digital ink, or directly, trough pen
gestures or pressure. This research contributes with two approaches for pen-based video
interaction: pen-based video annotations and video as ink.
The first uses pen-based annotations combined with motion tracking algorithms, in
order to augment video content with sketches or handwritten notes. It aims to study how
pen-based technology can be used to annotate a moving objects and how to maintain the
association between a pen-based annotations and the annotated moving object
The second concept replaces digital ink by video content, studding how pen gestures
and pressure can be used on video editing and what kind of changes are needed in the
interface, in order to provide a more familiar and creative interaction in this usage context.This work was partially funded by the UTAustin-Portugal, Digital Media, Program
(Ph.D. grant: SFRH/BD/42662/2007 - FCT/MCTES); by the HP Technology for Teaching
Grant Initiative 2006; by the project "TKB - A Transmedia Knowledge Base for contemporary
dance" (PTDC/EAT/AVP/098220/2008 funded by FCT/MCTES); and by CITI/DI/FCT/UNL (PEst-OE/EEI/UI0527/2011
MediaSync: Handbook on Multimedia Synchronization
This book provides an approachable overview of the most recent advances in the fascinating field of media synchronization (mediasync), gathering contributions from the most representative and influential experts. Understanding the challenges of this field in the current multi-sensory, multi-device, and multi-protocol world is not an easy task. The book revisits the foundations of mediasync, including theoretical frameworks and models, highlights ongoing research efforts, like hybrid broadband broadcast (HBB) delivery and users' perception modeling (i.e., Quality of Experience or QoE), and paves the way for the future (e.g., towards the deployment of multi-sensory and ultra-realistic experiences). Although many advances around mediasync have been devised and deployed, this area of research is getting renewed attention to overcome remaining challenges in the next-generation (heterogeneous and ubiquitous) media ecosystem. Given the significant advances in this research area, its current relevance and the multiple disciplines it involves, the availability of a reference book on mediasync becomes necessary. This book fills the gap in this context. In particular, it addresses key aspects and reviews the most relevant contributions within the mediasync research space, from different perspectives. Mediasync: Handbook on Multimedia Synchronization is the perfect companion for scholars and practitioners that want to acquire strong knowledge about this research area, and also approach the challenges behind ensuring the best mediated experiences, by providing the adequate synchronization between the media elements that constitute these experiences
Multimodal Data Analysis of Dyadic Interactions for an Automated Feedback System Supporting Parent Implementation of Pivotal Response Treatment
abstract: Parents fulfill a pivotal role in early childhood development of social and communication
skills. In children with autism, the development of these skills can be delayed. Applied
behavioral analysis (ABA) techniques have been created to aid in skill acquisition.
Among these, pivotal response treatment (PRT) has been empirically shown to foster
improvements. Research into PRT implementation has also shown that parents can be
trained to be effective interventionists for their children. The current difficulty in PRT
training is how to disseminate training to parents who need it, and how to support and
motivate practitioners after training.
Evaluation of the parents’ fidelity to implementation is often undertaken using video
probes that depict the dyadic interaction occurring between the parent and the child during
PRT sessions. These videos are time consuming for clinicians to process, and often result
in only minimal feedback for the parents. Current trends in technology could be utilized to
alleviate the manual cost of extracting data from the videos, affording greater
opportunities for providing clinician created feedback as well as automated assessments.
The naturalistic context of the video probes along with the dependence on ubiquitous
recording devices creates a difficult scenario for classification tasks. The domain of the
PRT video probes can be expected to have high levels of both aleatory and epistemic
uncertainty. Addressing these challenges requires examination of the multimodal data
along with implementation and evaluation of classification algorithms. This is explored
through the use of a new dataset of PRT videos.
The relationship between the parent and the clinician is important. The clinician can
provide support and help build self-efficacy in addition to providing knowledge and
modeling of treatment procedures. Facilitating this relationship along with automated
feedback not only provides the opportunity to present expert feedback to the parent, but
also allows the clinician to aid in personalizing the classification models. By utilizing a
human-in-the-loop framework, clinicians can aid in addressing the uncertainty in the
classification models by providing additional labeled samples. This will allow the system
to improve classification and provides a person-centered approach to extracting
multimodal data from PRT video probes.Dissertation/ThesisDoctoral Dissertation Computer Science 201
People and object tracking for video annotation
Dissertação para obtenção do Grau de Mestre em
Engenharia InformáticaObject tracking is a thoroughly researched problem, with a body of associated literature
dating at least as far back as the late 1970s. However, and despite the development of some satisfactory real-time trackers, it has not yet seen widespread use. This is not due to a lack of applications for the technology, since several interesting ones exist. In this document, it is postulated that this status quo is due, at least in part, to a lack of easy to use software libraries supporting object tracking. An overview of the problems associated with object tracking is presented and the process of developing one such library is documented. This discussion includes how to overcome problems like heterogeneities in
object representations and requirements for training or initial object position hints.
Video annotation is the process of associating data with a video’s content. Associating data with a video has numerous applications, ranging from making large video archives or long videos searchable, to enabling discussion about and augmentation of the video’s content. Object tracking is presented as a valid approach to both automatic and manual video annotation, and the integration of the developed object tracking library into an existing video annotator, running on a tablet computer, is described. The challenges involved in designing an interface to support the association of video annotations with tracked objects in real-time are also discussed. In particular, we discuss our interaction approaches to handle moving object selection on live video, which we have called “Hold and Overlay” and “Hold and Speed Up”. In addition, the results of a set of preliminary tests are reported.project “TKB – A Transmedia Knowledge Base
for contemporary dance” (PTDC/EA /AVP/098220/2008 funded by FCT/MCTES), the
UTAustin – Portugal, Digital Media Program (SFRH/BD/42662/2007 FCT/MCTES) and by CITI/DI/FCT/UNL (Pest-OE/EEI/UI0527/2011
Towards using artificial intelligence as tool in artistic gymnastics coaching : case backward giant circle
The objective of this thesis was to study whether it is possible to create a system that estimates artistic gymnast’s body joint angles based on a low-budget 2-dimensional single RGB video recording. To meet the objective, 54 video files were collected on gymnasts performing backward giant circle skill, together with assessments of the performances by two professional coaches. The video files contained total of 233 repetitions of the skill. A pilot system of computer vision algorithms was developed, using an open source human body pose recognition algorithm. An algorithm based on pixel grayscale value was developed and used to recognize starting and ending moment of a repetition and to sample each repetition at 7 key phases. Body joint angle estimates were calculated based on the body part location estimates of the 1631 samples. The work proved that it is possible to develop a system that estimates body joint angles of an artistic gymnast. It was found that rotation and cropping of the frames improved probability of yielding correct estimates. The angle estimate for knees had highest, up to 66%, correlation with coach evaluations. Hips and shoulders had weak but significant correlation with coach evaluations. The results indicate that it is possible to develop a low-budget system that could work as augmented tool in artistic gymnastics coaching. In addition, human body pose recognition provides a new method to biomechanical research of artistic gymnastics
This Track Contains Politics: The Culture of Sampling in Experimental Electronica
What does it mean to process field recordings from the Ukrainian war in an electronic music track? How can the sampling of an Armenian keyboard melody be read as a critique of traditional gender roles? And what does it say about voyeurism in our culture when a techno producer uses viral YouTube videos as the basic material of his compositions?
Across five detailed case studies, Hannes Liechti discusses the culture and politics of musical sampling from a new perspective. Giving particular attention to the reasons behind sampling processes, Liechti’s in-depth analysis of sampling strategies by artists such as COOL FOR YOU and Lara Sarkissian shows that sampling political material, and sampling with political intentions reveals
a complex net of contexts, meanings, and often deeply personal choices and creative decisions.
Offering tangible tools and concepts for further exploration of sample-based music, the book illustrates the potential of popular music to tell stories about the world, and it describes the habits, thoughts, and realities of the laptop producer, one of the core actors in 21st century music-making
Proceedings experiencing light 2009 : international conference on the effects of light on welbeing
no abstrac
Gesture and Speech in Interaction - 4th edition (GESPIN 4)
International audienceThe fourth edition of Gesture and Speech in Interaction (GESPIN) was held in Nantes, France. With more than 40 papers, these proceedings show just what a flourishing field of enquiry gesture studies continues to be. The keynote speeches of the conference addressed three different aspects of multimodal interaction:gesture and grammar, gesture acquisition, and gesture and social interaction. In a talk entitled Qualitiesof event construal in speech and gesture: Aspect and tense, Alan Cienki presented an ongoing researchproject on narratives in French, German and Russian, a project that focuses especially on the verbal andgestural expression of grammatical tense and aspect in narratives in the three languages. Jean-MarcColletta's talk, entitled Gesture and Language Development: towards a unified theoretical framework,described the joint acquisition and development of speech and early conventional and representationalgestures. In Grammar, deixis, and multimodality between code-manifestation and code-integration or whyKendon's Continuum should be transformed into a gestural circle, Ellen Fricke proposed a revisitedgrammar of noun phrases that integrates gestures as part of the semiotic and typological codes of individuallanguages. From a pragmatic and cognitive perspective, Judith Holler explored the use ofgaze and hand gestures as means of organizing turns at talk as well as establishing common ground in apresentation entitled On the pragmatics of multi-modal face-to-face communication: Gesture, speech andgaze in the coordination of mental states and social interaction.Among the talks and posters presented at the conference, the vast majority of topics related, quitenaturally, to gesture and speech in interaction - understood both in terms of mapping of units in differentsemiotic modes and of the use of gesture and speech in social interaction. Several presentations explored the effects of impairments(such as diseases or the natural ageing process) on gesture and speech. The communicative relevance ofgesture and speech and audience-design in natural interactions, as well as in more controlled settings liketelevision debates and reports, was another topic addressed during the conference. Some participantsalso presented research on first and second language learning, while others discussed the relationshipbetween gesture and intonation. While most participants presented research on gesture and speech froman observer's perspective, be it in semiotics or pragmatics, some nevertheless focused on another importantaspect: the cognitive processes involved in language production and perception. Last but not least,participants also presented talks and posters on the computational analysis of gestures, whether involvingexternal devices (e.g. mocap, kinect) or concerning the use of specially-designed computer software forthe post-treatment of gestural data. Importantly, new links were made between semiotics and mocap data
- …