35 research outputs found

    Skeleton-aided Articulated Motion Generation

    Full text link
    This work make the first attempt to generate articulated human motion sequence from a single image. On the one hand, we utilize paired inputs including human skeleton information as motion embedding and a single human image as appearance reference, to generate novel motion frames, based on the conditional GAN infrastructure. On the other hand, a triplet loss is employed to pursue appearance-smoothness between consecutive frames. As the proposed framework is capable of jointly exploiting the image appearance space and articulated/kinematic motion space, it generates realistic articulated motion sequence, in contrast to most previous video generation methods which yield blurred motion effects. We test our model on two human action datasets including KTH and Human3.6M, and the proposed framework generates very promising results on both datasets.Comment: ACM MM 201

    Visual tracking for sports applications

    Get PDF
    Visual tracking of the human body has attracted increasing attention due to the potential to perform high volume low cost analyses of motions in a wide range of applications, including sports training, rehabilitation and security. In this paper we present the development of a visual tracking module for a system aimed to be used as an autonomous instructional aid for amateur golfers. Postural information is captured visually and fused with information from a golf swing analyser mat and both visual and audio feedback given based on the golfer's mistakes. Results from the visual tracking module are presented

    Análise de movimento humano por visão computacional: uma síntese

    Get PDF
    O movimento humano é complexo, não linear e varia com o tempo. Nos últimos tempos, inúmeros investigadores têm-se dedicado ao desenvolvimento de sistemas automáticos capazes de realizar o seguimento, a análise e o reconhecimento deste tipo de movimento, utilizando técnicas de Visão Computacional. Neste artigo, serão resumidamente enumeradas e descritas algumas das técnicas actualmente empregues neste domínio

    Regression-Based Human Motion Capture From Voxel Data

    Full text link

    Utilizando visão computacional para reconstrução probabilística 3D e rastreamento de movimento

    Get PDF
    This paper presents an approach to the 3D visual tracking problem into multi-camera environments. This proposal executes the markerless visual tracking observing the environment through a model based in a volumetric reconstruction technique, called 3D Probabilistic Occupancy Grid, which is still seldom used for this purpose. The target is tracked by the use of Expectation-Maximization algorithm with an object representation model constructed with Gaussians blobs representing the object body parts.Este artigo apresenta um método não intrusivo para rastreamento de movimento 3D em ambientes monitorados por múltiplas câmeras. Primeiramente, se obtém uma reconstrução volumétrica 3D, através da técnica de Grid de Ocupação Probabilístico, tal técnica ainda foi pouco explorada no contexto de rastreamento de movimento. Então, utiliza-se o algoritmo Expectation-Maximization em conjunto com um modelo de representação do corpo do objeto de interesse baseado em blobs Gaussianas, para identificar e rastrear o movimento das partes do corpo do objeto de interesse

    Vision-Based Autonomous Human Tracking Mobile Robot

    Get PDF
    Tracking moving objects is one of the most important but problematic features of motion analysis and understanding. In order to effectively interact robots with people in close proximity, the systems must first be able to detect, track, and follow people. Following a human with a mobile robot arises in many different service robotic applications. This paper proposes to build an autonomous human tracking mobile robot which can solve the occlusion problem during tracking. The robot can make human tracking efficiently by analysing the information obtained from a camera which is equipped on the top of the robot. The system performs human detection by using Histogram of Oriented Gradient (HOG) and Support Vector Machine (SVM) algorithms and then uses HSV (Hue Saturation Value) color system for detecting stranger. If the detected human is stranger, robot will make tracking. During the process, the robot needs to track the stranger without missing. So, Kalman filter is used to solve this problem. Kalman filter can estimate the target human when the human is occluded with walls or something. This paper describes the processing results and experimental results of a mobile robot which will track unmarked human efficiently and handle the occlusion using vision sensor and Kalman filter

    Models and estimators for markerless human motion tracking

    Get PDF
    In this work, we analyze the diferent components of a model-based motion tracking system. The system consists in: a human body model, an estimator, and a likelihood or cost function

    Motion capture based on RGBD data from multiple sensors for avatar animation

    Get PDF
    With recent advances in technology and emergence of affordable RGB-D sensors for a wider range of users, markerless motion capture has become an active field of research both in computer vision and computer graphics. In this thesis, we designed a POC (Proof of Concept) for a new tool that enables us to perform motion capture by using a variable number of commodity RGB-D sensors of different brands and technical specifications on constraint-less layout environments. The main goal of this work is to provide a tool with motion capture capabilities by using a handful of RGB-D sensors, without imposing strong requirements in terms of lighting, background or extension of the motion capture area. Of course, the number of RGB-D sensors needed is inversely proportional to their resolution, and directly proportional to the size of the area to track to. Built on top of the OpenNI 2 library, we made this POC compatible with most of the nonhigh-end RGB-D sensors currently available in the market. Due to the lack of resources on a single computer, in order to support more than a couple of sensors working simultaneously, we need a setup composed of multiple computers. In order to keep data coherency and synchronization across sensors and computers, our tool makes use of a semi-automatic calibration method and a message-oriented network protocol. From color and depth data given by a sensor, we can also obtain a 3D pointcloud representation of the environment. By combining pointclouds from multiple sensors, we can collect a complete and animated 3D pointcloud that can be visualized from any viewpoint. Given a 3D avatar model and its corresponding attached skeleton, we can use an iterative optimization method (e.g. Simplex) to find a fit between each pointcloud frame and a skeleton configuration, resulting in 3D avatar animation when using such skeleton configurations as key frames
    corecore