18,446 research outputs found
Adaptive Temporal Encoding Network for Video Instance-level Human Parsing
Beyond the existing single-person and multiple-person human parsing tasks in
static images, this paper makes the first attempt to investigate a more
realistic video instance-level human parsing that simultaneously segments out
each person instance and parses each instance into more fine-grained parts
(e.g., head, leg, dress). We introduce a novel Adaptive Temporal Encoding
Network (ATEN) that alternatively performs temporal encoding among key frames
and flow-guided feature propagation from other consecutive frames between two
key frames. Specifically, ATEN first incorporates a Parsing-RCNN to produce the
instance-level parsing result for each key frame, which integrates both the
global human parsing and instance-level human segmentation into a unified
model. To balance between accuracy and efficiency, the flow-guided feature
propagation is used to directly parse consecutive frames according to their
identified temporal consistency with key frames. On the other hand, ATEN
leverages the convolution gated recurrent units (convGRU) to exploit temporal
changes over a series of key frames, which are further used to facilitate the
frame-level instance-level parsing. By alternatively performing direct feature
propagation between consistent frames and temporal encoding network among key
frames, our ATEN achieves a good balance between frame-level accuracy and time
efficiency, which is a common crucial problem in video object segmentation
research. To demonstrate the superiority of our ATEN, extensive experiments are
conducted on the most popular video segmentation benchmark (DAVIS) and a newly
collected Video Instance-level Parsing (VIP) dataset, which is the first video
instance-level human parsing dataset comprised of 404 sequences and over 20k
frames with instance-level and pixel-wise annotations.Comment: To appear in ACM MM 2018. Code link:
https://github.com/HCPLab-SYSU/ATEN. Dataset link: http://sysu-hcp.net/li
Fast object detection in compressed JPEG Images
Object detection in still images has drawn a lot of attention over past few
years, and with the advent of Deep Learning impressive performances have been
achieved with numerous industrial applications. Most of these deep learning
models rely on RGB images to localize and identify objects in the image.
However in some application scenarii, images are compressed either for storage
savings or fast transmission. Therefore a time consuming image decompression
step is compulsory in order to apply the aforementioned deep models. To
alleviate this drawback, we propose a fast deep architecture for object
detection in JPEG images, one of the most widespread compression format. We
train a neural network to detect objects based on the blockwise DCT (discrete
cosine transform) coefficients {issued from} the JPEG compression algorithm. We
modify the well-known Single Shot multibox Detector (SSD) by replacing its
first layers with one convolutional layer dedicated to process the DCT inputs.
Experimental evaluations on PASCAL VOC and industrial dataset comprising images
of road traffic surveillance show that the model is about faster than
regular SSD with promising detection performances. To the best of our
knowledge, this paper is the first to address detection in compressed JPEG
images
Weakly-Supervised Action Segmentation with Iterative Soft Boundary Assignment
In this work, we address the task of weakly-supervised human action
segmentation in long, untrimmed videos. Recent methods have relied on expensive
learning models, such as Recurrent Neural Networks (RNN) and Hidden Markov
Models (HMM). However, these methods suffer from expensive computational cost,
thus are unable to be deployed in large scale. To overcome the limitations, the
keys to our design are efficiency and scalability. We propose a novel action
modeling framework, which consists of a new temporal convolutional network,
named Temporal Convolutional Feature Pyramid Network (TCFPN), for predicting
frame-wise action labels, and a novel training strategy for weakly-supervised
sequence modeling, named Iterative Soft Boundary Assignment (ISBA), to align
action sequences and update the network in an iterative fashion. The proposed
framework is evaluated on two benchmark datasets, Breakfast and Hollywood
Extended, with four different evaluation metrics. Extensive experimental
results show that our methods achieve competitive or superior performance to
state-of-the-art methods.Comment: CVPR 201
Low computational complexity variable block size (VBS) partitioning for motion estimation using the Walsh Hadamard transform (WHT)
Variable Block Size (VBS) based motion estimation has
been adapted in state of the art video coding, such as
H.264/AVC, VC-1. However, a low complexity H.264/AVC
encoder cannot take advantage of VBS due to its power consumption
requirements. In this paper, we present a VBS partition
algorithm based on a binary motion edge map without
either initial motion estimation or Rate-Distortion (R-D)
optimization for selecting modes. The proposed algorithm
uses the Walsh Hadamard Transform (WHT) to create a binary
edge map, which provides a computational complexity
cost effectiveness compared to other light segmentation
methods typically used to detect the required region
- …