31,510 research outputs found
Object-driven block-based algorithm for the compression of stereo image pairs
In this paper, we propose a novel object driven, block based algorithm for the compression of stereo image pairs. The
algorithm effectively combines the simplicity and adaptability of the existing block based stereo image compression
techniques [1-6] with an edge/contour based object extraction technique to determine appropriate compression strategy for
various areas of the right image. Extensive experiments carried out support that significant improvements of up to 20% in
compression ratio can be achieved by the proposed algorithm, compared with the existing stereo image compression
techniques. Yet the reconstructed image quality is maintained at an equivalent level in terms of PSNR values. In terms of
visual quality, the right image reconstructed by the proposed algorithm does not incur any noticeable effect compared with
the outputs of the best algorithms.
The proposed algorithm performs object extraction and matching between the reconstructed left frame and the original right
frame to identify those objects that match but are displaced by varying amounts due to binocular parallax. Different coding
strategies are then applied separately to internal areas and the bounding areas for each identified object. Based on the mean
squared matching error of the internal blocks and a selected threshold, a decision is made whether or not to encode the
predictive errors inside these objects. The output bit stream includes entropy coding of object disparity, block disparity and
possibly some errors, which fail to meet the threshold requirement in the proposed algorith
Dynamic Domain Classification for Fractal Image Compression
Fractal image compression is attractive except for its high encoding time
requirements. The image is encoded as a set of contractive affine
transformations. The image is partitioned into non-overlapping range blocks,
and a best matching domain block larger than the range block is identified.
There are many attempts on improving the encoding time by reducing the size of
search pool for range-domain matching. But these methods are attempting to
prepare a static domain pool that remains unchanged throughout the encoding
process. This paper proposes dynamic preparation of separate domain pool for
each range block. This will result in significant reduction in the encoding
time. The domain pool for a particular range block can be selected based upon a
parametric value. Here we use classification based on local fractal dimension.Comment: 8 pages, 4 tables, 1 figur
Block Based Motion Vector Estimation Using FUHS16, UHDS16 and UHDS8 Algorithms for Video Sequence
Block-matching algorithm is the most common technique applied in block-based motion estimation technique. There are several block-matching algorithm based on block-based motion estimation techniques have been developed. Full search (FS), three step search (TSS), new three step search (NTSS), diamond search (DS) and hexagon based search (HS) are the most well known block-matching algorithm. These techniques are applied to video sequences to remove the temporal redundancy for compression purposes and to gauge the motion vector estimation. In addition, the mentioned block-matching algorithms are the baseline techniques that have been used to further develop all the enhanced or improved algorithms.
In order to develop the proposed methods, the baseline techniques are studied to develop the proposed algorithms. This chapter proposes modelling of fast unrestricted hexagon search (FUHS16) and unrestricted hexagon-diamond search (UHDS16) algorithms for motion vector estimation, which is based on the theory and application of block-based motion estimation. Both of these algorithms are designed using 16 Ă— 16 block size. In particular, the motion vector estimation, quality performance, computational complexity, and elapsed processing time are emphasised. These parameters have been used to measure the experimental results.
It is the aim of this study that this work provides a common framework with which to evaluate and understand block-based matching motion estimation performance. On the theoretical side, four fundamental issues are explored: (1) division of frame, (2) basic block-based matching, (3) motion vector estimation, and (4) block-matching algorithm development. Various existing block-matching motion estimation algorithms have been analysed to develop the fundamental research.
Based on the theoretical and fundamental research analysis the FUHS16 and UHDS16 algorithms using 16 Ă— 16 block-based motion estimation formulations were developed. To improve the UHDS16 algorithm, 8 Ă— 8 block-matching technique has been tested. The 8 Ă— 8 block-matching technique is known as UHDS8. The results show positive improvements. From an application perspective, the UHDS8 algorithm efficiently captured the motion vectors in many video sequences. For example, in video compression, the use of motion vectors on individual macro-blocks optimized the motion vector information. The UHDS8 algorithm also offers improvement in terms of image quality performance, computational complexity and elapsed processing time. Thus, this chapter offers contributions in certain areas such as reducing the mechanism of computational complexity in estimating the motion from the video sequences. In particular, the FUHS16, UHDS16 and UHDS8 algorithms were developed to estimate the motion vectors field in the video sequences. Theoretical analysis block-based matching criteria are adapted to FUHS16, UHDS16 and UHDS8 algorithms, which are based on search points technique. Basically, the proposed of FUHS16, UHDS16 and UHDS8 algorithm produces the best motion vector estimation finding based on the block-based matching criteria. Besides that, the UHDS8 algorithm also improves the image quality performances and the search points in terms of the computational complexity. Overall, the study shows that the UHDS8 algorithm produces better results compared to the FUHS16 and UHDS16 algorithm
AN EFFICIENT MOTION ESTIMATION ALGORITHM BASED ON PARTICLE SWARM OPTIMIZATION
The PSO algorithm reduce the search points without the degradation of the image quality. It provides accurate motion estimation with very low complexity in the context of video estimation. This algorithm is capable of reducing the computational complexity of block matching process. This algorithm maintains high estimation accuracy compared to the full search method. The critical component in most block-based video compression system is Motion Estimation because redundancy between successive frames of video sequence allows for compression of video data. These algorithms are used to reduce the computational requirement by checking only some points inside the search window, while keeping a good error performance when compared with Full Search and Diamond search algorithm. This algorithm should maintain high estimation accuracy compared to the Full search method and Diamond search algorithm. Here by using the PSO algorithm could get a high accuracy in the block-based motion estimation
The Use of Quadtree Range Domain Partitioning with Fast Double Moment Descriptors to Enhance FIC of Colored Image
In this paper, an enhanced fractal image compression system (FIC) is proposed; it is based on using both symmetry prediction and blocks indexing to speed up the blocks matching process. The proposed FIC uses quad tree as variable range block partitioning mechanism. two criteria’s for guiding the partitioning decision are used: The first one uses sobel-based edge magnitude, whereas the second uses the contrast of block. A new set of moment descriptors are introduced, they differ from the previously used descriptors by their ability to emphasize the weights of different parts of each block. The effectiveness of all possible combinations of double moments descriptors has been investigated. Furthermore, a fast computation mechanism is introduced to compute the moments attended to improve the overall computation cost. the results of applied tests on the system for the cases “variable and fixed range” block partitioning mechanism indicated that the variable partitioning scheme can produce better results than fixed partitioning one (that is, 4 × 4 block) in term of compression ratio, faster than and PSNR does not significantly decreased
A Perceptual Based Motion Compensation Technique for Video Coding
Motion estimation is one of the important procedures in the all video
encoders. Most of the complexity of the video coder depends on the complexity
of the motion estimation step. The original motion estimation algorithm has a
remarkable complexity and therefore many improvements were proposed to enhance
the crude version of the motion estimation. The basic idea of many of these
works were to optimize some distortion function for mean squared error (MSE) or
sum of absolute difference (SAD) in block matching But it is shown that these
metrics do not conclude the quality as it is, on the other hand, they are not
compatible with the human visual system (HVS). In this paper we explored the
usage of the image quality metrics in the video coding and more specific in the
motion estimation. We have utilized the perceptual image quality metrics
instead of MSE or SAD in the block based motion estimation. Three different
metrics have used: structural similarity or SSIM, complex wavelet structural
similarity or CW-SSIM, visual information fidelity or VIF. Experimental results
showed that usage of the quality criterions can improve the compression rate
while the quality remains fix and thus better quality in coded video at the
same bit budget
An Efficient Human Visual System Based Quality Metric for 3D Video
Stereoscopic video technologies have been introduced to the consumer market
in the past few years. A key factor in designing a 3D system is to understand
how different visual cues and distortions affect the perceptual quality of
stereoscopic video. The ultimate way to assess 3D video quality is through
subjective tests. However, subjective evaluation is time consuming, expensive,
and in some cases not possible. The other solution is developing objective
quality metrics, which attempt to model the Human Visual System (HVS) in order
to assess perceptual quality. Although several 2D quality metrics have been
proposed for still images and videos, in the case of 3D efforts are only at the
initial stages. In this paper, we propose a new full-reference quality metric
for 3D content. Our method mimics HVS by fusing information of both the left
and right views to construct the cyclopean view, as well as taking to account
the sensitivity of HVS to contrast and the disparity of the views. In addition,
a temporal pooling strategy is utilized to address the effect of temporal
variations of the quality in the video. Performance evaluations showed that our
3D quality metric quantifies quality degradation caused by several
representative types of distortions very accurately, with Pearson correlation
coefficient of 90.8 %, a competitive performance compared to the
state-of-the-art 3D quality metrics
PRNU-Based Source Device Attribution for YouTube Videos
Photo Response Non-Uniformity (PRNU) is a camera imaging sensor imperfection
which has earned a great interest for source device attribution of digital
videos. A majority of recent researches about PRNU-based source device
attribution for digital videos do not take into consideration the effects of
video compression on the PRNU noise in video frames, but rather consider video
frames as isolated images of equal importance. As a result, these methods
perform poorly on re-compressed or low bit-rate videos. This paper proposes a
novel method for PRNU fingerprint estimation from video frames taking into
account the effects of video compression on the PRNU noise in these frames.
With this method, we aim to determine whether two videos from unknown sources
originate from the same device or not. Experimental results on a large set of
videos show that the method we propose is more effective than existing
frame-based methods that use either only I frames or all (I-B-P) frames,
especially on YouTube videos.Comment: Revised (and accepted) version of the original submission. Minor
changes have been brought to the original manuscrip
Pyramid Attention Networks for Image Restoration
Self-similarity refers to the image prior widely used in image restoration
algorithms that small but similar patterns tend to occur at different locations
and scales. However, recent advanced deep convolutional neural network based
methods for image restoration do not take full advantage of self-similarities
by relying on self-attention neural modules that only process information at
the same scale. To solve this problem, we present a novel Pyramid Attention
module for image restoration, which captures long-range feature correspondences
from a multi-scale feature pyramid. Inspired by the fact that corruptions, such
as noise or compression artifacts, drop drastically at coarser image scales,
our attention module is designed to be able to borrow clean signals from their
"clean" correspondences at the coarser levels. The proposed pyramid attention
module is a generic building block that can be flexibly integrated into various
neural architectures. Its effectiveness is validated through extensive
experiments on multiple image restoration tasks: image denoising, demosaicing,
compression artifact reduction, and super resolution. Without any bells and
whistles, our PANet (pyramid attention module with simple network backbones)
can produce state-of-the-art results with superior accuracy and visual quality.
Our code will be available at
https://github.com/SHI-Labs/Pyramid-Attention-Network
- …