35,672 research outputs found
Regularity scalable image coding based on wavelet singularity detection
In this paper, we propose an adaptive algorithm for scalable wavelet image coding, which is based on the general feature, the regularity, of images. In pattern recognition or computer vision, regularity of images is estimated from the oriented wavelet coefficients and quantified by the Lipschitz exponents. To estimate the Lipschitz exponents, evaluating the interscale evolution of the wavelet transform modulus sum (WTMS) over the directional cone of influence was proven to be a better approach than tracing the wavelet transform modulus maxima (WTMM). This is because the irregular sampling nature of the WTMM complicates the reconstruction process. Moreover, examples were found to show that the WTMM representation cannot uniquely characterize a signal. It implies that the reconstruction of signal from its WTMM may not be consistently stable. Furthermore, the WTMM approach requires much more computational effort. Therefore, we use the WTMS approach to estimate the regularity of images from the separable wavelet transformed coefficients. Since we do not concern about the localization issue, we allow the decimation to occur when we evaluate the interscale evolution. After the regularity is estimated, this information is utilized in our proposed adaptive regularity scalable wavelet image coding algorithm. This algorithm can be simply embedded into any wavelet image coders, so it is compatible with the existing scalable coding techniques, such as the resolution scalable and signal-to-noise ratio (SNR) scalable coding techniques, without changing the bitstream format, but provides more scalable levels with higher peak signal-to-noise ratios (PSNRs) and lower bit rates. In comparison to the other feature-based wavelet scalable coding algorithms, the proposed algorithm outperforms them in terms of visual perception, computational complexity and coding efficienc
Factor versus palindromic complexity of uniformly recurrent infinite words
We study the relation between the palindromic and factor complexity of
infinite words. We show that for uniformly recurrent words one has P(n)+P(n+1)
\leq \Delta C(n) + 2, for all n \in N. For a large class of words it is a
better estimate of the palindromic complexity in terms of the factor complexity
then the one presented by Allouche et al. We provide several examples of
infinite words for which our estimate reaches its upper bound. In particular,
we derive an explicit prescription for the palindromic complexity of infinite
words coding r-interval exchange transformations. If the permutation \pi
connected with the transformation is given by \pi(k)=r+1-k for all k, then
there is exactly one palindrome of every even length, and exactly r palindromes
of every odd length.Comment: 16 pages, submitted to Theoretical Computer Scienc
A New Fast Motion Estimation and Mode Decision algorithm for H.264 Depth Maps encoding in Free Viewpoint TV
In this paper, we consider a scenario where 3D scenes are modeled through a View+Depth representation. This representation is to be used at the rendering side to generate synthetic views for free viewpoint video. The encoding of both type of data (view and depth) is carried out using two H.264/AVC encoders. In this scenario we address the reduction of the encoding complexity of depth data. Firstly, an analysis of the Mode Decision and Motion Estimation processes has been conducted for both view and depth sequences, in order to capture the correlation between them. Taking advantage of this correlation, we propose a fast mode decision and motion estimation algorithm for the depth encoding. Results show that the proposed algorithm reduces the computational burden with a negligible loss in terms of quality of the rendered synthetic views. Quality measurements have been conducted using the Video Quality Metric
Interaction between high-level and low-level image analysis for semantic video object extraction
Authors of articles published in EURASIP Journal on Advances in Signal Processing are the copyright holders of their articles and have granted to any third party, in advance and in perpetuity, the right to use, reproduce or disseminate the article, according to the SpringerOpen copyright and license agreement (http://www.springeropen.com/authors/license)
A Matlab-Based Tool for Video Quality Evaluation without Reference
This paper deals with the design of a Matlab based tool for measuring video quality with no use of a reference sequence. The main goals are described and the tool and its features are shown. The paper begins with a description of the existing pixel-based no-reference quality metrics. Then, a novel algorithm for simple PSNR estimation of H.264/AVC coded videos is presented as an alternative. The algorithm was designed and tested using publicly available video database of H.264/AVC coded videos. Cross-validation was used to confirm the consistency of results
Recommended from our members
A biologically inspired spiking model of visual processing for image feature detection
To enable fast reliable feature matching or tracking in scenes, features need to be discrete and meaningful, and hence edge or corner features, commonly called interest points are often used for this purpose. Experimental research has illustrated that biological vision systems use neuronal circuits to extract particular features such as edges or corners from visual scenes. Inspired by this biological behaviour, this paper proposes a biologically inspired spiking neural network for the purpose of image feature extraction. Standard digital images are processed and converted to spikes in a manner similar to the processing that transforms light into spikes in the retina. Using a hierarchical spiking network, various types of biologically inspired receptive fields are used to extract progressively complex image features. The performance of the network is assessed by examining the repeatability of extracted features with visual results presented using both synthetic and real images
Static/Dynamic Filtering for Mesh Geometry
The joint bilateral filter, which enables feature-preserving signal smoothing
according to the structural information from a guidance, has been applied for
various tasks in geometry processing. Existing methods either rely on a static
guidance that may be inconsistent with the input and lead to unsatisfactory
results, or a dynamic guidance that is automatically updated but sensitive to
noises and outliers. Inspired by recent advances in image filtering, we propose
a new geometry filtering technique called static/dynamic filter, which utilizes
both static and dynamic guidances to achieve state-of-the-art results. The
proposed filter is based on a nonlinear optimization that enforces smoothness
of the signal while preserving variations that correspond to features of
certain scales. We develop an efficient iterative solver for the problem, which
unifies existing filters that are based on static or dynamic guidances. The
filter can be applied to mesh face normals followed by vertex position update,
to achieve scale-aware and feature-preserving filtering of mesh geometry. It
also works well for other types of signals defined on mesh surfaces, such as
texture colors. Extensive experimental results demonstrate the effectiveness of
the proposed filter for various geometry processing applications such as mesh
denoising, geometry feature enhancement, and texture color filtering
The probabilistic nature of McShane's identity: planar tree coding of simple loops
In this article, we discuss a probabilistic interpretation of McShane's
identity as describing a finite measure on the space of embedded paths though a
point.Comment: 25 page
Low computational complexity variable block size (VBS) partitioning for motion estimation using the Walsh Hadamard transform (WHT)
Variable Block Size (VBS) based motion estimation has
been adapted in state of the art video coding, such as
H.264/AVC, VC-1. However, a low complexity H.264/AVC
encoder cannot take advantage of VBS due to its power consumption
requirements. In this paper, we present a VBS partition
algorithm based on a binary motion edge map without
either initial motion estimation or Rate-Distortion (R-D)
optimization for selecting modes. The proposed algorithm
uses the Walsh Hadamard Transform (WHT) to create a binary
edge map, which provides a computational complexity
cost effectiveness compared to other light segmentation
methods typically used to detect the required region
- …