2,525 research outputs found

    Fusion of Global and Local Motion Estimation Using Foreground Objects for Distributed Video Coding

    Get PDF
    International audienceThe side information in distributed video coding is estimated using the available decoded frames, and exploited for the decoding and reconstruction of other frames. The quality of the side information has a strong impact on the performance of distributed video coding. Here we propose a new approach that combines both global and local side information to improve coding performance. Since the background pixels in a frame are assigned to global estimation and the foreground objects to local estimation, one needs to estimate foreground objects in the side information using the backward and forward foreground objects, The background pixels are directly taken from the global side information. Specifically, elastic curves and local motion compensation are used to generate the foreground objects masks in the side information. Experimental results show that, as far as the rate-distortion performance is concerned, the proposed approach can achieve a PSNR improvement of up to 1.39 dB for a GOP size of 2, and up to 4.73 dB for larger GOP sizes, with respect to the reference DISCOVER codec. Index Terms A. ABOU-ELAILAH, F. DUFAUX, M. CAGNAZZO, and B. PESQUET-POPESCU are with the Signal and Image Processin

    A Fusion Framework for Camouflaged Moving Foreground Detection in the Wavelet Domain

    Full text link
    Detecting camouflaged moving foreground objects has been known to be difficult due to the similarity between the foreground objects and the background. Conventional methods cannot distinguish the foreground from background due to the small differences between them and thus suffer from under-detection of the camouflaged foreground objects. In this paper, we present a fusion framework to address this problem in the wavelet domain. We first show that the small differences in the image domain can be highlighted in certain wavelet bands. Then the likelihood of each wavelet coefficient being foreground is estimated by formulating foreground and background models for each wavelet band. The proposed framework effectively aggregates the likelihoods from different wavelet bands based on the characteristics of the wavelet transform. Experimental results demonstrated that the proposed method significantly outperformed existing methods in detecting camouflaged foreground objects. Specifically, the average F-measure for the proposed algorithm was 0.87, compared to 0.71 to 0.8 for the other state-of-the-art methods.Comment: 13 pages, accepted by IEEE TI

    Image sequence analysis for emerging interactive multimedia services - The European COST 211 framework

    Get PDF
    Cataloged from PDF version of article.Flexibility and efficiency of coding, content extraction, and content-based search are key research topics in the field of interactive multimedia. Ongoing ISO MPEG-4 and MPEG-7 activities are targeting standardization to facilitate such services. European COST Telecommunications activities provide a framework for research collaboration. COST 211bis and COST 211ter activities have been instrumental in the definition and development of the ITU-T H.261 and H.263 standards for videoconferencing over ISDN and videophony over regular phone lines, respectively. The group has also contributed significantly to the ISO MPEG-4 activities. At present a significant effort of the COST 211ter group activities is dedicated toward image and video sequence analysis and segmentation—an important technological aspect for the success of emerging object-based MPEG-4 and MPEG-7 multimedia applications. The current work of COST 211 is centered around the test model, called the Analysis Model (AM). The essential feature of the AM is its ability to fuse information from different sources to achieve a high-quality object segmentation. The current information sources are the intermediate results from frame-based (still) color segmentation, motion vector based segmentation, and changedetection-based segmentation. Motion vectors, which form the basis for the motion vector based intermediate segmentation, are estimated from consecutive frames. A recursive shortest spanning tree (RSST) algorithm is used to obtain intermediate color and motion vector based segmentation results. A rule-based region processor fuses the intermediate results; a postprocessor further refines the final segmentation output. The results of the current AM are satisfactory; it is expected that there will be further improvements of the AM within the COST 211 project

    A Low Power Architectural Framework for Automated Surveillance System with Low Bit Rate Transmission

    Get PDF
    Abstract The changed security scenario of the modern time has necessitated increased and sophisticated vigilance of the countries' borders. The technological challenges involved in accomplishing such feat of automated security system are many and require research at the components-and-algorithms as well as the architectural levels.  This paper proposes an architectural framework for automated video surveillance comprising a network of sensors and closed circuit television cameras as well as proposing algorithmic/component research of software and hardware for the core functioning of the framework, such as: communication protocols, object detection, data-integration, object identification, object tracking, video compression, threat identification, and alarm generation. In this paper, we are addressing some general topological and routing features that would be adopted in our system. There are two types of data with regard to data communication – video stream and object detection. The network is broken down into several disjoint, almost equal zones. A zone have one or more one cluster. A zone manager is chosen among the cluster heads depending on their relative residual energies. There are several levels of control that could be implemented with this arrangement with localized decision made, to get distributed effect at all levels. A cell tracks each target in its zone. If the target moves out of the range of a cell, the cell manager will send the target description to estimated next cell. The next cell starts tracking the target. If the estimated cell is wrongly chosen, corrections will be made by the cluster heads to get the new target-tracking. We also propose bitrate reduction algorithms to accommodate the limited bandwidth. One of the main feature of this paper is introducing a Low-Power Low-Bit rate video compression algorithm to accommodate the low power requirements at sensor nodes, and the low bit rate requirement for the communication protocol. We proposed two algorithms the ALBR and LPHSME. ALBR is addressing low bit rate required for sensors network with limited bandwidth which achieves a reduction in Average number of bits per Iframe by approximately 60% in case of low motion video sequences and 53% in case of fast motion video sequences . LPHSME addresses low power requirements of multi sensor network that has limited power batteries. The performance of the proposed LPHSME algorithm versus full search and three step search indicates  a reduction in motion estimation time by approximately 89% in case of low motion video sequences (e.g., Claire ) and 84% in case of fast motion video sequences. The reduced complexity of  LPHSME results in low power requirements

    Contextual cropping and scaling of TV productions

    Get PDF
    This is the author's accepted manuscript. The final publication is available at Springer via http://dx.doi.org/10.1007/s11042-011-0804-3. Copyright @ Springer Science+Business Media, LLC 2011.In this paper, an application is presented which automatically adapts SDTV (Standard Definition Television) sports productions to smaller displays through intelligent cropping and scaling. It crops regions of interest of sports productions based on a smart combination of production metadata and systematic video analysis methods. This approach allows a context-based composition of cropped images. It provides a differentiation between the original SD version of the production and the processed one adapted to the requirements for mobile TV. The system has been comprehensively evaluated by comparing the outcome of the proposed method with manually and statically cropped versions, as well as with non-cropped versions. Envisaged is the integration of the tool in post-production and live workflows

    Spatiotemporal Saliency Detection: State of Art

    Get PDF
    Saliency detection has become a very prominent subject for research in recent time. Many techniques has been defined for the saliency detection.In this paper number of techniques has been explained that include the saliency detection from the year 2000 to 2015, almost every technique has been included.all the methods are explained briefly including their advantages and disadvantages. Comparison between various techniques has been done. With the help of table which includes authors name,paper name,year,techniques,algorithms and challenges. A comparison between levels of acceptance rates and accuracy levels are made
    • …
    corecore