13,669 research outputs found
MASCOT : metadata for advanced scalable video coding tools : final report
The goal of the MASCOT project was to develop new video coding schemes and tools that provide both an increased coding efficiency as well as extended scalability features compared to technology that was available at the beginning of the project. Towards that goal the following tools would be used: - metadata-based coding tools; - new spatiotemporal decompositions; - new prediction schemes. Although the initial goal was to develop one single codec architecture that was able to combine all new coding tools that were foreseen when the project was formulated, it became clear that this would limit the selection of the new tools. Therefore the consortium decided to develop two codec frameworks within the project, a standard hybrid DCT-based codec and a 3D wavelet-based codec, which together are able to accommodate all tools developed during the course of the project
Survey of image-based representations and compression techniques
In this paper, we survey the techniques for image-based rendering (IBR) and for compressing image-based representations. Unlike traditional three-dimensional (3-D) computer graphics, in which 3-D geometry of the scene is known, IBR techniques render novel views directly from input images. IBR techniques can be classified into three categories according to how much geometric information is used: rendering without geometry, rendering with implicit geometry (i.e., correspondence), and rendering with explicit geometry (either with approximate or accurate geometry). We discuss the characteristics of these categories and their representative techniques. IBR techniques demonstrate a surprising diverse range in their extent of use of images and geometry in representing 3-D scenes. We explore the issues in trading off the use of images and geometry by revisiting plenoptic-sampling analysis and the notions of view dependency and geometric proxies. Finally, we highlight compression techniques specifically designed for image-based representations. Such compression techniques are important in making IBR techniques practical.published_or_final_versio
Anomaly Detection, Rule Adaptation and Rule Induction Methodologies in the Context of Automated Sports Video Annotation.
Automated video annotation is a topic of considerable interest in computer vision due to its applications in video search, object based video encoding and enhanced broadcast content. The domain of sport broadcasting is, in particular, the subject of current research attention due to its fixed, rule governed, content. This research work aims to develop, analyze and demonstrate novel methodologies that can be useful in the context of adaptive and automated video annotation systems. In this thesis, we present methodologies for addressing the problems of anomaly detection, rule adaptation and rule induction for court based sports such as tennis and badminton. We first introduce an HMM induction strategy for a court-model based method that uses the court structure in the form of a lattice for two related modalities of singles and doubles tennis to tackle the problems of anomaly detection and rectification. We also introduce another anomaly detection methodology that is based on the disparity between the low-level vision based classifiers and the high-level contextual classifier. Another approach to address the problem of rule adaptation is also proposed that employs Convex hulling of the anomalous states. We also investigate a number of novel hierarchical HMM generating methods for stochastic induction of game rules. These methodologies include, Cartesian product Label-based Hierarchical Bottom-up Clustering (CLHBC) that employs prior information within the label structures. A new constrained variant of the classical Chinese Restaurant Process (CRP) is also introduced that is relevant to sports games. We also propose two hybrid methodologies in this context and a comparative analysis is made against the flat Markov model. We also show that these methods are also generalizable to other rule based environments
Recommended from our members
James and Whitehead: Assemblage and Systematization of a Deeply Empiricist Mosaic Philosophy
This paper contributes to a growing body of philosophical and psychological work that draws parallels between the writings of William James and Alfred North Whitehead. In Part One I introduce Whitehead’s distinction between assemblage and systematization (section 1) and suggest that Whitehead’s philosophy was in part a systematization of James’ psychological and philosophical assemblage (section 2). The systematization is based on a rethinking of the entity/function contrast (section 3) by way of Whitehead’s concept of the actual entity/occasion (section 4). This permits a process-oriented ontological extension and James’ notion of pure experience (sections 5 & 6), which yields a deepened version of radical empiricism (section 7). The four sections of Part Two build a more specific argument that James’ often implicit distinctions between energetic, perceptual, conceptual and discursive modes of experience can be systematized by way of Whitehead’s concepts of causal efficacy, presentational immediacy and symbolic reference. Following the suggestion of Magritte’s famous Ceci n’est pas une Pipe artwork, this yields an analysis of the sum of human experience into four progressively integrated factors: power, image, proposition and enunciation
Variable Resolution & Dimensional Mapping For 3d Model Optimization
Three-dimensional computer models, especially geospatial architectural data sets, can be visualized in the same way humans experience the world, providing a realistic, interactive experience. Scene familiarization, architectural analysis, scientific visualization, and many other applications would benefit from finely detailed, high resolution, 3D models. Automated methods to construct these 3D models traditionally has produced data sets that are often low fidelity or inaccurate; otherwise, they are initially highly detailed, but are very labor and time intensive to construct. Such data sets are often not practical for common real-time usage and are not easily updated. This thesis proposes Variable Resolution & Dimensional Mapping (VRDM), a methodology that has been developed to address some of the limitations of existing approaches to model construction from images. Key components of VRDM are texture palettes, which enable variable and ultra-high resolution images to be easily composited; texture features, which allow image features to integrated as image or geometry, and have the ability to modify the geometric model structure to add detail. These components support a primary VRDM objective of facilitating model refinement with additional data. This can be done until the desired fidelity is achieved as practical limits of infinite detail are approached. Texture Levels, the third component, enable real-time interaction with a very detailed model, along with the flexibility of having alternate pixel data for a given area of the model and this is achieved through extra dimensions. Together these techniques have been used to construct models that can contain GBs of imagery data
Managing hybrid methods for integration and combination of data
This chapter looks at how monitoring can combine data from multiple sources, including basic observations coupled with auxiliary information and the use of reference data for classification and modelling. A vital component of monitoring research is to be able to combine and synthesize data in a systematic, transparent way that can integrate social and environmental factors and show how these reflect, overlap with, correlate to, and influence each other. Data types and relevant analytical methods are briefly discussed, as well as aspects of classification and semantics, showing best practice in analysis and some suitable methods for describing data properties such as data quality. Typical problems of incompleteness, lack of fit to semantic classes, thematic and geometric inaccuracy, and data redundancy are discussed, with a range of examples showing how these challenges can be met by identifying and filling gaps in datasets
- …