56 research outputs found

    Accurate and discernible photocollages

    Get PDF
    There currently exist several techniques for selecting and combining images from a digital image library into a single image so that the result meets certain prespecified visual criteria. Image mosaic methods, first explored by Connors and Trivedi[18], arrange library images according to some tiling arrangement, often a regular grid, so that the combination of images, when viewed as a whole, resembles some input target image. Other techniques, such as Autocollage of Rother et al.[78], seek only to combine images in an interesting and visually pleasing manner, according to certain composition principles, without attempting to approximate any target image. Each of these techniques provide a myriad of creative options for artists who wish to combine several levels of meaning into a single image or who wish to exploit the meaning and symbolism contained in each of a large set of images through an efficient and easy process. We first examine the most notable and successful of these methods, and summarize the advantages and limitations of each. We then formulate a set of goals for an image collage system that combines the advantages of these methods while addressing and mitigating the drawbacks. Particularly, we propose a system for creating photocollages that approximate a target image as an aggregation of smaller images, chosen from a large library, so that interesting visual correspondences between images are exploited. In this way, we allow users to create collages in which multiple layers of meaning are encoded, with meaningful visual links between each layer. In service of this goal, we ensure that the images used are as large as possible and are combined in such a way that boundaries between images are not immediately apparent, as in Autocollage. This has required us to apply a multiscale approach to searching and comparing images from a large database, which achieves both speed and accuracy. We also propose a new framework for color post-processing, and propose novel techniques for decomposing images according to object and texture information

    Detection algorithms for spatial data

    Get PDF
    This dissertation addresses the problem of anomaly detection in spatial data. The problem of landmine detection in airborne spatial data is chosen as the specific detection scenario. The first part of the dissertation deals with the development of a fast algorithm for kernel-based non-linear anomaly detection in the airborne spatial data. The original Kernel RX algorithm, proposed by Kwon et al. [2005a], suffers from the problem of high computational complexity, and has seen limited application. With the aim to reduce the computational complexity, a reformulated version of the Kernel RX, termed the Spatially Weighted Kernel RX (SW-KRX), is presented. It is shown that under this reformulation, the detector statistics can be obtained directly as a function of the centered kernel Gram matrix. Subsequently, a methodology for the fast computation of the centered kernel Gram matrix is proposed. The key idea behind the proposed methodology is to decompose the set of image pixels into clusters, and expediting the computations by approximating the effect of each cluster as a whole. The SW-KRX algorithm is implemented for a special case, and comparative results are compiled for the SW-KRX vis-à-vis the RX anomaly detector. In the second part of the dissertation, a detection methodology for buried mine detection is presented. The methodology is based on extraction of color texture information using cross-co-occurrence features. A feature selection methodology based on Bhattacharya coefficients and principal feature analysis is proposed and detection results with different feature-based detectors are presented, to demonstrate the effectiveness of the proposed methodology in the extraction of useful discriminatory information --Abstract, page iii

    Scales and Scale-like Structures

    Get PDF
    Scales are a visually striking feature that grows on many animals. These small, rigid plates embedded in the skin form an integral part of our description of fish and reptiles, some plants, and many extinct animals. Scales exist in many shapes and sizes, and serve as protection, camouflage, and plumage for animals. The variety of scales and the animals they grow from pose an interesting problem in the field of Computer Graphics. This dissertation presents a method for generating scales and scale-like structures on a polygonal mesh through surface replacement. A triangular mesh was covered with scales and one or more proxy-models were used as the scales shape. A user began scale generation by drawing a lateral line on the model to control the distribution and orientation of scales on the surface. Next, a vector field was created over the surface to control an anisotropic Voronoi tessellation, which represents the region occupied by each scale. Then these regions were replaced by cutting the proxy model to match the boundary of the Voronoi region and deform the cut model onto the surface. The final result is a fully connected 2-manifold that is suitable for subsequent post-processing applications, like surface subdivision

    Microstructural modeling and computational homogenization of the physically linear and nonlinear constitutive behavior of micro-heterogeneous materials

    Get PDF
    Engineering materials show a pronounced heterogeneity on a smaller scale that influences the macroscopic constitutive behavior. Algorithms for the periodic discretization of microstructures are presented. These are used within the Nonuniform Transformation Field Analysis (NTFA) which is an order reduction based nonlinear homogenization method with micro-mechanical background. Theoretical and numerical aspects of the method are discussed and its computational efficiency is validated

    Doctor of Philosophy

    Get PDF
    dissertationOne of the fundamental building blocks of many computational sciences is the construction and use of a discretized, geometric representation of a problem domain, often referred to as a mesh. Such a discretization enables an otherwise complex domain to be represented simply, and computation to be performed over that domain with a finite number of basis elements. As mesh generation techniques have become more sophisticated over the years, focus has largely shifted to quality mesh generation techniques that guarantee or empirically generate numerically well-behaved elements. In this dissertation, the two complementary meshing subproblems of vertex placement and element creation are analyzed, both separately and together. First, a dynamic particle system achieves adaptivity over domains by inferring feature size through a new information passing algorithm. Second, a new tetrahedral algorithm is constructed that carefully combines lattice-based stenciling and mesh warping to produce guaranteed quality meshes on multimaterial volumetric domains. Finally, the ideas of lattice cleaving and dynamic particle systems are merged into a unified framework for producing guaranteed quality, unstructured and adaptive meshing of multimaterial volumetric domains
    corecore