432 research outputs found

    Information embedding and retrieval in 3D printed objects

    Get PDF
    Deep learning and convolutional neural networks have become the main tools of computer vision. These techniques are good at using supervised learning to learn complex representations from data. In particular, under limited settings, the image recognition model now performs better than the human baseline. However, computer vision science aims to build machines that can see. It requires the model to be able to extract more valuable information from images and videos than recognition. Generally, it is much more challenging to apply these deep learning models from recognition to other problems in computer vision. This thesis presents end-to-end deep learning architectures for a new computer vision field: watermark retrieval from 3D printed objects. As it is a new area, there is no state-of-the-art on many challenging benchmarks. Hence, we first define the problems and introduce the traditional approach, Local Binary Pattern method, to set our baseline for further study. Our neural networks seem useful but straightfor- ward, which outperform traditional approaches. What is more, these networks have good generalization. However, because our research field is new, the problems we face are not only various unpredictable parameters but also limited and low-quality training data. To address this, we make two observations: (i) we do not need to learn everything from scratch, we know a lot about the image segmentation area, and (ii) we cannot know everything from data, our models should be aware what key features they should learn. This thesis explores these ideas and even explore more. We show how to use end-to-end deep learning models to learn to retrieve watermark bumps and tackle covariates from a few training images data. Secondly, we introduce ideas from synthetic image data and domain randomization to augment training data and understand various covariates that may affect retrieve real-world 3D watermark bumps. We also show how the illumination in synthetic images data to effect and even improve retrieval accuracy for real-world recognization applications

    Acquisition and modeling of material appearance

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2006.Includes bibliographical references (p. 131-143).In computer graphics, the realistic rendering of synthetic scenes requires a precise description of surface geometry, lighting, and material appearance. While 3D geometry scanning and modeling have advanced significantly in recent years, measurement and modeling of accurate material appearance have remained critical challenges. Analytical models are the main tools to describe material appearance in most current applications. They provide compact and smooth approximations to real materials but lack the expressiveness to represent complex materials. Data-driven approaches based on exhaustive measurements are fully general but the measurement process is difficult and the storage requirement is very high. In this thesis, we propose the use of hybrid representations that are more compact and easier to acquire than exhaustive measurement, while preserving much generality of a data-driven approach. To represent complex bidirectional reflectance distribution functions (BRDFs), we present a new method to estimate a general microfacet distribution from measured data. We show that this representation is able to reproduce complex materials that are impossible to model with purely analytical models.(cont.) We also propose a new method that significantly reduces measurement cost and time of the bidirectional texture function (BTF) through a statistical characterization of texture appearance. Our reconstruction method combines naturally aligned images and alignment-insensitive statistics to produce visually plausible results. We demonstrate our acquisition system which is able to capture intricate materials like fabrics in less than ten minutes with commodity equipments. In addition, we present a method to facilitate effective user design in the space of material appearance. We introduce a metric in the space of reflectance which corresponds roughly to perceptual measures. The main idea of our approach is to evaluate reflectance differences in terms of their induced rendered images, instead of the reflectance function itself defined in the angular domains. With rendered images, we show that even a simple computational metric can provide good perceptual spacing and enable intuitive navigation of the reflectance space.by Wai Kit Addy Ngan.Ph.D

    Accelerated volumetric reconstruction from uncalibrated camera views

    Get PDF
    While both work with images, computer graphics and computer vision are inverse problems. Computer graphics starts traditionally with input geometric models and produces image sequences. Computer vision starts with input image sequences and produces geometric models. In the last few years, there has been a convergence of research to bridge the gap between the two fields. This convergence has produced a new field called Image-based Rendering and Modeling (IBMR). IBMR represents the effort of using the geometric information recovered from real images to generate new images with the hope that the synthesized ones appear photorealistic, as well as reducing the time spent on model creation. In this dissertation, the capturing, geometric and photometric aspects of an IBMR system are studied. A versatile framework was developed that enables the reconstruction of scenes from images acquired with a handheld digital camera. The proposed system targets applications in areas such as Computer Gaming and Virtual Reality, from a lowcost perspective. In the spirit of IBMR, the human operator is allowed to provide the high-level information, while underlying algorithms are used to perform low-level computational work. Conforming to the latest architecture trends, we propose a streaming voxel carving method, allowing a fast GPU-based processing on commodity hardware

    Texture mapping with ray tracing

    Get PDF
    Includes bibliographical references leaves 56-59.Thesis (Master's)--Bilkent University, 1993.Ankara : The Department of Computer Engineering and Information Science and Institute of Engineering and Science of Bilkent University,1993.By using texture generation and global illumination techniques, it is possible to produce realistic computer images. Currently, ray tracing is one of the most popular global illumination techniques due to its simplicity, elegance, and easy implementation. In this thesis, texture mapping techniques are used with ray tracing to generate high quality visual effects. The implementation of the mapping process is presented and an approach for combining prefiltering techniques with ray tracing is introduced. General sweep surfaces produced by Topologybook are used for modeling.by Uğur Akdemir.M.S

    Ray Tracing in Non-Euclidean Spaces

    Get PDF
    This dissertation describes a method for modeling, simulating and real-time rendering piecewise linear approximations of generic non-Euclidean 3D spaces. The 3D rendering pipeline most commonly used, where one multiplies each vertex coordinate by a 4x4 matrix to project it on the screen does not work for all cases where the space does not obey Euclid’s postulates (non-Euclidean space). Furthermore, while other non-Euclidean rendering tools only work for a limited type of spaces, our approach allows us to model, simulate, and render any isometrically embeddable non-Euclidean space and eventual objects lying therein. We envision at least two main applications for our approach. The first for helping mathematicians get a better understanding of what arbitrary spaces look like (e.g., hyperconical space, hyper-spherical space, and so forth). The second for assisting physicists to visualize and simulate the effects of bent space (e.g., black holes, wormholes, Alcubierre drive, and so forth) on light, and on physical objectsEsta dissertação descreve um mĂ©todo para modelar, simular e renderizar aproximaçÔes lineares de espaços nĂŁo Euclideanos de forma genĂ©rica e em tempo real. A tĂ©cnica de renderização 3D mais comum, que multiplica a matriz de projeção 4 x 4 por cada vĂ©rtice para determinar as coordenadas do respetivo pixel no ecrĂŁ, nem sempre funciona quando o espaço nĂŁo obedece a um postulado de Euclides (espaço nĂŁo-Euclideano). AlĂ©m disso, enquanto outras ferramentas para renderizar espaços nĂŁo-Euclideanos sĂł funcionam para certos tipos de espaços, a nossa tĂ©cnica permite modelar, simular e renderizar qualquer espaço nĂŁo-Euclideano embebĂ­vel isometricamente, bem como eventuais objetos nele existentes. Antevemos pelo menos dois usos para a nossa tĂ©cnica. A primeira para ajudar matemĂĄticos a compreender melhor o aspeto de espaços arbitrĂĄrios (e.g., espaço hiper-cĂłnico, espaço hiper-esfĂ©rico, etc.). A segunda para ajudar fĂ­sicos a visualizar e simular os efeitos de espaço curvo (e.g., buracos negros, buracos de minhoca, deformaçÔes Alcubierra drive, etc.) em luz e objetos fĂ­sicos circundantes

    Advancements in multi-view processing for reconstruction, registration and visualization.

    Get PDF
    The ever-increasing diffusion of digital cameras and the advancements in computer vision, image processing and storage capabilities have lead, in the latest years, to the wide diffusion of digital image collections. A set of digital images is usually referred as a multi-view images set when the pictures cover different views of the same physical object or location. In multi-view datasets, correlations between images are exploited in many different ways to increase our capability to gather enhanced understanding and information on a scene. For example, a collection can be enhanced leveraging on the camera position and orientation, or with information about the 3D structure of the scene. The range of applications of multi-view data is really wide, encompassing diverse fields such as image-based reconstruction, image-based localization, navigation of virtual environments, collective photographic retouching, computational photography, object recognition, etc. For all these reasons, the development of new algorithms to effectively create, process, and visualize this type of data is an active research trend. The thesis will present four different advancements related to different aspects of the multi-view data processing: - Image-based 3D reconstruction: we present a pre-processing algorithm, that is a special color-to-gray conversion. This was developed with the aim to improve the accuracy of image-based reconstruction algorithms. In particular, we show how different dense stereo matching results can be enhanced by application of a domain separation approach that pre-computes a single optimized numerical value for each image location. - Image-based appearance reconstruction: we present a multi-view processing algorithm, this can enhance the quality of the color transfer from multi-view images to a geo-referenced 3D model of a location of interest. The proposed approach computes virtual shadows and allows to automatically segment shadowed regions from the input images preventing to use those pixels in subsequent texture synthesis. - 2D to 3D registration: we present an unsupervised localization and registration system. This system can recognize a site that has been framed in a multi-view data and calibrate it on a pre-existing 3D representation. The system has a very high accuracy and it can validate the result in a completely unsupervised manner. The system accuracy is enough to seamlessly view input images correctly super-imposed on the 3D location of interest. - Visualization: we present PhotoCloud, a real-time client-server system for interactive exploration of high resolution 3D models and up to several thousand photographs aligned over this 3D data. PhotoCloud supports any 3D models that can be rendered in a depth-coherent way and arbitrary multi-view image collections. Moreover, it tolerates 2D-to-2D and 2D-to-3D misalignments, and it provides scalable visualization of generic integrated 2D and 3D datasets by exploiting data duality. A set of effective 3D navigation controls, tightly integrated with innovative thumbnail bars, enhances the user navigation. These advancements have been developed in tourism and cultural heritage application contexts, but they are not limited to these
    • 

    corecore