1,809 research outputs found

    Towards Zero-Waste Furniture Design

    Get PDF
    In traditional design, shapes are first conceived, and then fabricated. While this decoupling simplifies the design process, it can result in inefficient material usage, especially where off-cut pieces are hard to reuse. The designer, in absence of explicit feedback on material usage remains helpless to effectively adapt the design -- even though design variabilities exist. In this paper, we investigate {\em waste minimizing furniture design} wherein based on the current design, the user is presented with design variations that result in more effective usage of materials. Technically, we dynamically analyze material space layout to determine {\em which} parts to change and {\em how}, while maintaining original design intent specified in the form of design constraints. We evaluate the approach on simple and complex furniture design scenarios, and demonstrate effective material usage that is difficult, if not impossible, to achieve without computational support

    A study of user perceptions of the relationship between bump-mapped and non-bump-mapped materials, and lighting intensity in a real-time virtual environment

    Get PDF
    The video and computer games industry has taken full advantage of the human sense of vision by producing games that utilize complex high-resolution textures and materials, and lighting technique. This results to the creation of an almost life-like real-time 3D virtual environment that can immerse the end-users. One of the visual techniques used is real-time display of bump-mapped materials. However, this sense of visual phenomenon has yet to be fully utilized for 3D design visualization in the architecture and construction domain. Virtual environments developed in the architecture and construction domain are often basic and use low-resolution images, which under represent the real physical environment. Such virtual environment is seen as being non-realistic to the user resulting in a misconception of the actual potential of it as a tool for 3D design visualization. A study was conducted to evaluate whether subjects can see the difference between bump-mapped and nonbump-mapped materials in different lighting conditions. The study utilized a real-time 3D virtual environment that was created using a custom-developed software application tool called BuildITC4. BuildITC4 was developed based upon the C4Engine which is classified as a next-generation 3D Game Engine. A total of thirty-five subjects were exposed to the virtual environment and were asked to compare the various types of material in different lighting conditions. The number of lights activated, the lighting intensity, and the materials used in the virtual environment were all interactive and changeable in real-time. The goal is to study how subjects perceived bump-mapped and non-bump mapped materials, and how different lighting conditions affect realistic representation. Results from this study indicate that subjects could tell the difference between the bump-mapped and non-bump mapped materials, and how different material reacts to different lighting condition

    The Potential for a GPU-Like Overlay Architecture for FPGAs

    Get PDF
    We propose a soft processor programming model and architecture inspired by graphics processing units (GPUs) that are well-matched to the strengths of FPGAs, namely, highly parallel and pipelinable computation. In particular, our soft processor architecture exploits multithreading, vector operations, and predication to supply a floating-point pipeline of 64 stages via hardware support for up to 256 concurrent thread contexts. The key new contributions of our architecture are mechanisms for managing threads and register files that maximize data-level and instruction-level parallelism while overcoming the challenges of port limitations of FPGA block memories as well as memory and pipeline latency. Through simulation of a system that (i) is programmable via NVIDIA's high-level Cg language, (ii) supports AMD's CTM r5xx GPU ISA, and (iii) is realizable on an XtremeData XD1000 FPGA-based accelerator system, we demonstrate the potential for such a system to achieve 100% utilization of a deeply pipelined floating-point datapath

    On the Information Rates of the Plenoptic Function

    Get PDF
    The {\it plenoptic function} (Adelson and Bergen, 91) describes the visual information available to an observer at any point in space and time. Samples of the plenoptic function (POF) are seen in video and in general visual content, and represent large amounts of information. In this paper we propose a stochastic model to study the compression limits of the plenoptic function. In the proposed framework, we isolate the two fundamental sources of information in the POF: the one representing the camera motion and the other representing the information complexity of the "reality" being acquired and transmitted. The sources of information are combined, generating a stochastic process that we study in detail. We first propose a model for ensembles of realities that do not change over time. The proposed model is simple in that it enables us to derive precise coding bounds in the information-theoretic sense that are sharp in a number of cases of practical interest. For this simple case of static realities and camera motion, our results indicate that coding practice is in accordance with optimal coding from an information-theoretic standpoint. The model is further extended to account for visual realities that change over time. We derive bounds on the lossless and lossy information rates for this dynamic reality model, stating conditions under which the bounds are tight. Examples with synthetic sources suggest that in the presence of scene dynamics, simple hybrid coding using motion/displacement estimation with DPCM performs considerably suboptimally relative to the true rate-distortion bound.Comment: submitted to IEEE Transactions in Information Theor
    • …
    corecore