78,990 research outputs found

    Posing 3D Models from Drawing

    Get PDF
    Inferring the 3D pose of a character from a drawing is a complex and under-constrained problem. Solving it may help automate various parts of an animation production pipeline such as pre-visualisation. In this paper, a novel way of inferring the 3D pose from a monocular 2D sketch is proposed. The proposed method does not make any external assumptions about the model, allowing it to be used on different types of characters. The inference of the 3D pose is formulated as an optimisation problem and a parallel variation of the Particle Swarm Optimisation algorithm called PARAC-LOAPSO is utilised for searching the minimum. Testing in isolation as well as part of a larger scene, the presented method is evaluated by posing a lamp, a horse and a human character. The results show that this method is robust, highly scalable and is able to be extended to various types of models

    Determining geometric primitives for a 3D GIS : easy as 1D, 2D, 3D?

    Get PDF
    Acquisition techniques such as photo modelling, using SfM-MVS algorithms, are being applied increasingly in several fields of research and render highly realistic and accurate 3D models. Nowadays, these 3D models are mainly deployed for documentation purposes. As these data generally encompass spatial data, the development of a 3D GIS would allow researchers to use these 3D models to their full extent. Such a GIS would allow a more elaborate analysis of these 3D models and thus support the comprehension of the objects that the features in the model represent. One of the first issues that has to be tackled in order to make the resulting 3D models compatible for implementation in a 3D GIS is the choice of a certain geometric primitive to spatially represent the input data. The chosen geometric primitive will not only influence the visualisation of the data, but also the way in which the data can be stored, exchanged, manipulated, queried and understood. Geometric primitives can be one-, two- and three-dimensional. By adding an extra dimension, the complexity of the data increases, but the user is allowed to understand the original situation more intuitively. This research paper tries to give an initial analysis of 1D, 2D and 3D primitives in the framework of the integration of SfM-MVS based 3D models in a 3D GIS

    Component-wise modeling of articulated objects

    Get PDF
    We introduce a novel framework for modeling articulated objects based on the aspects of their components. By decomposing the object into components, we divide the problem in smaller modeling tasks. After obtaining 3D models for each component aspect by employing a shape deformation paradigm, we merge them together, forming the object components. The final model is obtained by assembling the components using an optimization scheme which fits the respective 3D models to the corresponding apparent contours in a reference pose. The results suggest that our approach can produce realistic 3D models of articulated objects in reasonable time

    Detector and Event Visualization with SketchUp at the CMS Experiment

    Full text link
    We have created 3D models of the CMS detector and particle collision events in SketchUp, a 3D modelling program. SketchUp provides a Ruby API which we use to interface with the CMS Detector Description to create 3D models of the CMS detector. With the Ruby API, we also have created an interface to the JSON-based event format used for the iSpy event display to create 3D models of CMS events. These models have many applications related to 3D representation of the CMS detector and events. Figures produced based on these models were used in conference presentations, journal publications, technical design reports for the detector upgrades, art projects, outreach programs, and other presentations.Comment: 5 pages, 6 figures, Proceedings for CHEP 2013, 20th International Conference on Computing in High Energy and Nuclear Physic

    Developing Interaction 3D Models for E-Learning Applications

    Get PDF
    Some issues concerning the development of interactive 3D models for e-learning applications are considered. Given that 3D data sets are normally large and interactive display demands high performance computation, a natural solution would be placing the computational burden on the client machine rather than on the server. Mozilla and Google opted for a combination of client-side languages, JavaScript and OpenGL, to handle 3D graphics in a web browser (Mozilla 3D and O3D respectively). Based on the O3D model, core web technologies are considered and an example of the full process involving the generation of a 3D model and their interactive visualization in a web browser is described. The challenging issue of creating realistic 3D models of objects in the real world is discussed and a method based on line projection for fast 3D reconstruction is presented. The generated model is then visualized in a web browser. The experiments demonstrate that visualization of 3D data in a web browser can provide quality user experience. Moreover, the development of web applications are facilitated by O3D JavaScript extension allowing web designers to focus on 3D contents generation

    Creating Simplified 3D Models with High Quality Textures

    Get PDF
    This paper presents an extension to the KinectFusion algorithm which allows creating simplified 3D models with high quality RGB textures. This is achieved through (i) creating model textures using images from an HD RGB camera that is calibrated with Kinect depth camera, (ii) using a modified scheme to update model textures in an asymmetrical colour volume that contains a higher number of voxels than that of the geometry volume, (iii) simplifying dense polygon mesh model using quadric-based mesh decimation algorithm, and (iv) creating and mapping 2D textures to every polygon in the output 3D model. The proposed method is implemented in real-time by means of GPU parallel processing. Visualization via ray casting of both geometry and colour volumes provides users with a real-time feedback of the currently scanned 3D model. Experimental results show that the proposed method is capable of keeping the model texture quality even for a heavily decimated model and that, when reconstructing small objects, photorealistic RGB textures can still be reconstructed.Comment: 2015 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Page 1 -
    corecore