18,090 research outputs found

    Tracking of secondary and temporary objects in structural concrete work

    Get PDF
    Previous research has shown that “Scan-vs-BIM ” object recognition systems, that fuse 3D point clouds from Terrestrial Laser Scanning (TLS) or digital photogrammetry with 4D project BIM, provide valuable information for tracking structural works. However, until now, the potential of these systems has been demonstrated for tracking progress of permanent structures only; no work has been reported yet on tracking secondary or temporary structures. For structural concrete work, temporary structures include formwork, scaffolding and shoring, while secondary components include rebar. Together, they constitute most of the earned value in concrete work. The impact of tracking such elements would thus be added veracity and detail to earned value calculations, and subsequently better project control and performance. This paper presents three different techniques for recognizing concrete construction secondary and temporary objects in TLS point clouds. Two of the techniques are tested using real-life data collected from a reinforced concrete building construction site. The preliminary experimental results show that it is feasible to recognize secondary and temporary objects in TLS point clouds with good accuracy; but it is envisaged that superior results could be achieved by using additional cues such colour and 3D edge information

    3D freeform surfaces from planar sketches using neural networks

    Get PDF
    A novel intelligent approach into 3D freeform surface reconstruction from planar sketches is proposed. A multilayer perceptron (MLP) neural network is employed to induce 3D freeform surfaces from planar freehand curves. Planar curves were used to represent the boundaries of a freeform surface patch. The curves were varied iteratively and sampled to produce training data to train and test the neural network. The obtained results demonstrate that the network successfully learned the inverse-projection map and correctly inferred the respective surfaces from fresh curves

    On Recognizing Transparent Objects in Domestic Environments Using Fusion of Multiple Sensor Modalities

    Full text link
    Current object recognition methods fail on object sets that include both diffuse, reflective and transparent materials, although they are very common in domestic scenarios. We show that a combination of cues from multiple sensor modalities, including specular reflectance and unavailable depth information, allows us to capture a larger subset of household objects by extending a state of the art object recognition method. This leads to a significant increase in robustness of recognition over a larger set of commonly used objects.Comment: 12 page

    Form Perception

    Full text link
    National Science Foundation (SBE-0354378); Office of Naval Research (N00014-01-1-0624

    Learning from Millions of 3D Scans for Large-scale 3D Face Recognition

    Full text link
    Deep networks trained on millions of facial images are believed to be closely approaching human-level performance in face recognition. However, open world face recognition still remains a challenge. Although, 3D face recognition has an inherent edge over its 2D counterpart, it has not benefited from the recent developments in deep learning due to the unavailability of large training as well as large test datasets. Recognition accuracies have already saturated on existing 3D face datasets due to their small gallery sizes. Unlike 2D photographs, 3D facial scans cannot be sourced from the web causing a bottleneck in the development of deep 3D face recognition networks and datasets. In this backdrop, we propose a method for generating a large corpus of labeled 3D face identities and their multiple instances for training and a protocol for merging the most challenging existing 3D datasets for testing. We also propose the first deep CNN model designed specifically for 3D face recognition and trained on 3.1 Million 3D facial scans of 100K identities. Our test dataset comprises 1,853 identities with a single 3D scan in the gallery and another 31K scans as probes, which is several orders of magnitude larger than existing ones. Without fine tuning on this dataset, our network already outperforms state of the art face recognition by over 10%. We fine tune our network on the gallery set to perform end-to-end large scale 3D face recognition which further improves accuracy. Finally, we show the efficacy of our method for the open world face recognition problem.Comment: 11 page
    • 

    corecore