8 research outputs found

    Automatic Filtering of Lidar Building Point Cloud in Case of Trees Associated to Building Roof

    Get PDF
    This paper suggests a new algorithm for automatic building point cloud filtering based on the Z coordinate histogram. This operation aims to select the roof class points from the building point cloud, and the suggested algorithm considers the general case where high trees are associated with the building roof. The Z coordinate histogram is analyzed in order to divide the building point cloud into three zones: the surrounding terrain and low vegetation, the facades, and the tree crowns and/or the roof points. This operation allows the elimination of the first two classes which represent an obstacle toward distinguishing between the roof and the tree points. The analysis of the normal vectors, in addition to the change of curvature factor of the roof class leads to recognizing the high tree crown points. The suggested approach was tested on five datasets with different point densities and urban typology. Regarding the results’ accuracy quantification, the average values of the correctness, the completeness, and the quality indices are used. Their values are, respectively, equal to 97.9%, 97.6%, and 95.6%. These results confirm the high efficacy of the suggested approach

    A CNN Based Approach for Garments Texture Design Classification

    Get PDF
    Identifying garments texture design automatically for recommending the fashion trends is important nowadays because of the rapid growth of online shopping. By learning the properties of images efficiently, a machine can give better accuracy of classification. Several Hand-Engineered feature coding exists for identifying garments design classes. Recently, Deep Convolutional Neural Networks (CNNs) have shown better performances for different object recognition. Deep CNN uses multiple levels of representation and abstraction that helps a machine to understand the types of data more accurately. In this paper, a CNN model for identifying garments design classes has been proposed. Experimental results on two different datasets show better results than existing two well-known CNN models (AlexNet and VGGNet) and some state-of-the-art Hand-Engineered feature extraction methods

    Datasets for Aspect-Based Sentiment Analysis in Bangla and Its Baseline Evaluation

    No full text
    With the extensive growth of user interactions through prominent advances of the Web, sentiment analysis has obtained more focus from an academic and a commercial point of view. Recently, sentiment analysis in the Bangla language is progressively being considered as an important task, for which previous approaches have attempted to detect the overall polarity of a Bangla document. To the best of our knowledge, there is no research on the aspect-based sentiment analysis (ABSA) of Bangla text. This can be described as being due to the lack of available datasets for ABSA. In this paper, we provide two publicly available datasets to perform the ABSA task in Bangla. One of the datasets consists of human-annotated user comments on cricket, and the other dataset consists of customer reviews of restaurants. We also describe a baseline approach for the subtask of aspect category extraction to evaluate our datasets

    Computer Vision-Based Gender Detection from Facial Image

    No full text
    face detection, object detection, image processing Abstract � Computer vision-based gender detection from facial images is a challenging and important task for computer vision-based researchers. The automatic gender detection from face images has potential applications in visual surveillance and human-computer interaction systems (HCI). Human faces provide important visual information for gender perception. This research presents a novel approach for gender detection from facial images. The system can automatically detect face from input images and the detected facial area is taken as region of interest (ROI). Discrete Cosine Transformation (DCT) of that ROI plays an important role in gender detection. A gender knowledgebase of the processed DCT is created utilizing supervised learning. To detect gender, input image is passed through a classifier which is based on that knowledgebase. To improve the matching accuracy Local Binary Pattern (LBP) of the ROI is done before converting it into DCT. This research has experimented on a database of more than 4000 facial images which are mainly of this subcontinent in order to evaluate the performance of the proposed system. The average accuracy rate achieved by the system is more than 78%. 1

    An Automated System for Garment Texture Design Class Identification

    No full text
    Automatic identification of garment design class might play an important role in the garments and fashion industry. To achieve this, essential initial works are found in the literature. For example, construction of a garment database, automatic segmentation of garments from real life images, categorizing them into the type of garments such as shirts, jackets, tops, skirts, etc. It is now essential to find a system such that it will be possible to identify the particular design (printed, striped or single color) of garment product for an automated system to recommend the garment trends. In this paper, we have focused on this specific issue and thus propose two new descriptors namely Completed CENTRIST (cCENTRIST) and Ternary CENTRIST (tCENTRIST). To test these descriptors, we used two different publically available databases. The experimental results of these databases demonstrate that both cCENTRIST and tCENTRIST achieve nearly about 3% more accuracy than the existing state-of-the art methods

    Machine learning-based segmentation of aerial LiDAR point cloud data on building roof

    No full text
    ABSTRACTThree-dimensional (3D) reconstruction of a building can be facilitated by correctly segmenting different feature points (e.g. in the form of boundary, fold edge, and planar points) over the building roof, and then, establishing relationships among the constructed feature lines and planar patches using the segmented points. Present machine learning-based segmentation approaches of Light Detection and Ranging (LiDAR) point cloud data are confined only to different object classes or semantic labelling. In the context of fine-grained feature point classification over the extracted building roof, machine learning approaches have not yet been explored. In this paper, after generating the ground truth data for the extracted building roofs from three different datasets, we apply machine learning methods to segment the roof point cloud based on seven different effective geometric features. The goal is not to semantically enhance the point cloud, but rather to facilitate the application of 3D building reconstruction algorithms, making them easier to use. The calculated F1-scores for each class confirm the competitive performances over the state-of-the-art techniques, which are more than 95% almost in each area of the used datasets

    Full series algorithm of automatic building extraction and modelling from LiDAR data

    No full text
    This paper suggests an algorithm that automatically links the automatic building classification and modelling algorithms. To make this connection, the suggested algorithm applies two filters to the building classification results that enable processing of the failed cases of the classification algorithm. In this context, it filters the noisy terrain class and analyses the remaining points to detect missing buildings. Moreover, it filters the detected building to eliminate all undesirable points such as those associated with trees overhanging the building roof, the surrounding terrain and the façade points. In the modelling algorithm, the error map matrix is analysed to recognize the failed cases of the building modelling algorithm with these buildings being modelled with flat roofs. Finally, the region growing algorithm is applied on the building mask to detect each building and pass it to the modelling algorithm. The accuracy analysis of the classification and modelling algorithm within the global algorithm shows it to be highly effective. Hence, the total error of the building classification algorithm is 0.01% and only one building in the sample dataset is rejected by the modelling algorithm and even that is modelled, but with a flat roof. Most of the buildings have Segmentation Accuracy and Quality factor less than 5% (error less than 5%) which means that the resulting evaluation is excellent

    HeteroEdge: Addressing Asymmetry in Heterogeneous Collaborative Autonomous Systems

    Full text link
    Gathering knowledge about surroundings and generating situational awareness for IoT devices is of utmost importance for systems developed for smart urban and uncontested environments. For example, a large-area surveillance system is typically equipped with multi-modal sensors such as cameras and LIDARs and is required to execute deep learning algorithms for action, face, behavior, and object recognition. However, these systems face power and memory constraints due to their ubiquitous nature, making it crucial to optimize data processing, deep learning algorithm input, and model inference communication. In this paper, we propose a self-adaptive optimization framework for a testbed comprising two Unmanned Ground Vehicles (UGVs) and two NVIDIA Jetson devices. This framework efficiently manages multiple tasks (storage, processing, computation, transmission, inference) on heterogeneous nodes concurrently. It involves compressing and masking input image frames, identifying similar frames, and profiling devices to obtain boundary conditions for optimization.. Finally, we propose and optimize a novel parameter split-ratio, which indicates the proportion of the data required to be offloaded to another device while considering the networking bandwidth, busy factor, memory (CPU, GPU, RAM), and power constraints of the devices in the testbed. Our evaluations captured while executing multiple tasks (e.g., PoseNet, SegNet, ImageNet, DetectNet, DepthNet) simultaneously, reveal that executing 70% (split-ratio=70%) of the data on the auxiliary node minimizes the offloading latency by approx. 33% (18.7 ms/image to 12.5 ms/image) and the total operation time by approx. 47% (69.32s to 36.43s) compared to the baseline configuration (executing on the primary node)
    corecore