89 research outputs found

    Development of a Data Mining System for Subscriber Classification (Case Study: Electricity Distribution Company)

    Get PDF
    Currently, organizations and companies tend to provide customers with good and suitable services in accordance with their interests and behaviors. Thus, the better the customers are classified, the better the services provided will be. Data mining is an efficient process for helping companies discover patterns in the database and it is important to identify target customers in this process. In fact, customers are selected to provide new products and services. Customer classification is based on data mining techniques for customer identification. This study tends to classify customers using data mining algorithms such as decision tree CART, neural network and regression. The case study is customers of Electricity Distribution Company. Simulation results based on Clementine software show that population had the highest effect on the amount of power consumed in each of the six household, public, industrial, agricultural, road and commercial classes. This is consistent with the opinion of experts in the electric power industry, because higher number of subscribers of each class surely increases the amount of electricity consumed (not steadily). The second effective feature of power consumption in six classes is humidity, which in many classes has a relatively equivalent effect with the effect of temperature on power consumption

    Implementing the 2-D Wavelet Transform on SIMD-Enhanced General-Purpose Processors

    Full text link

    Copy-move forgery detection using convolutional neural network and K-mean clustering

    Get PDF
    Copying and pasting a patch of an image to hide or exaggerate something in a digital image is known as a copy-move forgery. Copy-move forgery detection (CMFD) is hard to detect because the copied part image from a scene has similar properties with the other parts of the image in terms of texture, light illumination, and objective. The CMFD is still a challenging issue in some attacks such as rotation, scaling, blurring, and noise. In this paper, an approach using the convolutional neural network (CNN) and k-mean clustering is for CMFD. To identify cloned parts candidates, a patch of an image is extracted using corner detection. Next, similar patches are detected using a pre-trained network inspired by the Siamese network. If two similar patches are not evidence of the CMFD, the post-process is performed using k-means clustering. Experimental analyses are done on MICC-F2000, MICC-F600, and MICC-F8 databases. The results showed that using the proposed algorithm we can receive a 94.13% and 96.98% precision and F1 score, respectively, which are the highest among all state-of-the-art algorithms

    Target Tracking Based on Virtual Grid in Wireless Sensor Networks

    Get PDF
    One of the most important and typical application of wireless sensor networks (WSNs) is target tracking. Although target tracking, can provide benefits for large-scale WSNs and organize them into clusters but tracking a moving target in cluster-based WSNs suffers a boundary problem. The main goal of this paper was to introduce an efficient and novel mobility management protocol namely Target Tracking Based on Virtual Grid (TTBVG), which integrates on-demand dynamic clustering into a cluster- based WSN for target tracking. This protocol converts on-demand dynamic clusters to scalable cluster-based WSNs, by using boundary nodes and facilitates sensors’ collaboration around clusters. In this manner, each sensor node has the probability of becoming a cluster head and apperceives the tradeoff between energy consumption and local sensor collaboration in cluster-based sensor networks. The simulation results of this study demonstrated that the efficiency of the proposed protocol in both one-hop and multi-hop cluster-based sensor networks

    Image retrieval using the combination of text-based and content-based algorithms

    Get PDF
    Image retrieval is an important research field which has received great attention in the last decades. In this paper, we present an approach for the image retrieval based on the combination of text-based and content-based features. For text-based features, keywords and for content-based features, color and texture features have been used. Query in this system contains some keywords and an input image. At first, the images are retrieved based on the input keywords. Then, visual features are extracted to retrieve ideal output images. For extraction of color features we have used color moments and for texture we have used color co-occurrence matrix. The COREL image database have been used for our experimental results. The experimental results show that the performance of the combination of both text- and content- based features is much higher than each of them which is applied separately

    Data sanitization in association rule mining based on impact factor

    Get PDF
    Data sanitization is a process that is used to promote the sharing of transactional databases among organizations and businesses, it alleviates concerns for individuals and organizations regarding the disclosure of sensitive patterns. It transforms the source database into a released database so that counterparts cannot discover the sensitive patterns and so data confidentiality is preserved against association rule mining method. This process strongly rely on the minimizing the impact of data sanitization on the data utility by minimizing the number of lost patterns in the form of non-sensitive patterns which are not mined from sanitized database. This study proposes a data sanitization algorithm to hide sensitive patterns in the form of frequent itemsets from the database while controls the impact of sanitization on the data utility using estimation of impact factor of each modification on non-sensitive itemsets. The proposed algorithm has been compared with Sliding Window size Algorithm (SWA) and Max-Min1 in term of execution time, data utility and data accuracy. The data accuracy is defined as the ratio of deleted items to the total support values of sensitive itemsets in the source dataset. Experimental results demonstrate that proposed algorithm outperforms SWA and Max-Min1 in terms of maximizing the data utility and data accuracy and it provides better execution time over SWA and Max-Min1 in high scalability for sensitive itemsets and transactions
    corecore