6 research outputs found

    Algorithms for Data Cleaning in Knowledge Bases

    No full text
    Data cleaning is an action which includes a process of correcting and identifying the inconsistencies and errors in data warehouse. Different terms are uses in these papers like data cleaning also called data scrubbing. Using data scrubbing to get high quality data and this is one the data ETL (extraction transformation and loading tools). Now a day there is a need of authentic information for better decision-making. So we conduct a review paper in which six papers are reviewed related to data cleaning. Relating papers discussed different algorithms, methods, problems, their solutions, and approaches etc. Each paper has their own methods to solve a problem in an efficient way, but all the paper have a common problem of data cleaning and inconsistencies. In these papers data inconsistencies, identification of the errors, conflicting, duplicate records etc problems are discussed in detail and also provided the solutions. These algorithms increase the quality of data. At ETL process stage, there are almost thirty-five different sources and causes of poor quality constraints
    corecore