57 research outputs found

    Splash 데이터 μ „μ²˜λ¦¬ μ—°μ‚°μžλ₯Ό μ΄μš©ν•œ 지도 ν•™μŠ΅ 데이터 증강과 필터링

    Get PDF
    ν•™μœ„λ…Όλ¬Έ(석사) -- μ„œμšΈλŒ€ν•™κ΅λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ 전기·정보곡학뢀, 2021.8. κΉ€μ§€ν™˜.Splash is a graphical user interface programming framework designed to support artificial intelligence application development. Artificial intelligence experts in various fields including data, modeling, control engineers can easily develop artificial intelligence applications without profound programming knowledge through Splash’s programming abstraction. To further increase Splash’s functionality for supporting artificial intelligence application development, we are adding a language construct in Splash for data preprocessing. This language construct provides an easy-to-use data augmenter and data filter, which are the main tasks of data preprocessing for data engineers in supervised learning. Data augmentation and filtering are particularly important tasks in supervised learning because the training dataset's quality and quantity directly affect the accuracy of the model. Datasets such as MNIST and datasets prepared in person have data with accurate labels yet lack an amount of data and labels, so the datasets need augmentation for an increase in dataset quantity. When using a data label platform such as crowdsourcing or an automated label program to utilize numerous datasets for training, the datasets need filtering because they often include noisy labels. In this thesis, we implement basic data augmentation and filtering techniques as a Splash language construct, called data preprocessor, to support data engineers. Data augmentation function in Splash data preprocessor increases dataset quantity by using seven augmentation techniques: horizontal and vertical shift, horizontal and vertical flip, random rotation, random brightness, and random zoom. The data filtering function finds duplicated images with different and same labels, then removes those images to improve the quality of the training dataset. To demonstrate the feasibility of using Splash data preprocessor and to confirm the correctness of the data preprocessor implementation, we trained the CIFAR-10 dataset as an experiment using Splash data preprocessor. This experiment shows that training data filtering and augmentation can be easily performed using the Splash data preprocessor.SplashλŠ” 인곡 지λŠ₯ μ‘μš© κ°œλ°œμ„ μ§€μ›ν•˜κΈ° μœ„ν•΄ λ§Œλ“€μ–΄μ§„ GUI ν”„λ‘œκ·Έλž˜λ° ν”„λ ˆμž„μ›Œν¬μ΄λ‹€. SplashλŠ” ν”„λ‘œκ·Έλž˜λ° 좔상화λ₯Ό 톡해 데이터, AI λͺ¨λΈλ§, μ œμ–΄ μ—”μ§€λ‹ˆμ–΄λ₯Ό ν¬ν•¨ν•œ μ—¬λŸ¬ λΆ„μ•Ό 전문가듀이 ν”„λ‘œκ·Έλž˜λ°μ  지식 없이도 μ†μ‰½κ²Œ μ‚¬μš©ν•  수 μžˆλ„λ‘ λ§Œλ“€μ–΄μ‘Œλ‹€. 인곡 지λŠ₯ μ‘μš© κ°œλ°œμ„ μ§€μ›ν•˜λŠ” Splash의 κΈ°λŠ₯을 λ”μš± ν–₯μƒμ‹œν‚€κΈ° μœ„ν•˜μ—¬ 데이터 μ „μ²˜λ¦¬ κΈ°λŠ₯을 Splash의 μ–Έμ–΄ ꡬ쑰둜 μΆ”κ°€ν•˜μ˜€λ‹€. 이 μ–Έμ–΄ κ΅¬μ‘°λŠ” 데이터 μ—”μ§€λ‹ˆμ–΄μ˜ μ£Όμš” 업무인 데이터 μ „μ²˜λ¦¬ 쀑 데이터 필터링과 증강 κΈ°λŠ₯을 μ§€μ›ν•œλ‹€. 지도 ν•™μŠ΅(supervised learning)μ—μ„œ 데이터 필터링과 증강은 특히 μ€‘μš”ν•œ μž‘μ—…μ΄λ‹€. μ§€λ„ν•™μŠ΅μ„ μœ„ν•΄μ„œλŠ” λ ˆμ΄λΈ”μ΄ λ˜μ–΄μžˆλŠ” 데이터가 ν•„μš”ν•œλ°, μ‰½κ²Œ ꡬ할 수 μžˆλŠ” MNIST와 같은 ν•™μŠ΅ λ°μ΄ν„°μ…‹μ΄λ‚˜ 직접 λ ˆμ΄λΈ”λ§ ν•œ 데이터셋은 μˆ˜κ°€ ν•œμ •μ μ΄λ‹€. λ”°λΌμ„œ λ°μ΄ν„°μ˜ 수λ₯Ό μ¦κ°€μ‹œν‚€κΈ° μœ„ν•˜μ—¬ 데이터 증강 기술이 ν•„μš”ν•˜λ‹€. λ§Žμ€ 수의 데이터셋을 ν™œμš©ν•˜κΈ° μœ„ν•΄μ„œ ν¬λΌμš°λ“œμ†Œμ‹± 같은 데이터 λ ˆμ΄λΈ” ν”Œλž«νΌμ΄λ‚˜ μžλ™ λ ˆμ΄λΈ” ν”„λ‘œκ·Έλž¨μ„ μ΄μš©ν•˜λŠ” 경우, λ ˆμ΄λΈ”μ΄ 잘λͺ»λ˜μ–΄ μžˆλŠ” κ²½μš°κ°€ 많기 λ•Œλ¬Έμ— 이λ₯Ό 필터링해야 ν•œλ‹€. λ³Έ λ…Όλ¬Έμ—μ„œλŠ” 지도 ν•™μŠ΅μ—μ„œ ν•„μš”ν•œ 기본적인 데이터 필터링 기법과 데이터 증강 기법을 Splash에 κ΅¬ν˜„ν•˜μ—¬ 데이터 μ—”μ§€λ‹ˆμ–΄κ°€ μ†μ‰½κ²Œ μ΄μš©ν•  수 μžˆλ„λ‘ ν•œλ‹€. Splash 데이터 μ „μ²˜λ¦¬ μ—°μ‚°μžλŠ” μ΄λ―Έμ§€μ˜ 쀑볡성을 νŒλ‹¨ν•˜μ—¬ ν•„ν„°λ§ν•˜κ³ , 일곱 가지 λ°©λ²•μœΌλ‘œ 이미지λ₯Ό μ¦κ°•μ‹œν‚¨λ‹€. μš°λ¦¬λŠ” Splash 데이터 μ „μ²˜λ¦¬ μ—°μ‚°μžλ₯Ό μ‚¬μš©ν•˜μ—¬ 지도 ν•™μŠ΅ 데이터 필터링 및 증강을 μ‰½κ²Œ μˆ˜ν–‰ ν•  수 μžˆμŒμ„ λ³΄μ˜€λ‹€.Chapter 1. Introduction 1 Chapter 2. Splash programming language 4 Chapter 3. Splash data preprocessor 9 Chapter 4. Splash data preprocessor experiment 14 Chapter 5. Conclusion 18 References 19 Abstract in Korean 21석

    Model and Data Agreement for Learning with Noisy Labels

    Full text link
    Learning with noisy labels is a vital topic for practical deep learning as models should be robust to noisy open-world datasets in the wild. The state-of-the-art noisy label learning approach JoCoR fails when faced with a large ratio of noisy labels. Moreover, selecting small-loss samples can also cause error accumulation as once the noisy samples are mistakenly selected as small-loss samples, they are more likely to be selected again. In this paper, we try to deal with error accumulation in noisy label learning from both model and data perspectives. We introduce mean point ensemble to utilize a more robust loss function and more information from unselected samples to reduce error accumulation from the model perspective. Furthermore, as the flip images have the same semantic meaning as the original images, we select small-loss samples according to the loss values of flip images instead of the original ones to reduce error accumulation from the data perspective. Extensive experiments on CIFAR-10, CIFAR-100, and large-scale Clothing1M show that our method outperforms state-of-the-art noisy label learning methods with different levels of label noise. Our method can also be seamlessly combined with other noisy label learning methods to further improve their performance and generalize well to other tasks. The code is available in https://github.com/zyh-uaiaaaa/MDA-noisy-label-learning.Comment: Accepted by AAAI2023 Worksho

    Meta Soft Label Generation for Noisy Labels

    Full text link
    The existence of noisy labels in the dataset causes significant performance degradation for deep neural networks (DNNs). To address this problem, we propose a Meta Soft Label Generation algorithm called MSLG, which can jointly generate soft labels using meta-learning techniques and learn DNN parameters in an end-to-end fashion. Our approach adapts the meta-learning paradigm to estimate optimal label distribution by checking gradient directions on both noisy training data and noise-free meta-data. In order to iteratively update soft labels, meta-gradient descent step is performed on estimated labels, which would minimize the loss of noise-free meta samples. In each iteration, the base classifier is trained on estimated meta labels. MSLG is model-agnostic and can be added on top of any existing model at hand with ease. We performed extensive experiments on CIFAR10, Clothing1M and Food101N datasets. Results show that our approach outperforms other state-of-the-art methods by a large margin.Comment: Accepted by ICPR 202
    • …
    corecore