57 research outputs found
Splash λ°μ΄ν° μ μ²λ¦¬ μ°μ°μλ₯Ό μ΄μ©ν μ§λ νμ΅ λ°μ΄ν° μ¦κ°κ³Ό νν°λ§
νμλ
Όλ¬Έ(μμ¬) -- μμΈλνκ΅λνμ : 곡과λν μ κΈ°Β·μ 보곡νλΆ, 2021.8. κΉμ§ν.Splash is a graphical user interface programming framework designed to support artificial intelligence application development. Artificial intelligence experts in various fields including data, modeling, control engineers can easily develop artificial intelligence applications without profound programming knowledge through Splashβs programming abstraction. To further increase Splashβs functionality for supporting artificial intelligence application development, we are adding a language construct in Splash for data preprocessing. This language construct provides an easy-to-use data augmenter and data filter, which are the main tasks of data preprocessing for data engineers in supervised learning.
Data augmentation and filtering are particularly important tasks in supervised learning because the training dataset's quality and quantity directly affect the accuracy of the model. Datasets such as MNIST and datasets prepared in person have data with accurate labels yet lack an amount of data and labels, so the datasets need augmentation for an increase in dataset quantity. When using a data label platform such as crowdsourcing or an automated label program to utilize numerous datasets for training, the datasets need filtering because they often include noisy labels. In this thesis, we implement basic data augmentation and filtering techniques as a Splash language construct, called data preprocessor, to support data engineers.
Data augmentation function in Splash data preprocessor increases dataset quantity by using seven augmentation techniques: horizontal and vertical shift, horizontal and vertical flip, random rotation, random brightness, and random zoom. The data filtering function finds duplicated images with different and same labels, then removes those images to improve the quality of the training dataset. To demonstrate the feasibility of using Splash data preprocessor and to confirm the correctness of the data preprocessor implementation, we trained the CIFAR-10 dataset as an experiment using Splash data preprocessor. This experiment shows that training data filtering and augmentation can be easily performed using the Splash data preprocessor.Splashλ μΈκ³΅ μ§λ₯ μμ© κ°λ°μ μ§μνκΈ° μν΄ λ§λ€μ΄μ§ GUI νλ‘κ·Έλλ° νλ μμν¬μ΄λ€. Splashλ νλ‘κ·Έλλ° μΆμνλ₯Ό ν΅ν΄ λ°μ΄ν°, AI λͺ¨λΈλ§, μ μ΄ μμ§λμ΄λ₯Ό ν¬ν¨ν μ¬λ¬ λΆμΌ μ λ¬Έκ°λ€μ΄ νλ‘κ·Έλλ°μ μ§μ μμ΄λ μμ½κ² μ¬μ©ν μ μλλ‘ λ§λ€μ΄μ‘λ€. μΈκ³΅ μ§λ₯ μμ© κ°λ°μ μ§μνλ Splashμ κΈ°λ₯μ λμ± ν₯μμν€κΈ° μνμ¬ λ°μ΄ν° μ μ²λ¦¬ κΈ°λ₯μ Splashμ μΈμ΄ κ΅¬μ‘°λ‘ μΆκ°νμλ€. μ΄ μΈμ΄ ꡬ쑰λ λ°μ΄ν° μμ§λμ΄μ μ£Όμ μ
λ¬΄μΈ λ°μ΄ν° μ μ²λ¦¬ μ€ λ°μ΄ν° νν°λ§κ³Ό μ¦κ° κΈ°λ₯μ μ§μνλ€.
μ§λ νμ΅(supervised learning)μμ λ°μ΄ν° νν°λ§κ³Ό μ¦κ°μ νΉν μ€μν μμ
μ΄λ€. μ§λνμ΅μ μν΄μλ λ μ΄λΈμ΄ λμ΄μλ λ°μ΄ν°κ° νμνλ°, μ½κ² ꡬν μ μλ MNISTμ κ°μ νμ΅ λ°μ΄ν°μ
μ΄λ μ§μ λ μ΄λΈλ§ ν λ°μ΄ν°μ
μ μκ° νμ μ μ΄λ€. λ°λΌμ λ°μ΄ν°μ μλ₯Ό μ¦κ°μν€κΈ° μνμ¬ λ°μ΄ν° μ¦κ° κΈ°μ μ΄ νμνλ€. λ§μ μμ λ°μ΄ν°μ
μ νμ©νκΈ° μν΄μ ν¬λΌμ°λμμ± κ°μ λ°μ΄ν° λ μ΄λΈ νλ«νΌμ΄λ μλ λ μ΄λΈ νλ‘κ·Έλ¨μ μ΄μ©νλ κ²½μ°, λ μ΄λΈμ΄ μλͺ»λμ΄ μλ κ²½μ°κ° λ§κΈ° λλ¬Έμ μ΄λ₯Ό νν°λ§ν΄μΌ νλ€. λ³Έ λ
Όλ¬Έμμλ μ§λ νμ΅μμ νμν κΈ°λ³Έμ μΈ λ°μ΄ν° νν°λ§ κΈ°λ²κ³Ό λ°μ΄ν° μ¦κ° κΈ°λ²μ Splashμ ꡬννμ¬ λ°μ΄ν° μμ§λμ΄κ° μμ½κ² μ΄μ©ν μ μλλ‘ νλ€. Splash λ°μ΄ν° μ μ²λ¦¬ μ°μ°μλ μ΄λ―Έμ§μ μ€λ³΅μ±μ νλ¨νμ¬ νν°λ§νκ³ , μΌκ³± κ°μ§ λ°©λ²μΌλ‘ μ΄λ―Έμ§λ₯Ό μ¦κ°μν¨λ€. μ°λ¦¬λ Splash λ°μ΄ν° μ μ²λ¦¬ μ°μ°μλ₯Ό μ¬μ©νμ¬ μ§λ νμ΅ λ°μ΄ν° νν°λ§ λ° μ¦κ°μ μ½κ² μν ν μ μμμ 보μλ€.Chapter 1. Introduction 1
Chapter 2. Splash programming language 4
Chapter 3. Splash data preprocessor 9
Chapter 4. Splash data preprocessor experiment 14
Chapter 5. Conclusion 18
References 19
Abstract in Korean 21μ
Model and Data Agreement for Learning with Noisy Labels
Learning with noisy labels is a vital topic for practical deep learning as
models should be robust to noisy open-world datasets in the wild. The
state-of-the-art noisy label learning approach JoCoR fails when faced with a
large ratio of noisy labels. Moreover, selecting small-loss samples can also
cause error accumulation as once the noisy samples are mistakenly selected as
small-loss samples, they are more likely to be selected again. In this paper,
we try to deal with error accumulation in noisy label learning from both model
and data perspectives. We introduce mean point ensemble to utilize a more
robust loss function and more information from unselected samples to reduce
error accumulation from the model perspective. Furthermore, as the flip images
have the same semantic meaning as the original images, we select small-loss
samples according to the loss values of flip images instead of the original
ones to reduce error accumulation from the data perspective. Extensive
experiments on CIFAR-10, CIFAR-100, and large-scale Clothing1M show that our
method outperforms state-of-the-art noisy label learning methods with different
levels of label noise. Our method can also be seamlessly combined with other
noisy label learning methods to further improve their performance and
generalize well to other tasks. The code is available in
https://github.com/zyh-uaiaaaa/MDA-noisy-label-learning.Comment: Accepted by AAAI2023 Worksho
Meta Soft Label Generation for Noisy Labels
The existence of noisy labels in the dataset causes significant performance
degradation for deep neural networks (DNNs). To address this problem, we
propose a Meta Soft Label Generation algorithm called MSLG, which can jointly
generate soft labels using meta-learning techniques and learn DNN parameters in
an end-to-end fashion. Our approach adapts the meta-learning paradigm to
estimate optimal label distribution by checking gradient directions on both
noisy training data and noise-free meta-data. In order to iteratively update
soft labels, meta-gradient descent step is performed on estimated labels, which
would minimize the loss of noise-free meta samples. In each iteration, the base
classifier is trained on estimated meta labels. MSLG is model-agnostic and can
be added on top of any existing model at hand with ease. We performed extensive
experiments on CIFAR10, Clothing1M and Food101N datasets. Results show that our
approach outperforms other state-of-the-art methods by a large margin.Comment: Accepted by ICPR 202
- β¦