702 research outputs found

    On the class overlap problem in imbalanced data classification.

    Get PDF
    Class imbalance is an active research area in the machine learning community. However, existing and recent literature showed that class overlap had a higher negative impact on the performance of learning algorithms. This paper provides detailed critical discussion and objective evaluation of class overlap in the context of imbalanced data and its impact on classification accuracy. First, we present a thorough experimental comparison of class overlap and class imbalance. Unlike previous work, our experiment was carried out on the full scale of class overlap and an extreme range of class imbalance degrees. Second, we provide an in-depth critical technical review of existing approaches to handle imbalanced datasets. Existing solutions from selective literature are critically reviewed and categorised as class distribution-based and class overlap-based methods. Emerging techniques and the latest development in this area are also discussed in detail. Experimental results in this paper are consistent with existing literature and show clearly that the performance of the learning algorithm deteriorates across varying degrees of class overlap whereas class imbalance does not always have an effect. The review emphasises the need for further research towards handling class overlap in imbalanced datasets to effectively improve learning algorithmsโ€™ performance

    Understanding the apparent superiority of over-sampling through an analysis of local information for class-imbalanced data

    Get PDF
    Data plays a key role in the design of expert and intelligent systems and therefore, data preprocessing appears to be a critical step to produce high-quality data and build accurate machine learning models. Over the past decades, increasing attention has been paid towards the issue of class imbalance and this is now a research hotspot in a variety of fields. Although the resampling methods, either by under-sampling the majority class or by over-sampling the minority class, stand among the most powerful techniques to face this problem, their strengths and weaknesses have typically been discussed based only on the class imbalance ratio. However, several questions remain open and need further exploration. For instance, the subtle differences in performance between the over- and under-sampling algorithms are still under-comprehended, and we hypothesize that they could be better explained by analyzing the inner structure of the data sets. Consequently, this paper attempts to investigate and illustrate the effects of the resampling methods on the inner structure of a data set by exploiting local neighborhood information, identifying the sample types in both classes and analyzing their distribution in each resampled set. Experimental results indicate that the resampling methods that produce the highest proportion of safe samples and the lowest proportion of unsafe samples correspond to those with the highest overall performance. The significance of this paper lies in the fact that our findings may contribute to gain a better understanding of how these techniques perform on class-imbalanced data and why over-sampling has been reported to be usually more efficient than under-sampling. The outcomes in this study may have impact on both research and practice in the design of expert and intelligent systems since a priori knowledge about the internal structure of the imbalanced data sets could be incorporated to the learning algorithms

    ๋”ฅ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์˜ ๋น ๋ฅด๊ณ  ์ •ํ™•ํ•œ ์ •๋ณด ์ „๋‹ฌ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ „๊ธฐยท์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2021. 2. ์œค์„ฑ๋กœ.์˜ค๋Š˜ ๋‚  ๋”ฅ๋Ÿฌ๋‹์˜ ํฐ ์„ฑ๊ณต์€ ๊ณ ์„ฑ๋Šฅ ๋ณ‘๋ ฌ ์ปดํ“จํŒ… ์‹œ์Šคํ…œ์˜ ๋ฐœ์ „๊ณผ ๋ณต์žกํ•œ ๋ชจ๋ธ์„ ํ•™์Šตํ•˜๊ธฐ ์œ„ํ•ด ํ•„์š”ํ•œ ๋งŽ์€ ์–‘์˜ ๋ฐ์ดํ„ฐ๊ฐ€ ์ˆ˜์ง‘๋˜์–ด ์ ‘๊ทผ์ด ๊ฐ€๋Šฅํ•ด์ง„ ์ ์ด๋ผ๊ณ  ํ•  ์ˆ˜ ์žˆ๋‹ค. ํ•˜์ง€๋งŒ ์‹ค์ œ ์„ธ์ƒ์— ์กด์žฌํ•˜๋Š” ๋” ์–ด๋ ค์šด ๋ฌธ์ œ๋“ค์„ ํ’€๊ณ ์žํ•  ๋•Œ๋Š” ๋”์šฑ ๋” ์„ฌ์„ธํ•˜๊ณ  ๋ณต์žกํ•œ ๋ชจ๋ธ๊ณผ ์ด ๋ชจ๋ธ์„ ์„ฑ๊ณต์ ์œผ๋กœ ํ•™์Šตํ•  ์ˆ˜ ์žˆ๋Š” ๋ฐฉ๋Œ€ํ•œ ์–‘์˜ ๋ฐ์ดํ„ฐ๋ฅผ ํ•„์š”ํ•œ๋‹ค. ํ•˜์ง€๋งŒ ์ด๋Ÿฌํ•œ ์ ๋“ค์€ ๋ชจ๋ธ ์ˆ˜ํ–‰ ์‹œ ์—ฐ์‚ฐ ์˜ค๋ฒ„ํ—ค๋“œ์™€ ์ „๋ ฅ ์†Œ๋ชจ๋ฅผ ๊ธ‰๊ฒฉํ•˜๊ฒŒ ์ฆ๊ฐ€์‹œํ‚ฌ ์ˆ˜ ๋ฐ–์— ์—†๋‹ค. ์ด๋Ÿฌํ•œ ๋ฌธ์ œ์ ๋“ค์„ ๊ทน๋ณตํ•˜๋Š” ์—ฌ๋Ÿฌ ๋ฐฉ๋ฒ•๋“ค ์ค‘ ํ•˜๋‚˜๋กœ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๊ฐ€ ์ตœ๊ทผ ๋งŽ์€ ์ฃผ๋ชฉ์„ ๋ฐ›๊ณ  ์žˆ๋‹ค. ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๋Š” ์ œ 3์„ธ๋Œ€ ์ธ๊ณต ์‹ ๊ฒฝ๋ง์œผ๋กœ ๋ถˆ๋ฆฌ๋ฉฐ ์ด๋ฒคํŠธ ์ค‘์‹ฌ์˜ ๋™์ž‘์„ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•˜์—ฌ ์ €์ „๋ ฅ์ด ๊ฐ€์žฅ ํฐ ์žฅ์ ์ด๋‹ค. ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๋Š” ์‹ค์ œ ์ธ๊ฐ„์˜ ๋‡Œ์—์„œ ๋‰ด๋Ÿฐ๋“ค ๊ฐ„ ์ •๋ณด๋ฅผ ์ „๋‹ฌํ•˜๋Š” ๋ฐฉ์‹์„ ๋ชจ๋ฐฉํ•˜๋ฉฐ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿฐ์„ ์—ฐ์‚ฐ ๋‹จ์œ„๋กœ ์‚ฌ์šฉํ•˜๊ณ  ์žˆ๋‹ค. ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๋Š” ์ƒ๋ฌผํ•™์  ์‹ ๊ฒฝ๊ณ„์™€ ๋™์ผํ•˜๊ฒŒ ์‹œ๊ฐ„์  ์ •๋ณด๋ฅผ ํ™œ์šฉํ•  ์ˆ˜ ์žˆ๊ธฐ ๋•Œ๋ฌธ์— ๋งค์šฐ ๋›ฐ์–ด๋‚œ ์—ฐ์‚ฐ ๋Šฅ๋ ฅ์„ ๊ฐ€์ง€๊ณ  ์žˆ๋‹ค. ํ•˜์ง€๋งŒ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๋Š” ์ด๋ฏธ์ง€ ๋ถ„๋ฅ˜์™€ ๊ฐ™์€ ๋น„๊ต์  ์‰ฌ์šด ์‘์šฉ์—๋งŒ ์ฃผ๋กœ ์‚ฌ์šฉ๋˜๊ณ  ์žˆ์œผ๋ฉฐ ์–•์€ ์ธ๊ณต ์‹ ๊ฒฝ๋ง๊ณผ ๊ฐ„๋‹จํ•œ ๋ฐ์ดํ„ฐ์…‹์—์„œ๋งŒ ์ฃผ๋กœ ์ˆ˜ํ–‰๋˜๊ณ  ์žˆ๋‹ค. ์ด๋Ÿฌํ•œ ์ œ์•ฝ์ด ์กด์žฌํ•˜๋Š” ๊ฐ€์žฅ ํฐ ์š”์ธ ์ค‘ ํ•˜๋‚˜๋Š” ์ŠคํŒŒ์ดํฌ ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์— ์ ํ•ฉํ•œ ํ•™์Šต ์•Œ๊ณ ๋ฆฌ์ฆ˜์ด ์•„์ง ์กด์žฌํ•˜์ง€ ์•Š๊ธฐ ๋•Œ๋ฌธ์ด๋‹ค. ์ŠคํŒŒ์ดํฌ๋กœ ์ •๋ณด๋ฅผ ์ „๋‹ฌํ•˜๊ณ  ์—ฐ์‚ฐ์„ ์ˆ˜ํ–‰ํ•˜๊ธฐ ๋•Œ๋ฌธ์— ๋ฏธ๋ถ„์ด ๋ถˆ๊ฐ€๋Šฅํ•˜๋‹ค. ๋”ฐ๋ผ์„œ ๋”ฅ ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์—์„œ ์ฃผ๋กœ ์‚ฌ์šฉ๋˜๋Š” ์—ญ์ „ํŒŒ ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ์‚ฌ์šฉ์ด ๋ถˆ๊ฐ€๋Šฅํ•˜๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ ๋”ฅ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๋ฅผ ์ด๋ฏธ์ง€ ๋ถ„๋ฅ˜๋ณด๋‹ค ๋” ์–ด๋ ค์šด ํšŒ๊ท€ ๋ฌธ์ œ (๊ฐ์ฒด ์ธ์‹)์— ์ ์šฉํ•ด ๋ณด๊ณ , ๋”ฅ ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์˜ ์„ฑ๋Šฅ์— ๋ฒ„๊ธˆ๊ฐ€๋Š” ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ์„ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œ์—์„œ ์ฒ˜์Œ์œผ๋กœ ์ œ์•ˆํ•œ๋‹ค. ๋” ๋‚˜์•„๊ฐ€, ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ์˜ ์„ฑ๋Šฅ๊ณผ ์ง€์—ฐ์‹œ๊ฐ„, ์—๋„ˆ์ง€ ํšจ์œจ์„ฑ์„ ํ–ฅ์ƒ ์‹œํ‚ฌ ์ˆ˜ ์žˆ๋Š” ์—ฌ๋Ÿฌ ๋ฐฉ๋ฒ•๋“ค์„ ์ œ์•ˆํ•œ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์€ ํฌ๊ฒŒ ๋‘ ๊ฐ€์ง€ ์ฃผ์ œ๋กœ ๋‚˜๋ˆ„์–ด ์„ค๋ช…ํ•œ๋‹ค: (a) ๋”ฅ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์—์„œ์˜ ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ, (b) ๋”ฅ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์—์„œ์˜ ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ์˜ ์„ฑ๋Šฅ ๋ฐ ํšจ์œจ์„ฑ ํ–ฅ์ƒ. ์ œ์•ˆํ•˜๋Š” ๋ฐฉ๋ฒ•๋“ค์„ ํ†ตํ•ด ๋น ๋ฅด๊ณ  ์ •ํ™•ํ•œ ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ์„ ๋”ฅ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์—์„œ ์ˆ˜ํ–‰ํ•  ์ˆ˜ ์žˆ๋‹ค. ์ฒซ ๋ฒˆ์งธ ๋ฐฉ๋ฒ•์€ ๋”ฅ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์—์„œ์˜ ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ์ด๋‹ค. ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ์€ Spiking-YOLO๋กœ ๋ถ€๋ฅด๊ณ , ์ €์ž๋“ค์ด ์•„๋Š” ๋ฐ”์— ์˜ํ•˜๋ฉด PASCAL VOC, MS COCO์™€ ๊ฐ™์€ ๋ฐ์ดํ„ฐ ์…‹์—์„œ ๋”ฅ ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์˜ ์„ฑ๋Šฅ์— ๋ฒ„๊ธˆ๊ฐ€๋Š” ๊ฒฐ๊ณผ๋ฅผ ๋ณด์—ฌ์ค€ ์ฒซ ๋ฒˆ์งธ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•˜๋Š” ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ์ด๋‹ค. Spiking-YOLO์—์„œ๋Š” ํฌ๊ฒŒ ๋‘ ๊ฐ€์ง€ ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•œ๋‹ค. ์ฒซ๋ฒˆ ์งธ๋Š” ์ฑ„๋„ ๋ณ„ ๊ฐ€์ค‘์น˜ ์ •๊ทœํ™”์ด๊ณ  ๋‘๋ฒˆ์งธ๋Š” ๋ถˆ๊ท ํ˜• ํ•œ๊ณ„ ์ „์••์„ ๊ฐ€์ง€๋Š” ์–‘์Œ์ˆ˜ ๋‰ด๋Ÿฐ์ด๋‹ค. ๋‘ ๊ฐ€์ง€ ๋ฐฉ๋ฒ•์„ ํ†ตํ•ด ๋น ๋ฅด๊ณ  ์ •ํ™•ํ•œ ์ •๋ณด๋ฅผ ๋”ฅ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์—์„œ ์ „๋‹ฌ ๊ฐ€๋Šฅํ•˜๊ฒŒ ํ•œ๋‹ค. ์‹คํ—˜ ๊ฒฐ๊ณผ, Spiking-YOLO๋Š” PASCAL VOC์™€ MS COCO ๋ฐ์ดํ„ฐ์…‹์—์„œ ๋”ฅ ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์˜ ๊ฐ์ฒด ์ธ์‹๋ฅ ์˜ 98%์— ๋›ฐ์–ด๋‚œ ์„ฑ๋Šฅ์„ ๋ณด์˜€๋‹ค. ๋˜ํ•œ Spiking-YOLO๊ฐ€ ๋‰ด๋กœ๋ชจํ”ฝ ์นฉ์— ๊ตฌํ˜„๋˜์—ˆ์Œ ๊ฐ€์ •ํ•˜์˜€์„ ๋•Œ, Tiny YOLO๋ณด๋‹ค ์•ฝ 280์˜ ์—๋„ˆ์ง€๋ฅผ ์ ๊ฒŒ ์†Œ๋ชจํ•˜์˜€๊ณ  ๊ธฐ์กด์˜ DNN-to-SNN ์ „ํ™˜ ๋ฐฉ๋ฒ•๋“ค ๋ณด๋‹ค 2.3๋ฐฐ์—์„œ 4๋ฐฐ ๋” ๋น ๋ฅด๊ฒŒ ์ˆ˜๋ ดํ•˜๋Š” ๊ฒƒ์„ ํ™•์ธํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ๋‘ ๋ฒˆ์งธ ๋ฐฉ๋ฒ•์€ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์— ์กฐ๊ธˆ ๋” ํšจ์œจ์ ์ธ ์—ฐ์‚ฐ ๋Šฅ๋ ฅ์„ ๋ถ€์—ฌํ•˜๋Š”๋ฐ ์ค‘์ ์„ ์ฃผ๊ณ  ์žˆ๋‹ค. ๋น„๋ก ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ๊ฐ€ ํฌ๋ฐ•ํ•œ ์–‘์˜ ์ŠคํŒŒ์ดํฌ๋กœ ์ •๋ณด๋ฅผ ํšจ์œจ์ ์œผ๋กœ ์ „๋‹ฌํ•˜๋ฉฐ ์—ฐ์‚ฐ ์˜ค๋ฒ„ํ—ค๋“œ์™€ ์—๋„ˆ์ง€ ์†Œ๋ชจ๊ฐ€ ์ ์ง€๋งŒ, ๋‘ ๊ฐ€์ง€ ๋งค์šฐ ์ค‘์š”ํ•œ ๋ฌธ์ œ๋“ค์ด ์กด์žฌํ•œ๋‹ค: (a) ์ง€์—ฐ์†๋„: ์ข‹์€ ์„ฑ๋Šฅ์„ ๋‚ด๊ธฐ ์œ„ํ•ด ํ•„์š”ํ•œ ํƒ€์ž„์Šคํƒญ, (b) ์‹œ๋ƒ…ํ‹ฑ ์—ฐ์‚ฐ์ˆ˜: ์ถ”๋ก  ์‹œ ์ƒ์„ฑ๋œ ์ด ์ŠคํŒŒ์ดํฌ์˜ ์ˆ˜. ์ด๋Ÿฌํ•œ ๋ฌธ์ œ๋“ค์„ ์ ์ ˆํžˆ ํ•ด๊ฒฐํ•˜์ง€ ๋ชปํ•œ๋‹ค๋ฉด ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์˜ ํฐ ์žฅ์ ์ด๋ผ๊ณ  ํ•  ์ˆ˜ ์žˆ๋Š” ์—๋„ˆ์ง€์™€ ์ „๋ ฅ ํšจ์œจ์„ฑ์ด ํฌ๊ฒŒ ์ €ํ•˜๋  ์ˆ˜ ์žˆ๋‹ค. ์ด๋ฅผ ํ•ด๊ฒฐํ•˜๊ธฐ ์œ„ํ•ด ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ํ•œ๊ณ„ ์ „์•• ๊ท ํ˜• ๋ฐฉ๋ฒ•๋ก ์„ ์ƒˆ๋กœ ์ œ์•ˆํ•œ๋‹ค. ์ œ์•ˆํ•˜๋Š” ๋ฐฉ๋ฒ•๋ก ์€ ๋ฒ ์ด์‹œ์•ˆ ์ตœ์ ํ™” ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์‚ฌ์šฉํ•˜์—ฌ ๊ฐ€์žฅ ์ตœ์ ์˜ ํ•œ๊ณ„์ „์•• ๊ฐ’์„ ์ฐพ๋Š”๋‹ค. ๋˜ํ•œ ๋ฒ ์ด์‹œ์•ˆ ์ตœ์ ํ™” ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ง€์—ฐ์†๋„๋‚˜ ์‹œ๋ƒ…ํ‹ฑ ์—ฐ์‚ฐ์ˆ˜ ๋“ฑ์˜ ์ŠคํŒŒ์ดํ‚น ๋‰ด๋Ÿด ๋„คํŠธ์›Œํฌ์˜ ํŠน์„ฑ์„ ๊ณ ๋ คํ•  ์ˆ˜ ์žˆ๊ฒŒ ๋””์ž์ธํ•œ๋‹ค. ๋” ๋‚˜์•„๊ฐ€, ๋‘ ๋‹จ๊ณ„์˜ ํ•œ๊ณ„ ์ „์••์„ ์ œ์•ˆํ•˜์—ฌ ๋†’์€ ์—๋„ˆ์ง€ ํšจ์œจ์„ ๊ฐ€์ง€๋ฉฐ ๋” ๋น ๋ฅด๊ณ  ๋” ์ •ํ™•ํ•œ ๊ฐ์ฒด ์ธ์‹ ๋ชจ๋ธ์„ ๊ฐ€๋Šฅํ•˜๊ฒŒ ํ•œ๋‹ค. ์‹คํ—˜ ๊ฒฐ๊ณผ์— ๋”ฐ๋ฅด๋ฉด ์ œ์•ˆํ•˜๋Š” ๋ฐฉ๋ฒ•๋“ค์„ ํ†ตํ•ด state-of-the-art ๊ฐ์ฒด ์ธ์‹๋ฅ ์„ ๋‹ฌ์„ฑํ•˜์˜€๊ณ  ๊ธฐ์กด์˜ ๋ฐฉ๋ฒ•๋“ค๋ณด๋‹ค PASCAL VOC์—์„œ๋Š” 2๋ฐฐ, MS COCO์—์„œ๋Š” 1.85๋ฐฐ ๋น ๋ฅด๊ฒŒ ์ˆ˜๋ ดํ•˜๋Š” ๊ฒƒ์„ ํ™•์ธํ•˜์˜€๋‹ค. ๋˜ํ•œ ์‹œ๋ƒ…ํ‹ฑ ์—ฐ์‚ฐ์ˆ˜๋„ PASCAL VOC์—์„œ๋Š” 40.33%, MS COCO์—์„œ๋Š” 45.31%๋ฅผ ์ค„์ผ ์ˆ˜ ์žˆ์—ˆ๋‹ค.One of the primary reasons behind the recent success of deep neural networks (DNNs) lies in the development of high-performance parallel computing systems and the availability of enormous amounts of data for training a complex model. Nonetheless, solving such advanced machine learning problems in real world applications requires a more sophisticated model with a vast number of parameters and training data, which leads to substantial amounts of computational overhead and power consumption. Given these circumstances, spiking neural networks (SNNs) have attracted growing interest as the third generation of neural networks due to their event-driven and low-powered nature. SNNs were introduced to mimic how information is encoded and processed in the human brain by employing spiking neurons as computation units. SNNs utilize temporal aspects in information transmission as in biological neural systems, thus providing sparse yet powerful computing ability. SNNs have been successfully applied in several applications, but these applications only include relatively simple tasks such as image classification, and are limited to shallow neural networks and datasets. One of the primary reasons for the limited application scope is the lack of scalable training algorithms attained from non-differential spiking neurons. In this dissertation, we investigate deep SNNs in a much more challenging regression problem (i.e., object detection), and propose a first object detection model in deep SNNs which is able to achieve comparable results to those of DNNs in non-trivial datasets. Furthermore, we introduce novel approaches to improve performance of the object detection model in terms of accuracy, latency and energy efficiency. This dissertation contains mainly two approaches: (a) object detection model in deep SNNs, and (b) improving performance of object detection model in deep SNNs. Consequently, the two approaches enable fast and accurate object detection in deep SNNs. The first approach is an object detection model in deep SNNs. We present a spiked-based object detection model, called Spiking-YOLO. To the best of our knowledge, Spiking-YOLO is the first spiked-based object detection model that is able to achieve comparable results to those of DNNs on a non-trivial dataset, namely PASCAL VOC and MS COCO. In doing so, we introduce two novel methods: a channel-wise weight normalization and a signed neuron with imbalanced threshold, both of which provide fast and accurate information transmission in deep SNNs. Our experiments show that Spiking-YOLO achieves remarkable results that are comparable (up to 98%) to those of Tiny YOLO (DNNs) on PASCAL VOC and MS COCO. Furthermore, Spiking-YOLO on a neuromorphic chip consumes approximately 280 times less energy than Tiny YOLO, and converges 2.3 to 4 times faster than previous DNN-to-SNN conversion methods. The second approach aims to provide a more effective form of computational capabilities in SNNs. Even though, SNNs enable sparse yet efficient information transmission through spike trains, leading to exceptional computational and energy efficiency, the critical challenges in SNNs to date are two-fold: (a) latency: the number of time steps required to achieve competitive results and (b) synaptic operations: the total number of spikes generated during inference. Without addressing these challenges properly, the potential impact of SNNs may be diminished in terms of energy and power efficiency. We present a threshold voltage balancing method for object detection in SNNs, which utilizes Bayesian optimization to find optimal threshold voltages in SNNs. We specifically design Bayesian optimization to consider important characteristics of SNNs, such as latency and number of synaptic operations. Furthermore, we introduce two-phase threshold voltages to provide faster and more accurate object detection, while providing high energy efficiency. According to experimental results, the proposed methods achieve the state-of-the-art object detection accuracy in SNNs, and converge 2x and 1.85x faster than conventional methods on PASCAL VOC and MS COCO, respectively. Moreover, the total number of synaptic operations is reduced by 40.33% and 45.31% on PASCAL VOC and MS COCO, respectively.Abstract i List of Figures ix List of Tables x 1 Introduction 1 2 Background 10 2.1 Object detection 10 2.2 Spiking Neural Networks 16 2.3 DNN-to-SNN conversion 18 2.4 Hyper-parameter optimization 21 3 Object detection model in deep SNNs 25 3.1 Introduction 25 3.2 Channel-wise weight normalization 27 3.2.1 Conventional weight normalization methods 27 3.2.2 Analysis of limitations in layer-wise weight normalization 29 3.2.3 Proposed weight normalization method 30 3.2.4 Analysis of the improved firing rate 38 3.3 Signed neuron with imbalanced threshold 39 3.3.1 Limitation of leaky-ReLU implementation in SNNs 39 3.3.2 The notion of imbalanced threshold 41 3.4 Evaluation 43 3.4.1 Spiking-YOLO detection results 43 3.4.2 Spiking-YOLO energy efficiency 57 4 Improving performance and efficiency of deep SNNs 60 4.1 Introduction 60 4.2 Threshold voltage balancing through Bayesian optimization 62 4.2.1 Motivation 62 4.2.2 Overall process and setup 67 4.2.3 Design of Bayesian optimization for SNNs 69 4.3 Fast and accurate object detection with two-phase threshold voltages 74 4.3.1 Motivation 74 4.3.2 Phase-1 threshold voltages: fast object detection 76 4.3.3 Phase-2 threshold voltages: accurate detection 76 4.4 Evaluation 79 4.4.1 Experimental setup 79 4.4.2 Experimental results 79 5 Conclusion 85 5.1 Dissertation summary 86 5.2 Discussion 88 5.2.1 Overview of the proposed methods and their usages 88 5.3 Challenges in SNNs 90 5.4 Future Work 92 5.4.1 Extension to various applications and DNN models 92 5.4.2 Further improve efficiency of SNNs 93 5.4.3 Optimization of deep SNNs 94 Bibliography 95 Abstract (In Korean) 110Docto

    MRL-Seg: Overcoming Imbalance in Medical Image Segmentation With Multi-Step Reinforcement Learning

    Get PDF
    Medical image segmentation is a critical task for clinical diagnosis and research. However, dealing with highly imbalanced data remains a significant challenge in this domain, where the region of interest (ROI) may exhibit substantial variations across different slices. This presents a significant hurdle to medical image segmentation, as conventional segmentation methods may either overlook the minority class or overly emphasize the majority class, ultimately leading to a decrease in the overall generalization ability of the segmentation results. To overcome this, we propose a novel approach based on multi-step reinforcement learning, which integrates prior knowledge of medical images and pixel-wise segmentation difficulty into the reward function. Our method treats each pixel as an individual agent, utilizing diverse actions to evaluate its relevance for segmentation. To validate the effectiveness of our approach, we conduct experiments on four imbalanced medical datasets, and the results show that our approach surpasses other state-of-the-art methods in highly imbalanced scenarios. These findings hold substantial implications for clinical diagnosis and research

    Constrained Twin Variational Auto-Encoder for Intrusion Detection in IoT Systems

    Full text link
    Intrusion detection systems (IDSs) play a critical role in protecting billions of IoT devices from malicious attacks. However, the IDSs for IoT devices face inherent challenges of IoT systems, including the heterogeneity of IoT data/devices, the high dimensionality of training data, and the imbalanced data. Moreover, the deployment of IDSs on IoT systems is challenging, and sometimes impossible, due to the limited resources such as memory/storage and computing capability of typical IoT devices. To tackle these challenges, this article proposes a novel deep neural network/architecture called Constrained Twin Variational Auto-Encoder (CTVAE) that can feed classifiers of IDSs with more separable/distinguishable and lower-dimensional representation data. Additionally, in comparison to the state-of-the-art neural networks used in IDSs, CTVAE requires less memory/storage and computing power, hence making it more suitable for IoT IDS systems. Extensive experiments with the 11 most popular IoT botnet datasets show that CTVAE can boost around 1% in terms of accuracy and Fscore in detection attack compared to the state-of-the-art machine learning and representation learning methods, whilst the running time for attack detection is lower than 2E-6 seconds and the model size is lower than 1 MB. We also further investigate various characteristics of CTVAE in the latent space and in the reconstruction representation to demonstrate its efficacy compared with current well-known methods

    A Deep Unsupervised Learning Approach for Airspace Complexity Evaluation

    Get PDF
    Airspace complexity is a critical metric in current Air Traffic Management systems for indicating the security degree of airspace operations. Airspace complexity can be affected by many coupling factors in a complicated and nonlinear way, making it extremely difficult to be evaluated. In recent years, machine learning has been proved as a promising approach and achieved significant results in evaluating airspace complexity. However, existing machine learning based approaches require a large number of airspace operational data labeled by experts. Due to the high cost in labeling the operational data and the dynamical nature of the airspace operating environment, such data are often limited and may not be suitable for the changing airspace situation. In light of these, we propose a novel unsupervised learning approach for airspace complexity evaluation based on a deep neural network trained by unlabeled samples. We introduce a new loss function to better address the characteristics pertaining to airspace complexity data, including dimension coupling, category imbalance, and overlapped boundaries. Due to these characteristics, the generalization ability of existing unsupervised models is adversely impacted. The proposed approach is validated through extensive experiments based on the real-world data of six sectors in Southwestern China airspace. Experimental results show that our deep unsupervised model outperforms the state-of-the-art methods in terms of airspace complexity evaluation accuracy

    SMOTE for Learning from Imbalanced Data: Progress and Challenges, Marking the 15-year Anniversary

    Get PDF
    The Synthetic Minority Oversampling Technique (SMOTE) preprocessing algorithm is considered \de facto" standard in the framework of learning from imbalanced data. This is due to its simplicity in the design of the procedure, as well as its robustness when applied to di erent type of problems. Since its publication in 2002, SMOTE has proven successful in a variety of applications from several di erent domains. SMOTE has also inspired several approaches to counter the issue of class imbalance, and has also signi cantly contributed to new supervised learning paradigms, including multilabel classi cation, incremental learning, semi-supervised learning, multi-instance learning, among others. It is standard benchmark for learning from imbalanced data. It is also featured in a number of di erent software packages | from open source to commercial. In this paper, marking the fteen year anniversary of SMOTE, we re ect on the SMOTE journey, discuss the current state of a airs with SMOTE, its applications, and also identify the next set of challenges to extend SMOTE for Big Data problems.This work have been partially supported by the Spanish Ministry of Science and Technology under projects TIN2014-57251-P, TIN2015-68454-R and TIN2017-89517-P; the Project 887 BigDaP-TOOLS - Ayudas Fundaci on BBVA a Equipos de Investigaci on Cient ca 2016; and the National Science Foundation (NSF) Grant IIS-1447795

    Empowering One-vs-One Decomposition with Ensemble Learning for Multi-Class Imbalanced Data

    Get PDF
    Zhongliang Zhang was supported by the National Science Foundation of China (NSFC Proj. 61273204) and CSC Scholarship Program (CSC NO. 201406080059). Bartosz Krawczyk was supported by the Polish National Science Center under the grant no. UMO-2015/19/B/ST6/01597. Salvador Garcia and Francisco Herrera were partially supported by the Spanish Ministry of Education and Science under Project TIN2014-57251-P and the Andalusian Research Plan P10-TIC-6858, P11-TIC-7765. Alejandro Rosales-Perez was supported by the CONACyT grant 329013.Multi-class imbalance classification problems occur in many real-world applications, which suffer from the quite different distribution of classes. Decomposition strategies are well-known techniques to address the classification problems involving multiple classes. Among them binary approaches using one-vs-one and one-vs-all has gained a significant attention from the research community. They allow to divide multi-class problems into several easier-to-solve two-class sub-problems. In this study we develop an exhaustive empirical analysis to explore the possibility of empowering the one-vs-one scheme for multi-class imbalance classification problems with applying binary ensemble learning approaches. We examine several state-of-the-art ensemble learning methods proposed for addressing the imbalance problems to solve the pairwise tasks derived from the multi-class data set. Then the aggregation strategy is employed to combine the binary ensemble outputs to reconstruct the original multi-class task. We present a detailed experimental study of the proposed approach, supported by the statistical analysis. The results indicate the high effectiveness of ensemble learning with one-vs-one scheme in dealing with the multi-class imbalance classification problems.National Natural Science Foundation of China (NSFC) 61273204CSC Scholarship Program (CSC) 201406080059Polish National Science Center UMO-2015/19/B/ST6/01597Spanish Government TIN2014-57251-PAndalusian Research Plan P10-TIC-6858 P11-TIC-7765Consejo Nacional de Ciencia y Tecnologia (CONACyT) 32901
    • โ€ฆ
    corecore