753 research outputs found

    FPGA acceleration of a quantized neural network for remote-sensed cloud detection

    Get PDF
    The capture and transmission of remote-sensed imagery for Earth observation is both computationally and bandwidth expensive. In the analyses of remote-sensed imagery in the visual band, atmospheric cloud cover can obstruct up to two-thirds of observations, resulting in costly imagery being discarded. Mission objectives and satellite operational details vary; however, assuming a cloud-free observation requirement, a doubling of useful data downlinked with an associated halving of delivery cost is possible through effective cloud detection. A minimal-resource, real-time inference neural network is ideally suited to perform automatic cloud detection, both for pre-processing captured images prior to transmission and preventing unnecessary images being taken by larger payload sensors. Much of the hardware complexity of modern neural network implementations resides in high-precision floating-point calculation pipelines. In recent years, research has been conducted in identifying quantized, or low-integer precision equivalents to known deep learning models, which do not require the extensive resources of their floating-point, full-precision counterparts. Our work leverages existing research on binary and quantized neural networks to develop a real-time, remote-sensed cloud detection solution using a commodity field-programmable gate array. This follows on developments of the Forwards Looking Imager for predictive cloud detection developed by Craft Prospect, a space engineering practice based in Glasgow, UK. The synthesized cloud detection accelerator achieved an inference throughput of 358.1 images per second with a maximum power consumption of 2.4 W. This throughput is an order of magnitude faster than alternate algorithmic options for the Forwards Looking Imager at around one third reduction in classification accuracy, and approximately two orders of magnitude faster than the CloudScout deep neural network, deployed with HyperScout 2 on the European Space Agency PhiSat-1 mission. Strategies for incorporating fault tolerance mechanisms are expounded

    Object Detection in 20 Years: A Survey

    Full text link
    Object detection, as of one the most fundamental and challenging problems in computer vision, has received great attention in recent years. Its development in the past two decades can be regarded as an epitome of computer vision history. If we think of today's object detection as a technical aesthetics under the power of deep learning, then turning back the clock 20 years we would witness the wisdom of cold weapon era. This paper extensively reviews 400+ papers of object detection in the light of its technical evolution, spanning over a quarter-century's time (from the 1990s to 2019). A number of topics have been covered in this paper, including the milestone detectors in history, detection datasets, metrics, fundamental building blocks of the detection system, speed up techniques, and the recent state of the art detection methods. This paper also reviews some important detection applications, such as pedestrian detection, face detection, text detection, etc, and makes an in-deep analysis of their challenges as well as technical improvements in recent years.Comment: This work has been submitted to the IEEE TPAMI for possible publicatio

    CNN๊ธฐ๋ฐ˜์˜ FusionNet ์‹ ๊ฒฝ๋ง๊ณผ ๋†์ง€ ๊ฒฝ๊ณ„์ถ”์ถœ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ด์šฉํ•œ ํ† ์ง€ํ”ผ๋ณต๋ถ„๋ฅ˜๋ชจ๋ธ ๊ฐœ๋ฐœ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (์„์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๋†์—…์ƒ๋ช…๊ณผํ•™๋Œ€ํ•™ ์ƒํƒœ์กฐ๊ฒฝ.์ง€์—ญ์‹œ์Šคํ…œ๊ณตํ•™๋ถ€(์ง€์—ญ์‹œ์Šคํ…œ๊ณตํ•™์ „๊ณต), 2021. 2. ์†ก์ธํ™.ํ† ์ง€์ด์šฉ์ด ๋น ๋ฅด๊ฒŒ ๋ณ€ํ™”ํ•จ์— ๋”ฐ๋ผ, ํ† ์ง€ ํ”ผ๋ณต์— ๋Œ€ํ•œ ๊ณต๊ฐ„์ •๋ณด๋ฅผ ๋‹ด๊ณ  ์žˆ๋Š” ํ† ์ง€ ํ”ผ๋ณต ์ง€๋„์˜ ์‹ ์†ํ•œ ์ตœ์‹ ํ™”๋Š” ํ•„์ˆ˜์ ์ด๋‹ค. ํ•˜์ง€๋งŒ, ํ˜„ ํ† ์ง€ ํ”ผ๋ณต ์ง€๋„๋Š” ๋งŽ์€ ์‹œ๊ฐ„๊ณผ ๋…ธ๋™๋ ฅ์„ ์š”๊ตฌํ•˜๋Š” manual digitizing ๋ฐฉ๋ฒ•์œผ๋กœ ์ œ์ž‘๋จ์— ๋”ฐ๋ผ, ํ† ์ง€ ํ”ผ๋ณต ์ง€๋„์˜ ์—…๋ฐ์ดํŠธ ๋ฐ ๋ฐฐํฌ์— ๊ธด ์‹œ๊ฐ„ ๊ฐ„๊ฒฉ์ด ๋ฐœ์ƒํ•˜๋Š” ์‹ค์ •์ด๋‹ค. ๋ณธ ์—ฐ๊ตฌ์—์„œ๋Š” convolutional neural network (CNN) ๊ธฐ๋ฐ˜์˜ ์ธ๊ณต์‹ ๊ฒฝ๋ง์„ ์ด์šฉํ•˜์—ฌ high-resolution remote sensing (HRRS) ์˜์ƒ์œผ๋กœ๋ถ€ํ„ฐ ํ† ์ง€ ํ”ผ๋ณต์„ ๋ถ„๋ฅ˜ํ•˜๋Š” ๋ชจ๋ธ์„ ๊ฐœ๋ฐœํ•˜๊ณ , ํŠนํžˆ ๋†์ง€ ๊ฒฝ๊ณ„์ถ”์ถœ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ ์šฉํ•˜์—ฌ ๋†์—…์ง€์—ญ์—์„œ ๋ถ„๋ฅ˜ ์ •ํ™•๋„๋ฅผ ๊ฐœ์„ ํ•˜๊ณ ์ž ํ•˜์˜€๋‹ค. ๊ฐœ๋ฐœ๋œ ํ† ์ง€ ํ”ผ๋ณต ๋ถ„๋ฅ˜๋ชจ๋ธ์€ ์ „์ฒ˜๋ฆฌ(pre-processing) ๋ชจ๋“ˆ, ํ† ์ง€ ํ”ผ๋ณต ๋ถ„๋ฅ˜(land cover classification) ๋ชจ๋“ˆ, ๊ทธ๋ฆฌ๊ณ  ํ›„์ฒ˜๋ฆฌ(post-processing) ๋ชจ๋“ˆ์˜ ์„ธ ๋ชจ๋“ˆ๋กœ ๊ตฌ์„ฑ๋œ๋‹ค. ์ „์ฒ˜๋ฆฌ ๋ชจ๋“ˆ์€ ์ž…๋ ฅ๋œ HRRS ์˜์ƒ์„ 75%์”ฉ ์ค‘์ฒฉ ๋ถ„ํ• ํ•˜์—ฌ ๊ด€์ ์„ ๋‹ค์–‘ํ™”ํ•˜๋Š” ๋ชจ๋“ˆ๋กœ, ํ•œ ๊ด€์ ์—์„œ ํ† ์ง€ ํ”ผ๋ณต์„ ๋ถ„๋ฅ˜ํ•  ๋•Œ ๋ฐœ์ƒํ•  ์ˆ˜ ์žˆ๋Š” ์˜ค๋ถ„๋ฅ˜๋ฅผ ์ค„์ด๊ณ ์ž ํ•˜์˜€๋‹ค. ํ† ์ง€ ํ”ผ๋ณต ๋ถ„๋ฅ˜ ๋ชจ๋“ˆ์€ FusionNet model ๊ตฌ์กฐ๋ฅผ ๋ฐ”ํƒ•์œผ๋กœ ๊ฐœ๋ฐœ๋˜์—ˆ๊ณ , ์ด๋Š” ๋ถ„ํ• ๋œ HRRS ์ด๋ฏธ์ง€์˜ ํ”ฝ์…€๋ณ„๋กœ ์ตœ์  ํ† ์ง€ ํ”ผ๋ณต์„ ๋ถ€์—ฌํ•˜๋„๋ก ์„ค๊ณ„๋˜์—ˆ๋‹ค. ํ›„์ฒ˜๋ฆฌ ๋ชจ๋“ˆ์€ ํ”ฝ์…€๋ณ„ ์ตœ์ข… ํ† ์ง€ ํ”ผ๋ณต์„ ๊ฒฐ์ •ํ•˜๋Š” ๋ชจ๋“ˆ๋กœ, ๋ถ„ํ• ๋œ HRRS ์ด๋ฏธ์ง€์˜ ๋ถ„๋ฅ˜๊ฒฐ๊ณผ๋ฅผ ์ทจํ•ฉํ•˜์—ฌ ์ตœ๋นˆ๊ฐ’์„ ์ตœ์ข… ํ† ์ง€ ํ”ผ๋ณต์œผ๋กœ ๊ฒฐ์ •ํ•œ๋‹ค. ์ถ”๊ฐ€๋กœ ๋†์ง€์—์„œ๋Š” ๋†์ง€๊ฒฝ๊ณ„๋ฅผ ์ถ”์ถœํ•˜๊ณ , ํ•„์ง€๋ณ„ ๋ถ„๋ฅ˜๋œ ํ† ์ง€ ํ”ผ๋ณต์„ ์ง‘๊ณ„ํ•˜์—ฌ ํ•œ ํ•„์ง€์— ๊ฐ™์€ ํ† ์ง€ ํ”ผ๋ณต์„ ๋ถ€์—ฌํ•˜์˜€๋‹ค. ๊ฐœ๋ฐœ๋œ ํ† ์ง€ ํ”ผ๋ณต ๋ถ„๋ฅ˜๋ชจ๋ธ์€ ์ „๋ผ๋‚จ๋„ ์ง€์—ญ(๋ฉด์ : 547 km2)์˜ 2018๋…„ ์ •์‚ฌ์˜์ƒ๊ณผ ํ† ์ง€ ํ”ผ๋ณต ์ง€๋„๋ฅผ ์ด์šฉํ•˜์—ฌ ํ•™์Šต๋˜์—ˆ๋‹ค. ํ† ์ง€ ํ”ผ๋ณต ๋ถ„๋ฅ˜๋ชจ๋ธ ๊ฒ€์ฆ์€ ํ•™์Šต์ง€์—ญ๊ณผ ์‹œ๊ฐ„, ๊ณต๊ฐ„์ ์œผ๋กœ ๊ตฌ๋ถ„๋œ, 2018๋…„ ์ „๋ผ๋‚จ๋„ ์ˆ˜๋ถ๋ฉด๊ณผ 2016๋…„ ์ถฉ์ฒญ๋ถ๋„ ๋Œ€์†Œ๋ฉด์˜ ๋‘ ๊ฒ€์ฆ์ง€์—ญ์—์„œ ์ˆ˜ํ–‰๋˜์—ˆ๋‹ค. ๊ฐ ๊ฒ€์ฆ์ง€์—ญ์—์„œ overall accuracy๋Š” 0.81, 0.71๋กœ ์ง‘๊ณ„๋˜์—ˆ๊ณ , kappa coefficients๋Š” 0.75, 0.64๋กœ ์‚ฐ์ •๋˜์–ด substantial ์ˆ˜์ค€์˜ ํ† ์ง€ ํ”ผ๋ณต ๋ถ„๋ฅ˜ ์ •ํ™•๋„๋ฅผ ํ™•์ธํ•˜์˜€๋‹ค. ํŠนํžˆ, ๊ฐœ๋ฐœ๋œ ๋ชจ๋ธ์€ ํ•„์ง€ ๊ฒฝ๊ณ„๋ฅผ ๊ณ ๋ คํ•œ ๋†์—…์ง€์—ญ์—์„œ overall accuracy 0.89, kappa coefficient 0.81๋กœ almost perfect ์ˆ˜์ค€์˜ ์šฐ์ˆ˜ํ•œ ๋ถ„๋ฅ˜ ์ •ํ™•๋„๋ฅผ ๋ณด์˜€๋‹ค. ์ด์— ๊ฐœ๋ฐœ๋œ ํ† ์ง€ ํ”ผ๋ณต ๋ถ„๋ฅ˜๋ชจ๋ธ์€ ํŠนํžˆ ๋†์—…์ง€์—ญ์—์„œ ํ˜„ ํ† ์ง€ ํ”ผ๋ณต ๋ถ„๋ฅ˜ ๋ฐฉ๋ฒ•์„ ์ง€์›ํ•˜์—ฌ ํ† ์ง€ ํ”ผ๋ณต ์ง€๋„์˜ ๋น ๋ฅด๊ณ  ์ •ํ™•ํ•œ ์ตœ์‹ ํ™”์— ๊ธฐ์—ฌํ•  ์ˆ˜ ์žˆ์„ ๊ฒƒ์œผ๋กœ ์ƒ๊ฐ๋œ๋‹ค.The rapid update of land cover maps is necessary because spatial information of land cover is widely used in various areas. However, these maps have been released or updated in the interval of several years primarily owing to the manual digitizing method, which is time-consuming and labor-intensive. This study was aimed to develop a land cover classification model using the concept of a convolutional neural network (CNN) that classifies land cover labels from high-resolution remote sensing (HRRS) images and to increase the classification accuracy in agricultural areas using the parcel boundary extraction algorithm. The developed model comprises three modules, namely the pre-processing, land cover classification, and post-processing modules. The pre-processing module diversifies the perspective of the HRRS images by separating images with 75% overlaps to reduce the misclassification that can occur in a single image. The land cover classification module was designed based on the FusionNet model structure, and the optimal land cover type was assigned for each pixel of the separated HRRS images. The post-processing module determines the ultimate land cover types for each pixel unit by summing up the several-perspective classification results and aggregating the pixel-classification result for the parcel-boundary unit in agricultural areas. The developed model was trained with land cover maps and orthographic images (area: 547 km2) from the Jeonnam province in Korea. Model validation was conducted with two spatially and temporally different sites including Subuk-myeon of Jeonnam province in 2018 and Daseo-myeon of Chungbuk province in 2016. In the respective validation sites, the models overall accuracies were 0.81 and 0.71, and kappa coefficients were 0.75 and 0.64, implying substantial model performance. The model performance was particularly better when considering parcel boundaries in agricultural areas, exhibiting an overall accuracy of 0.89 and kappa coefficient 0.81 (almost perfect). It was concluded that the developed model may help perform rapid and accurate land cover updates especially for agricultural areas.Chapter 1. Introduction 1 1.1. Study background 1 1.2. Objective of thesis 4 Chapter 2. Literature review 6 2.1. Development of remote sensing technique 6 2.2. Land cover segmentation 9 2.3. Land boundary extraction 13 Chapter 3. Development of the land cover classification model 15 3.1. Conceptual structure of the land cover classification model 15 3.2. Pre-processing module 16 3.3. CNN based land cover classification module 17 3.4. Post processing module 22 3.4.1 Determination of land cover in a pixel unit 22 3.4.2 Aggregation of land cover to parcel boundary 24 Chapter 4. Verification of the land cover classification model 30 4.1. Study area and data acquisition 31 4.1.1. Training area 31 4.1.2. Verification area 32 4.1.3. Data acquisition 33 4.2. Training the land cover classification model 36 4.3. Verification method 37 4.3.1. The performance measurement methods of land cover classification model 37 4.3.2. Accuracy estimation methods of agricultural parcel boundary 39 4.3.3. Comparison of boundary based classification result with ERDAS Imagine 41 4.4. Verification of land cover classification model 42 4.4.1. Performance of land cover classification at the child subcategory 42 4.4.2. Classification accuracy of the aggregated land cover to main category 46 4.4.3. Classification accuracy of boundary based aggregation in agricultural area 57 Chapter 5. Conclusions 71 Reference 73 ๊ตญ ๋ฌธ ์ดˆ ๋ก 83Maste

    Tiny Machine Learning Environment: Enabling Intelligence on Constrained Devices

    Get PDF
    Running machine learning algorithms (ML) on constrained devices at the extreme edge of the network is problematic due to the computational overhead of ML algorithms, available resources on the embedded platform, and application budget (i.e., real-time requirements, power constraints, etc.). This required the development of specific solutions and development tools for what is now referred to as TinyML. In this dissertation, we focus on improving the deployment and performance of TinyML applications, taking into consideration the aforementioned challenges, especially memory requirements. This dissertation contributed to the construction of the Edge Learning Machine environment (ELM), a platform-independent open-source framework that provides three main TinyML services, namely shallow ML, self-supervised ML, and binary deep learning on constrained devices. In this context, this work includes the following steps, which are reflected in the thesis structure. First, we present the performance analysis of state-of-the-art shallow ML algorithms including dense neural networks, implemented on mainstream microcontrollers. The comprehensive analysis in terms of algorithms, hardware platforms, datasets, preprocessing techniques, and configurations shows similar performance results compared to a desktop machine and highlights the impact of these factors on overall performance. Second, despite the assumption that TinyML only permits models inference provided by the scarcity of resources, we have gone a step further and enabled self-supervised on-device training on microcontrollers and tiny IoT devices by developing the Autonomous Edge Pipeline (AEP) system. AEP achieves comparable accuracy compared to the typical TinyML paradigm, i.e., models trained on resource-abundant devices and then deployed on microcontrollers. Next, we present the development of a memory allocation strategy for convolutional neural networks (CNNs) layers, that optimizes memory requirements. This approach reduces the memory footprint without affecting accuracy nor latency. Moreover, e-skin systems share the main requirements of the TinyML fields: enabling intelligence with low memory, low power consumption, and low latency. Therefore, we designed an efficient Tiny CNN architecture for e-skin applications. The architecture leverages the memory allocation strategy presented earlier and provides better performance than existing solutions. A major contribution of the thesis is given by CBin-NN, a library of functions for implementing extremely efficient binary neural networks on constrained devices. The library outperforms state of the art NN deployment solutions by drastically reducing memory footprint and inference latency. All the solutions proposed in this thesis have been implemented on representative devices and tested in relevant applications, of which results are reported and discussed. The ELM framework is open source, and this work is clearly becoming a useful, versatile toolkit for the IoT and TinyML research and development community

    Machine Vision Identification of Plants

    Get PDF
    • โ€ฆ
    corecore