2,016 research outputs found

    κ°•μΈν•œ λŒ€ν™”ν˜• μ˜μƒ λΆ„ν•  μ•Œκ³ λ¦¬μ¦˜μ„ μœ„ν•œ μ‹œλ“œ 정보 ν™•μž₯ 기법에 λŒ€ν•œ 연ꡬ

    Get PDF
    ν•™μœ„λ…Όλ¬Έ (박사) -- μ„œμšΈλŒ€ν•™κ΅ λŒ€ν•™μ› : κ³΅κ³ΌλŒ€ν•™ 전기·컴퓨터곡학뢀, 2021. 2. 이경무.Segmentation of an area corresponding to a desired object in an image is essential to computer vision problems. This is because most algorithms are performed in semantic units when interpreting or analyzing images. However, segmenting the desired object from a given image is an ambiguous issue. The target object varies depending on user and purpose. To solve this problem, an interactive segmentation technique has been proposed. In this approach, segmentation was performed in the desired direction according to interaction with the user. In this case, seed information provided by the user plays an important role. If the seed provided by a user contain abundant information, the accuracy of segmentation increases. However, providing rich seed information places much burden on the users. Therefore, the main goal of the present study was to obtain satisfactory segmentation results using simple seed information. We primarily focused on converting the provided sparse seed information to a rich state so that accurate segmentation results can be derived. To this end, a minimum user input was taken and enriched it through various seed enrichment techniques. A total of three interactive segmentation techniques was proposed based on: (1) Seed Expansion, (2) Seed Generation, (3) Seed Attention. Our seed enriching type comprised expansion of area around a seed, generation of new seed in a new position, and attention to semantic information. First, in seed expansion, we expanded the scope of the seed. We integrated reliable pixels around the initial seed into the seed set through an expansion step composed of two stages. Through the extended seed covering a wider area than the initial seed, the seed's scarcity and imbalance problems was resolved. Next, in seed generation, we created a seed at a new point, but not around the seed. We trained the system by imitating the user behavior through providing a new seed point in the erroneous region. By learning the user's intention, our model could e ciently create a new seed point. The generated seed helped segmentation and could be used as additional information for weakly supervised learning. Finally, through seed attention, we put semantic information in the seed. Unlike the previous models, we integrated both the segmentation process and seed enrichment process. We reinforced the seed information by adding semantic information to the seed instead of spatial expansion. The seed information was enriched through mutual attention with feature maps generated during the segmentation process. The proposed models show superiority compared to the existing techniques through various experiments. To note, even with sparse seed information, our proposed seed enrichment technique gave by far more accurate segmentation results than the other existing methods.μ˜μƒμ—μ„œ μ›ν•˜λŠ” 물체 μ˜μ—­μ„ μž˜λΌλ‚΄λŠ” 것은 컴퓨터 λΉ„μ „ λ¬Έμ œμ—μ„œ ν•„μˆ˜μ μΈ μš”μ†Œμ΄λ‹€. μ˜μƒμ„ ν•΄μ„ν•˜κ±°λ‚˜ 뢄석할 λ•Œ, λŒ€λΆ€λΆ„μ˜ μ•Œκ³ λ¦¬μ¦˜λ“€μ΄ 의미둠적인 λ‹¨μœ„ 기반으둜 λ™μž‘ν•˜κΈ° λ•Œλ¬Έμ΄λ‹€. κ·ΈλŸ¬λ‚˜ μ˜μƒμ—μ„œ 물체 μ˜μ—­μ„ λΆ„ν• ν•˜λŠ” 것은 λͺ¨ν˜Έν•œ λ¬Έμ œμ΄λ‹€. μ‚¬μš©μžμ™€ λͺ©μ μ— 따라 μ›ν•˜λŠ” 물체 μ˜μ—­μ΄ 달라지기 λ•Œλ¬Έμ΄λ‹€. 이λ₯Ό ν•΄κ²°ν•˜κΈ° μœ„ν•΄ μ‚¬μš©μžμ™€μ˜ ꡐλ₯˜λ₯Ό 톡해 μ›ν•˜λŠ” λ°©ν–₯으둜 μ˜μƒ 뢄할을 μ§„ν–‰ν•˜λŠ” λŒ€ν™”ν˜• μ˜μƒ λΆ„ν•  기법이 μ‚¬μš©λœλ‹€. μ—¬κΈ°μ„œ μ‚¬μš©μžκ°€ μ œκ³΅ν•˜λŠ” μ‹œλ“œ 정보가 μ€‘μš”ν•œ 역할을 ν•œλ‹€. μ‚¬μš©μžμ˜ μ˜λ„λ₯Ό λ‹΄κ³  μžˆλŠ” μ‹œλ“œ 정보가 μ •ν™•ν• μˆ˜λ‘ μ˜μƒ λΆ„ν• μ˜ 정확도도 μ¦κ°€ν•˜κ²Œ λœλ‹€. κ·ΈλŸ¬λ‚˜ ν’λΆ€ν•œ μ‹œλ“œ 정보λ₯Ό μ œκ³΅ν•˜λŠ” 것은 μ‚¬μš©μžμ—κ²Œ λ§Žμ€ 뢀담을 주게 λœλ‹€. κ·ΈλŸ¬λ―€λ‘œ κ°„λ‹¨ν•œ μ‹œλ“œ 정보λ₯Ό μ‚¬μš©ν•˜μ—¬ λ§Œμ‘±ν• λ§Œν•œ λΆ„ν•  κ²°κ³Όλ₯Ό μ–»λŠ” 것이 μ£Όμš” λͺ©μ μ΄ λœλ‹€. μš°λ¦¬λŠ” 제곡된 ν¬μ†Œν•œ μ‹œλ“œ 정보λ₯Ό λ³€ν™˜ν•˜λŠ” μž‘μ—…μ— μ΄ˆμ μ„ λ‘μ—ˆλ‹€. λ§Œμ•½ μ‹œλ“œ 정보가 ν’λΆ€ν•˜κ²Œ λ³€ν™˜λœλ‹€λ©΄ μ •ν™•ν•œ μ˜μƒ λΆ„ν•  κ²°κ³Όλ₯Ό 얻을 수 있기 λ•Œλ¬Έμ΄λ‹€. κ·ΈλŸ¬λ―€λ‘œ λ³Έ ν•™μœ„ λ…Όλ¬Έμ—μ„œλŠ” μ‹œλ“œ 정보λ₯Ό ν’λΆ€ν•˜κ²Œ ν•˜λŠ” 기법듀을 μ œμ•ˆν•œλ‹€. μ΅œμ†Œν•œμ˜ μ‚¬μš©μž μž…λ ₯을 κ°€μ •ν•˜κ³  이λ₯Ό λ‹€μ–‘ν•œ μ‹œλ“œ ν™•μž₯ 기법을 톡해 λ³€ν™˜ν•œλ‹€. μš°λ¦¬λŠ” μ‹œλ“œ ν™•λŒ€, μ‹œλ“œ 생성, μ‹œλ“œ 주의 집쀑에 κΈ°λ°˜ν•œ 총 μ„Έ κ°€μ§€μ˜ λŒ€ν™”ν˜• μ˜μƒ λΆ„ν•  기법을 μ œμ•ˆν•œλ‹€. 각각 μ‹œλ“œ μ£Όλ³€μœΌλ‘œμ˜ μ˜μ—­ ν™•λŒ€, μƒˆλ‘œμš΄ 지점에 μ‹œλ“œ 생성, 의미둠적 정보에 μ£Όλͺ©ν•˜λŠ” ν˜•νƒœμ˜ μ‹œλ“œ ν™•μž₯ 기법을 μ‚¬μš©ν•œλ‹€. λ¨Όμ € μ‹œλ“œ ν™•λŒ€μ— κΈ°λ°˜ν•œ κΈ°λ²•μ—μ„œ μš°λ¦¬λŠ” μ‹œλ“œμ˜ μ˜μ—­ ν™•μž₯을 λͺ©ν‘œλ‘œ ν•œλ‹€. 두 λ‹¨κ³„λ‘œ κ΅¬μ„±λœ ν™•λŒ€ 과정을 톡해 처음 μ‹œλ“œ μ£Όλ³€μ˜ λΉ„μŠ·ν•œ 픽셀듀을 μ‹œλ“œ μ˜μ—­μœΌλ‘œ νŽΈμž…ν•œλ‹€. μ΄λ ‡κ²Œ ν™•μž₯된 μ‹œλ“œλ₯Ό μ‚¬μš©ν•¨μœΌλ‘œμ¨ μ‹œλ“œμ˜ ν¬μ†Œν•¨κ³Ό λΆˆκ· ν˜•μœΌλ‘œ μΈν•œ 문제λ₯Ό ν•΄κ²°ν•  수 μžˆλ‹€. λ‹€μŒμœΌλ‘œ μ‹œλ“œ 생성에 κΈ°λ°˜ν•œ κΈ°λ²•μ—μ„œ μš°λ¦¬λŠ” μ‹œλ“œ 주변이 μ•„λ‹Œ μƒˆλ‘œμš΄ 지점에 μ‹œλ“œλ₯Ό μƒμ„±ν•œλ‹€. μš°λ¦¬λŠ” μ˜€μ°¨κ°€ λ°œμƒν•œ μ˜μ—­μ— μ‚¬μš©μžκ°€ μƒˆλ‘œμš΄ μ‹œλ“œλ₯Ό μ œκ³΅ν•˜λŠ” λ™μž‘μ„ λͺ¨λ°©ν•˜μ—¬ μ‹œμŠ€ν…œμ„ ν•™μŠ΅ν•˜μ˜€λ‹€. μ‚¬μš©μžμ˜ μ˜λ„λ₯Ό ν•™μŠ΅ν•¨μœΌλ‘œμ¨ 효과적으둜 μ‹œλ“œλ₯Ό 생성할 수 μžˆλ‹€. μƒμ„±λœ μ‹œλ“œλŠ” μ˜μƒ λΆ„ν• μ˜ 정확도λ₯Ό 높일 뿐만 μ•„λ‹ˆλΌ μ•½μ§€λ„ν•™μŠ΅μ„ μœ„ν•œ λ°μ΄ν„°λ‘œμ¨ ν™œμš©λ  수 μžˆλ‹€. λ§ˆμ§€λ§‰μœΌλ‘œ μ‹œλ“œ 주의 집쀑을 ν™œμš©ν•œ κΈ°λ²•μ—μ„œ μš°λ¦¬λŠ” 의미둠적 정보λ₯Ό μ‹œλ“œμ— λ‹΄λŠ”λ‹€. 기쑴에 μ œμ•ˆν•œ 기법듀과 달리 μ˜μƒ λΆ„ν•  λ™μž‘κ³Ό μ‹œλ“œ ν™•μž₯ λ™μž‘μ΄ ν†΅ν•©λœ λͺ¨λΈμ„ μ œμ•ˆν•œλ‹€. μ‹œλ“œ μ •λ³΄λŠ” μ˜μƒ λΆ„ν•  λ„€νŠΈμ›Œν¬μ˜ νŠΉμ§•λ§΅κ³Ό μƒν˜Έ ꡐλ₯˜ν•˜λ©° κ·Έ 정보가 풍뢀해진닀. μ œμ•ˆν•œ λͺ¨λΈλ“€μ€ λ‹€μ–‘ν•œ μ‹€ν—˜μ„ 톡해 κΈ°μ‘΄ 기법 λŒ€λΉ„ μš°μˆ˜ν•œ μ„±λŠ₯을 κΈ°λ‘ν•˜μ˜€λ‹€. 특히 μ‹œλ“œκ°€ λΆ€μ‘±ν•œ μƒν™©μ—μ„œ μ‹œλ“œ ν™•μž₯ 기법듀은 ν›Œλ₯­ν•œ λŒ€ν™”ν˜• μ˜μƒ λΆ„ν•  μ„±λŠ₯을 λ³΄μ˜€λ‹€.1 Introduction 1 1.1 Previous Works 2 1.2 Proposed Methods 4 2 Interactive Segmentation with Seed Expansion 9 2.1 Introduction 9 2.2 Proposed Method 12 2.2.1 Background 13 2.2.2 Pyramidal RWR 16 2.2.3 Seed Expansion 19 2.2.4 Re nement with Global Information 24 2.3 Experiments 27 2.3.1 Dataset 27 2.3.2 Implement Details 28 2.3.3 Performance 29 2.3.4 Contribution of Each Part 30 2.3.5 Seed Consistency 31 2.3.6 Running Time 33 2.4 Summary 34 3 Interactive Segmentation with Seed Generation 37 3.1 Introduction 37 3.2 Related Works 40 3.3 Proposed Method 41 3.3.1 System Overview 41 3.3.2 Markov Decision Process 42 3.3.3 Deep Q-Network 46 3.3.4 Model Architecture 47 3.4 Experiments 48 3.4.1 Implement Details 48 3.4.2 Performance 49 3.4.3 Ablation Study 53 3.4.4 Other Datasets 55 3.5 Summary 58 4 Interactive Segmentation with Seed Attention 61 4.1 Introduction 61 4.2 Related Works 64 4.3 Proposed Method 65 4.3.1 Interactive Segmentation Network 65 4.3.2 Bi-directional Seed Attention Module 67 4.4 Experiments 70 4.4.1 Datasets 70 4.4.2 Metrics 70 4.4.3 Implement Details 71 4.4.4 Performance 71 4.4.5 Ablation Study 76 4.4.6 Seed enrichment methods 79 4.5 Summary 82 5 Conclusions 87 5.1 Summary 89 Bibliography 90 ꡭ문초둝 103Docto

    Application of Fast Deviation Correction Algorithm Based on Shape Matching Algorithm in Component Placement

    Get PDF
    For contradiction PC template matching between accuracy and speed, combined with the advantages of FPGA high speed parallel computing. This paper presents a FPGA-based rapid correction shape matching algorithm. Mainly in the FPGA, using shape matching and least squares method to calculate the angular deviation chip components. Use single instruction stream algorithm acceleration. Experimental results show that compared with traditional PC template matching algorithms, this algorithm to further improve the correction accuracy and greatly reducing correction time. And SMT machine vision correction can be obtained in a stable and efficient use

    Robust surface modelling of visual hull from multiple silhouettes

    Get PDF
    Reconstructing depth information from images is one of the actively researched themes in computer vision and its application involves most vision research areas from object recognition to realistic visualisation. Amongst other useful vision-based reconstruction techniques, this thesis extensively investigates the visual hull (VH) concept for volume approximation and its robust surface modelling when various views of an object are available. Assuming that multiple images are captured from a circular motion, projection matrices are generally parameterised in terms of a rotation angle from a reference position in order to facilitate the multi-camera calibration. However, this assumption is often violated in practice, i.e., a pure rotation in a planar motion with accurate rotation angle is hardly realisable. To address this problem, at first, this thesis proposes a calibration method associated with the approximate circular motion. With these modified projection matrices, a resulting VH is represented by a hierarchical tree structure of voxels from which surfaces are extracted by the Marching cubes (MC) algorithm. However, the surfaces may have unexpected artefacts caused by a coarser volume reconstruction, the topological ambiguity of the MC algorithm, and imperfect image processing or calibration result. To avoid this sensitivity, this thesis proposes a robust surface construction algorithm which initially classifies local convex regions from imperfect MC vertices and then aggregates local surfaces constructed by the 3D convex hull algorithm. Furthermore, this thesis also explores the use of wide baseline images to refine a coarse VH using an affine invariant region descriptor. This improves the quality of VH when a small number of initial views is given. In conclusion, the proposed methods achieve a 3D model with enhanced accuracy. Also, robust surface modelling is retained when silhouette images are degraded by practical noise

    A novel approach to rainfall measuring: methodology, field test and business opportunity

    Get PDF
    Being able to measure rainfall is crucial in everyday life. The more rainfall measures are accurate, spatially distributed and detailed in time, the more forecast models - be they meteorological or hydrological - can be accurate. Safety on travel networks could be increased by informing users about the nearby roads’ conditions in real time. In the agricultural sector, being able to gain a detailed knowledge of rainfalls would allow for an optimal management of irrigation, nutrients and phytosanitary treatments. In the sport sector, a better measurement of rainfalls for outdoor events (e.g., motor, motorcycle or bike races) would increase athletes’ safety. Rain gauges are the most common and widely used tools for rainfall measurement. However, the existent monitoring networks still fail in providing accurate spatial representations of localized precipitation events due to the sparseness. This effect is magnified by the intrinsic nature of intense precipitation events, as they are naturally characterized by a great spatial and temporal variability. Potentially, coupling at-ground measures (i.e., coming from pluviometric and disdrometric networks) with remote measurement (e.g., radars or meteorological satellites) could allow to describe the rainfall phenomena in a more continuous and spatially detailed way. However, this kind of approach requires that at-ground measurements are used to calibrate the remote sensors relationships, which leads us back to the dearth of ground networks diffusion. Hence the need to increase the presence of ground measures, in order to gain a better description of the events, and to make a more productive use of the remote sensing technologies. The ambitious aim of the methodology developed in this thesis is to repurpose other sensors already available at ground (e.g., surveillance cameras, webcams, smartphones, cars, etc.) into new source of rain rate measures widely distributed over space and time. The technology, developed to function in daylight conditions, requires that the pictures collected during rainfall events are analyzed to identify and characterize each raindrop. The process leads to an instant measurement of the rain rate associated with the captured image. To improve the robustness of the measurement, we propose to elaborate a higher number of images within a predefined time span (i.e., 12 or more pictures per minute) and to provide an averaged measure over the observed time interval. A schematic summary of how the method works for each acquired image is represented hereinafter : 1. background removal; 2. identification of the rain drops; 3. positioning of each drop in the control volume, by using the blur effect; 4. estimation of drops’ diameters, under the hypothesis that each drop falls at its terminal velocity; 5. rain rate estimation, as the sum of the contributions of each drop. Different techniques for background recognition, drops detection and selection and noise reduction were investigated. Each solution has been applied to the same images sample, in order to identify the combination producing accuracy in the rainfall estimate. The best performing procedure was then validated, by applying it to a wider sample of images. Such a sample was acquired by an experimental station installed on the roof of the Laboratory of Hydraulics of the Politecnico di Torino. The sample includes rainfall events which took place between May 15th, 2016 and February 15th, 2017. Seasonal variability allowed to record events characterized by different intensity in varied light conditions. Moreover, the technology developed during this program of research was patented (2015) and represents the heart of WaterView, spinoff of the Politecnico di Torino founded in February 2015, which is currently in charge of the further development of this technology, its dissemination, and its commercial exploitation
    • …
    corecore