5,427 research outputs found

    Automatic detection of welding defects using the convolutional neural network

    Get PDF
    Quality control of welded joints is an important step before commissioning of various types of metal structures. The main obstacles to the commissioning of such facilities are the areas where the welded joint deviates from acceptable defective standards. The defects of welded joints include non-welded, foreign inclusions, cracks, pores, etc. The article describes an approach to the detection of the main types of defects of welded joints using a combination of convolutional neural networks and support vector machine methods. Convolutional neural networks are used for primary classification. The support vector machine is used to accurately define defect boundaries. As a preprocessing in our work, we use the methods of morphological filtration. A series of experiments confirms the high efficiency of the proposed method in comparison with pure CNN method for detecting defects

    Automatic Lumbar Vertebrae Segmentation in Fluoroscopic Images via Optimised Concurrent Hough Transform

    No full text
    Low back pain is a very common problem in the industrialised countries and its associated cost is enormous. Diagnosis of the underlying causes can be extremely difficult. Many studies have focused on mechanical disorders of the spine. Digital videofluoroscopy (DVF) was widely used to obtain images for motion studies. This can provide motion sequences of the lumbar spine, but the images obtained often suffer due to noise, exacerbated by the very low radiation dosage. Thus determining vertebrae position within the image sequence presents a considerable challenge. In this paper, we show how our new approach can automatically detect the positions and borders of vertebrae concurrently, relieving many of the problems experienced in other approaches. First, we use phase congruency to relieve difficulty associated with threshold selection in edge detection of the illumination variant DVF images. Then, our new Hough transform approach is applied to determine the moving vertebrae, concurrently. We include optimisation via a genetic algorithm as without it the extraction of moving multiple vertebrae is computationally daunting. Our results show that this new approach can indeed provide extractions of position and rotation which appear to be of sufficient quality to aid therapy and diagnosis of spinal disorders

    Machine Vision Application For Automatic Defect Segmentation In Weld Radiographs

    Get PDF
    Objektif penyelidikan ini adalah untuk membangunkan satu kaedah peruasan kecacatan kimpalan automatik yang boleh meruas pelbagai jenis kecacatan kimpalan yang wujud dalam imej radiografi kimpalan. Kaedah segmentasi kecacatan automatik yang dibangunkan terdir:i daripada tiga algoritma utama, iaitu algoritma penyingkiran label, algoritma pengenalpastian bahagian kimpalan dan algoritma segmentasi kecacatan kimpalan. Algoritma penyingkiran label dibangunkan untuk mengenalpasti dan menyingkirkan label yang terdapat pada imej radiograf kimpalan secara automatik, sebelum algoritma pengenalpastian bahagian kimpalan dan algortima segmentasi kecacatan diaplikasikan ke atas imej radiografi. Satu algoritma pengenalpastian bahagian kimpalan juga dibangunkan dengan tujuan mengenalpasti bahagian kimpalan dalam imej radiogaf secara automatik dengan menggunakan profil keamatan yang diperoleh daripada imej radiografi. The objective of the research is to develop an automatic weld defect segmentation methodology to segment different types of defects in radiographic images of welds. The segmentation methodology consists of three main algorithms. namely label removal algorithm. weld extraction algorithm and defect segmentation algorithm. The label removal algorithm was developed to detect and remove labels that are printed on weld radiographs automatically before weld extraction algorithm and defect detection algorithm are applied. The weld extraction algorithm was developed to locate and extract welds automatically from the intensity profiles taken across the image by using graphical analysis. This algorithm was able to extract weld from a radiograph regardless of whether the intensity profile is Gaussian or otherwise. This method is an improvement compared to the previous weld extraction methods which are limited to weld image with Gaussian intensity profiles. Finally. a defect segmentation algorithm was developed to segment the defects automatically from the image using background subtraction and rank leveling method

    A study of hough transform for weld extraction

    Get PDF
    The process of joining metals is called welding. At times, selecting a poor quality material or improper usage of welding technologies may cause defects in welded joints. Some of these welded joints have to be tested nondestructively, because their failure can cause lot of damage, for instance in power plants. Radiography is a very common method for non-destructive testing of welds. It is done by certified weld inspectors who have knowledge about weld flaws, looking at the radiograph of the welded joint with naked eye. The judgment of the weld inspector can be biased; subjective, because it is dependent on his/her experience. This manual method can also become very time consuming. Many researches were exploring computer aided examination of radiographic images in early 1990โ€™s. With much advancement in computer vision and image processing technologies, they are being used to find more effective ways of automatic weld inspection. These days, fuzzy based methods are being widely used in this area too. The first step in automatic weld inspection is to locate the welds or find a Region of Interest (ROI) in the radiographic image [7]. In this thesis, a Standard Hough Transform (SHT) based methodology is developed for weld extraction. Firstly, we have done binarization of image to remove the background and non-welds. For binarization, optimal binary threshold is found by a metaheuristic โ€“Simulated annealing. Secondly, we use SHT to generate the Hough Transform matrix of all non-zero points in the binary image. Thirdly, we have explored two different paths to find a meaningful set of lines in the binarized image that are welds. Finally, these lines are verified as weld using a weld-peak detection procedure. Weld-peak detection is also helpful to remove any non-welds that were remaining. We have used 25 digitized radiographic images containing 100 welds to test the method in terms of true detection and false alarm rate

    ์‹ฌ์ธต์‹ ๊ฒฝ๋ง์„ ์ด์šฉํ•œ ์ž๋™ํ™”๋œ ์น˜๊ณผ ์˜๋ฃŒ์˜์ƒ ๋ถ„์„

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ์น˜๊ณผ๋Œ€ํ•™ ์น˜์˜๊ณผํ•™๊ณผ, 2021.8. ํ•œ์ค‘์„.๋ชฉ ์ : ์น˜๊ณผ ์˜์—ญ์—์„œ๋„ ์‹ฌ์ธต์‹ ๊ฒฝ๋ง(Deep Neural Network) ๋ชจ๋ธ์„ ์ด์šฉํ•œ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์—์„œ์˜ ์ž„ํ”Œ๋ž€ํŠธ ๋ถ„๋ฅ˜, ๋ณ‘์†Œ ์œ„์น˜ ํƒ์ง€ ๋“ฑ์˜ ์—ฐ๊ตฌ๋“ค์ด ์ง„ํ–‰๋˜์—ˆ์œผ๋‚˜, ์ตœ๊ทผ ๊ฐœ๋ฐœ๋œ ํ‚คํฌ์ธํŠธ ํƒ์ง€(keypoint detection) ๋ชจ๋ธ ๋˜๋Š” ์ „์ฒด์  ๊ตฌํšํ™”(panoptic segmentation) ๋ชจ๋ธ์„ ์˜๋ฃŒ๋ถ„์•ผ์— ์ ์šฉํ•œ ์—ฐ๊ตฌ๋Š” ์•„์ง ๋ฏธ๋น„ํ•˜๋‹ค. ๋ณธ ์—ฐ๊ตฌ์˜ ๋ชฉ์ ์€ ์น˜๊ทผ๋‹จ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์—์„œ ํ‚คํฌ์ธํŠธ ํƒ์ง€๋ฅผ ์ด์šฉํ•ด ์ž„ํ”Œ๋ž€ํŠธ ๊ณจ ์†Œ์‹ค ์ •๋„๋ฅผ ํŒŒ์•…ํ•˜๋Š” ๋ชจ๋ธ๊ณผ panoptic segmentation์„ ํŒŒ๋…ธ๋ผ๋งˆ์˜์ƒ์— ์ ์šฉํ•˜์—ฌ ๋‹ค์–‘ํ•œ ๊ตฌ์กฐ๋ฌผ๋“ค์„ ๊ตฌํšํ™”ํ•˜๋Š” ๋ชจ๋ธ์„ ํ•™์Šต์‹œ์ผœ ์ง„๋ฃŒ์— ๋ณด์กฐ์ ์œผ๋กœ ํ™œ์šฉ๋˜๋„๋ก ๋งŒ๋“ค์–ด๋ณด๊ณ , ์ด ๋ชจ๋ธ๋“ค์˜ ์ถ”๋ก ๊ฒฐ๊ณผ๋ฅผ ํ‰๊ฐ€ํ•ด๋ณด๋Š” ๊ฒƒ์ด๋‹ค. ๋ฐฉ ๋ฒ•: ๊ฐ์ฒด ํƒ์ง€ ๋ฐ ๊ตฌํšํ™”์— ์žˆ์–ด ๋„๋ฆฌ ์—ฐ๊ตฌ๋œ ํ•ฉ์„ฑ๊ณฑ ์‹ ๊ฒฝ๋ง ๋ชจ๋ธ์ธ Mask-RCNN์„ ํ‚คํฌ์ธํŠธ ํƒ์ง€๊ฐ€ ๊ฐ€๋Šฅํ•œ ํ˜•ํƒœ๋กœ ์ค€๋น„ํ•˜์—ฌ ์น˜๊ทผ๋‹จ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์—์„œ ์ž„ํ”Œ๋ž€ํŠธ์˜ top, apex, ๊ทธ๋ฆฌ๊ณ  bone level ์ง€์ ์„ ์ขŒ์šฐ๋กœ ์ด 6์ง€์  ํƒ์ง€ํ•˜๊ฒŒ๋” ํ•™์Šต์‹œํ‚จ ๋’ค, ํ•™์Šต์— ์‚ฌ์šฉ๋˜์ง€ ์•Š์€ ์‹œํ—˜ ๋ฐ์ดํ„ฐ์…‹์„ ๋Œ€์ƒ์œผ๋กœ ํƒ์ง€์‹œํ‚จ๋‹ค. ํ‚คํฌ์ธํŠธ ํƒ์ง€ ํ‰๊ฐ€์šฉ ์ง€ํ‘œ์ธ object keypoint similarity (OKS) ๋ฐ ์ด๋ฅผ ์ด์šฉํ•œ average precision (AP) ๊ฐ’์„ ๊ณ„์‚ฐํ•˜๊ณ , ํ‰๊ท  OKS๊ฐ’์„ ํ†ตํ•ด ๋ชจ๋ธ ๋ฐ ์น˜๊ณผ์˜์‚ฌ์˜ ๊ฒฐ๊ณผ๋ฅผ ๋น„๊ตํ•œ๋‹ค. ๋˜ํ•œ, ํƒ์ง€๋œ ํ‚คํฌ์ธํŠธ๋ฅผ ๋ฐ”ํƒ•์œผ๋กœ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์ƒ์—์„œ์˜ ๊ณจ ์†Œ์‹ค ์ •๋„๋ฅผ ์ˆ˜์น˜ํ™”ํ•œ๋‹ค. Panoptic segmentation์„ ์œ„ํ•ด์„œ๋Š” ๊ธฐ์กด์˜ ๋ฒค์น˜๋งˆํฌ์—์„œ ์šฐ์ˆ˜ํ•œ ์„ฑ์ ์„ ๊ฑฐ๋‘” ์‹ ๊ฒฝ๋ง ๋ชจ๋ธ์ธ Panoptic DeepLab์„ ํŒŒ๋…ธ๋ผ๋งˆ์˜์ƒ์—์„œ ์ฃผ์š” ๊ตฌ์กฐ๋ฌผ(์ƒ์•…๋™, ์ƒ์•…๊ณจ, ํ•˜์•…๊ด€, ํ•˜์•…๊ณจ, ์ž์—ฐ์น˜, ์น˜๋ฃŒ๋œ ์น˜์•„, ์ž„ํ”Œ๋ž€ํŠธ)์„ ๊ตฌํšํ™”ํ•˜๋„๋ก ํ•™์Šต์‹œํ‚จ ๋’ค, ์‹œํ—˜ ๋ฐ์ดํ„ฐ์…‹์—์„œ์˜ ๊ตฌํšํ™” ๊ฒฐ๊ณผ์— panoptic / semantic / instance segmentation ๊ฐ๊ฐ์˜ ํ‰๊ฐ€์ง€ํ‘œ๋“ค์„ ์ ์šฉํ•˜๊ณ , ํ”ฝ์…€๋“ค์˜ ์ •๋‹ต(ground truth) ํด๋ž˜์Šค์™€ ๋ชจ๋ธ์ด ์ถ”๋ก ํ•œ ํด๋ž˜์Šค์— ๋Œ€ํ•œ confusion matrix๋ฅผ ๊ณ„์‚ฐํ•œ๋‹ค. ๊ฒฐ ๊ณผ: OKS๊ฐ’์„ ๊ธฐ๋ฐ˜์œผ๋กœ ๊ณ„์‚ฐํ•œ ํ‚คํฌ์ธํŠธ ํƒ์ง€ AP๋Š”, ๋ชจ๋“  OKS threshold์— ๋Œ€ํ•œ ํ‰๊ท ์˜ ๊ฒฝ์šฐ, ์ƒ์•… ์ž„ํ”Œ๋ž€ํŠธ์—์„œ๋Š” 0.761, ํ•˜์•… ์ž„ํ”Œ๋ž€ํŠธ์—์„œ๋Š” 0.786์ด์—ˆ๋‹ค. ํ‰๊ท  OKS๋Š” ๋ชจ๋ธ์ด 0.8885, ์น˜๊ณผ์˜์‚ฌ๊ฐ€ 0.9012๋กœ, ํ†ต๊ณ„์ ์œผ๋กœ ์œ ์˜๋ฏธํ•œ ์ฐจ์ด๊ฐ€ ์—†์—ˆ๋‹ค (p = 0.41). ๋ชจ๋ธ์˜ ํ‰๊ท  OKS ๊ฐ’์€ ์‚ฌ๋žŒ์˜ ํ‚คํฌ์ธํŠธ ์–ด๋…ธํ…Œ์ด์…˜ ์ •๊ทœ๋ถ„ํฌ์ƒ์—์„œ ์ƒ์œ„ 66.92% ์ˆ˜์ค€์ด์—ˆ๋‹ค. ํŒŒ๋…ธ๋ผ๋งˆ์˜์ƒ ๊ตฌ์กฐ๋ฌผ ๊ตฌํšํ™”์—์„œ๋Š”, panoptic segmentation ํ‰๊ฐ€์ง€ํ‘œ์ธ panoptic quality ๊ฐ’์˜ ๊ฒฝ์šฐ ๋ชจ๋“  ํด๋ž˜์Šค์˜ ํ‰๊ท ์€ 80.47์ด์—ˆ์œผ๋ฉฐ, ์น˜๋ฃŒ๋œ ์น˜์•„๊ฐ€ 57.13์œผ๋กœ ๊ฐ€์žฅ ๋‚ฎ์•˜๊ณ  ํ•˜์•…๊ด€์ด 65.97๋กœ ๋‘๋ฒˆ์งธ๋กœ ๋‚ฎ์€ ๊ฐ’์„ ๋ณด์˜€๋‹ค. Semantic segmentation ํ‰๊ฐ€์ง€ํ‘œ์ธ globalํ•œ Intersection over Union (IoU) ๊ฐ’์€ ๋ชจ๋“  ํด๋ž˜์Šค ํ‰๊ท  0.795์˜€์œผ๋ฉฐ, ํ•˜์•…๊ด€์ด 0.639๋กœ ๊ฐ€์žฅ ๋‚ฎ์•˜๊ณ  ์น˜๋ฃŒ๋œ ์น˜์•„๊ฐ€ 0.656์œผ๋กœ ๋‘๋ฒˆ์งธ๋กœ ๋‚ฎ์€ ๊ฐ’์„ ๋ณด์˜€๋‹ค. Confusion matrix ๊ณ„์‚ฐ ๊ฒฐ๊ณผ, ground truth ํ”ฝ์…€๋“ค ์ค‘ ์˜ฌ๋ฐ”๋ฅด๊ฒŒ ์ถ”๋ก ๋œ ํ”ฝ์…€๋“ค์˜ ๋น„์œจ์€ ํ•˜์•…๊ด€์ด 0.802๋กœ ๊ฐ€์žฅ ๋‚ฎ์•˜๋‹ค. ๊ฐœ๋ณ„ ๊ฐ์ฒด์— ๋Œ€ํ•œ IoU๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ๊ณ„์‚ฐํ•œ Instance segmentation ํ‰๊ฐ€์ง€ํ‘œ์ธ AP๊ฐ’์€, ๋ชจ๋“  IoU threshold์— ๋Œ€ํ•œ ํ‰๊ท ์˜ ๊ฒฝ์šฐ, ์น˜๋ฃŒ๋œ ์น˜์•„๊ฐ€ 0.316, ์ž„ํ”Œ๋ž€ํŠธ๊ฐ€ 0.414, ์ž์—ฐ์น˜๊ฐ€ 0.520์ด์—ˆ๋‹ค. ๊ฒฐ ๋ก : ํ‚คํฌ์ธํŠธ ํƒ์ง€ ์‹ ๊ฒฝ๋ง ๋ชจ๋ธ์„ ์ด์šฉํ•˜์—ฌ, ์น˜๊ทผ๋‹จ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์—์„œ ์ž„ํ”Œ๋ž€ํŠธ์˜ ์ฃผ์š” ์ง€์ ์„ ์‚ฌ๋žŒ๊ณผ ๋‹ค์†Œ ์œ ์‚ฌํ•œ ์ˆ˜์ค€์œผ๋กœ ํƒ์ง€ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ๋˜ํ•œ, ํƒ์ง€๋œ ์ง€์ ๋“ค์„ ๊ธฐ๋ฐ˜์œผ๋กœ ๋ฐฉ์‚ฌ์„ ์‚ฌ์ง„์ƒ์—์„œ์˜ ์ž„ํ”Œ๋ž€ํŠธ ์ฃผ์œ„ ๊ณจ ์†Œ์‹ค ๋น„์œจ ๊ณ„์‚ฐ์„ ์ž๋™ํ™”ํ•  ์ˆ˜ ์žˆ๊ณ , ์ด ๊ฐ’์€ ์ž„ํ”Œ๋ž€ํŠธ ์ฃผ์œ„์—ผ์˜ ์‹ฌ๋„ ๋ถ„๋ฅ˜์— ์‚ฌ์šฉ๋  ์ˆ˜ ์žˆ๋‹ค. ํŒŒ๋…ธ๋ผ๋งˆ ์˜์ƒ์—์„œ๋Š” panoptic segmentation์ด ๊ฐ€๋Šฅํ•œ ์‹ ๊ฒฝ๋ง ๋ชจ๋ธ์„ ์ด์šฉํ•˜์—ฌ ์ƒ์•…๋™๊ณผ ํ•˜์•…๊ด€์„ ํฌํ•จํ•œ ์ฃผ์š” ๊ตฌ์กฐ๋ฌผ๋“ค์„ ๊ตฌํšํ™”ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ๋”ฐ๋ผ์„œ, ์ด์™€ ๊ฐ™์ด ๊ฐ ์ž‘์—…์— ๋งž๋Š” ์‹ฌ์ธต์‹ ๊ฒฝ๋ง์„ ์ ์ ˆํ•œ ๋ฐ์ดํ„ฐ๋กœ ํ•™์Šต์‹œํ‚จ๋‹ค๋ฉด ์ง„๋ฃŒ ๋ณด์กฐ ์ˆ˜๋‹จ์œผ๋กœ ํ™œ์šฉ๋  ์ˆ˜ ์žˆ๋‹ค.Purpose: In dentistry, deep neural network models have been applied in areas such as implant classification or lesion detection in radiographs. However, few studies have applied the recently developed keypoint detection model or panoptic segmentation model to medical or dental images. The purpose of this study is to train two neural network models to be used as aids in clinical practice and evaluate them: a model to determine the extent of implant bone loss using keypoint detection in periapical radiographs and a model that segments various structures on panoramic radiographs using panoptic segmentation. Methods: Mask-RCNN, a widely studied convolutional neural network for object detection and instance segmentation, was constructed in a form that is capable of keypoint detection, and trained to detect six points of an implant in a periapical radiograph: left and right of the top, apex, and bone level. Next, a test dataset was used to evaluate the inference results. Object keypoint similarity (OKS), a metric to evaluate the keypoint detection task, and average precision (AP), based on the OKS values, were calculated. Furthermore, the results of the model and those arrived at by a dentist were compared using the mean OKS. Based on the detected keypoint, the peri-implant bone loss ratio was obtained from the radiograph. For panoptic segmentation, Panoptic DeepLab, a neural network model ranked high in the previous benchmark, was trained to segment key structures in panoramic radiographs: maxillary sinus, maxilla, mandibular canal, mandible, natural tooth, treated tooth, and dental implant. Then, each evaluation metric of panoptic, semantic, and instance segmentation was applied to the inference results of the test dataset. Finally, the confusion matrix for the ground truth class of pixels and the class inferred by the model was obtained. Results: The AP of keypoint detection for the average of all OKS thresholds was 0.761 for the upper implants and 0.786 for the lower implants. The mean OKS was 0.8885 for the model and 0.9012 for the dentist; thus, the difference was not statistically significant (p = 0.41). The mean OKS of the model was in the top 66.92% of the normal distribution of human keypoint annotations. In panoramic radiograph segmentation, the average panoptic quality (PQ) of all classes was 80.47. The treated teeth showed the lowest PQ of 57.13, and the mandibular canal showed the second lowest PQ of 65.97. The Intersection over Union (IoU) was 0.795 on average for all classes, where the mandibular canal showed the lowest IoU of 0.639, and the treated tooth showed the second lowest IoU of 0.656. In the confusion matrix, the proportion of correctly inferred pixels among the ground truth pixels was the lowest in the mandibular canal at 0.802. The AP, averaged for all IoU thresholds, was 0.316 for the treated tooth, 0.414 for the dental implant, and 0.520 for the normal tooth. Conclusion: Using the keypoint detection neural network model, it was possible to detect major landmarks around dental implants in periapical radiographs to a degree similar to that of human experts. In addition, it was possible to automate the calculation of the peri-implant bone loss ratio on periapical radiographs based on the detected keypoints, and this value could be used to classify the degree of peri-implantitis. In panoramic radiographs, the major structures including the maxillary sinus and the mandibular canal could be segmented using a neural network model capable of panoptic segmentation. Thus, if deep neural networks suitable for each task are trained using suitable datasets, the proposed approach can be used to assist dental clinicians.Chapter 1. Introduction 1 Chapter 2. Materials and methods 5 Chapter 3. Results 23 Chapter 4. Discussion 32 Chapter 5. Conclusions 45 Published papers related to this study 46 References 47 Abbreviations 52 Abstract in Korean 53 Acknowledgements 56๋ฐ•
    • โ€ฆ
    corecore