47 research outputs found

    Vision-Based 2D and 3D Human Activity Recognition

    Get PDF

    Memory-Based Active Visual Search for Humanoid Robots

    Get PDF

    Automated interpretation of benthic stereo imagery

    Get PDF
    Autonomous benthic imaging, reduces human risk and increases the amount of collected data. However, manually interpreting these high volumes of data is onerous, time consuming and in many cases, infeasible. The objective of this thesis is to improve the scientific utility of the large image datasets. Fine-scale terrain complexity is typically quantified by rugosity and measured by divers using chains and tape measures. This thesis proposes a new technique for measuring terrain complexity from 3D stereo image reconstructions, which is non-contact and can be calculated at multiple scales over large spatial extents. Using robots, terrain complexity can be measured without endangering humans, beyond scuba depths. Results show that this approach is more robust, flexible and easily repeatable than traditional methods. These proposed terrain complexity features are combined with visual colour and texture descriptors and applied to classifying imagery. New multi-dataset feature selection methods are proposed for performing feature selection across multiple datasets, and are shown to improve the overall classification performance. The results show that the most informative predictors of benthic habitat types are the new terrain complexity measurements. This thesis presents a method that aims to reduce human labelling effort, while maximising classification performance by combining pre-clustering with active learning. The results support that utilising the structure of the unlabelled data in conjunction with uncertainty sampling can significantly reduce the number of labels required for a given level of accuracy. Typically 0.00001–0.00007% of image data is annotated and processed for science purposes (20–50 points in 1–2% of the images). This thesis proposes a framework that uses existing human-annotated point labels to train a superpixel-based automated classification system, which can extrapolate the classified results to every pixel across all the images of an entire survey

    Automated interpretation of benthic stereo imagery

    Get PDF
    Autonomous benthic imaging, reduces human risk and increases the amount of collected data. However, manually interpreting these high volumes of data is onerous, time consuming and in many cases, infeasible. The objective of this thesis is to improve the scientific utility of the large image datasets. Fine-scale terrain complexity is typically quantified by rugosity and measured by divers using chains and tape measures. This thesis proposes a new technique for measuring terrain complexity from 3D stereo image reconstructions, which is non-contact and can be calculated at multiple scales over large spatial extents. Using robots, terrain complexity can be measured without endangering humans, beyond scuba depths. Results show that this approach is more robust, flexible and easily repeatable than traditional methods. These proposed terrain complexity features are combined with visual colour and texture descriptors and applied to classifying imagery. New multi-dataset feature selection methods are proposed for performing feature selection across multiple datasets, and are shown to improve the overall classification performance. The results show that the most informative predictors of benthic habitat types are the new terrain complexity measurements. This thesis presents a method that aims to reduce human labelling effort, while maximising classification performance by combining pre-clustering with active learning. The results support that utilising the structure of the unlabelled data in conjunction with uncertainty sampling can significantly reduce the number of labels required for a given level of accuracy. Typically 0.00001–0.00007% of image data is annotated and processed for science purposes (20–50 points in 1–2% of the images). This thesis proposes a framework that uses existing human-annotated point labels to train a superpixel-based automated classification system, which can extrapolate the classified results to every pixel across all the images of an entire survey

    ACCURATE AND FAST STEREO VISION

    Get PDF
    Stereo vision from short-baseline image pairs is one of the most active research fields in computer vision. The estimation of dense disparity maps from stereo image pairs is still a challenging task and there is further space for improving accuracy, minimizing the computational cost and handling more efficiently outliers, low-textured areas, repeated textures, disparity discontinuities and light variations. This PhD thesis presents two novel methodologies relating to stereo vision from short-baseline image pairs: I. The first methodology combines three different cost metrics, defined using colour, the CENSUS transform and SIFT (Scale Invariant Feature Transform) coefficients. The selected cost metrics are aggregated based on an adaptive weights approach, in order to calculate their corresponding cost volumes. The resulting cost volumes are merged into a combined one, following a novel two-phase strategy, which is further refined by exploiting semi-global optimization. A mean-shift segmentation-driven approach is exploited to deal with outliers in the disparity maps. Additionally, low-textured areas are handled using disparity histogram analysis, which allows for reliable disparity plane fitting on these areas. II. The second methodology relies on content-based guided image filtering and weighted semi-global optimization. Initially, the approach uses a pixel-based cost term that combines gradient, Gabor-Feature and colour information. The pixel-based matching costs are filtered by applying guided image filtering, which relies on support windows of two different sizes. In this way, two filtered costs are estimated for each pixel. Among the two filtered costs, the one that will be finally assigned to each pixel, depends on the local image content around this pixel. The filtered cost volume is further refined by exploiting weighted semi-global optimization, which improves the disparity accuracy. The handling of the occluded areas is enhanced by incorporating a straightforward and time efficient scheme. The evaluation results show that both methodologies are very accurate, since they handle efficiently low-textured/occluded areas and disparity discontinuities. Additionally, the second approach has very low computational complexity. Except for the aforementioned two methodologies that use as input short-baseline image pairs, this PhD thesis presents a novel methodology for generating 3D point clouds of good accuracy from wide-baseline stereo pairs

    Similarity Measure Based on Entropy and Census and Multi-Resolution Disparity Estimation Technique for Stereo Matching

    Get PDF
    Stereo matching is one of the most active research areas in the field of computer vision. Stereo matching aims to obtain 3D information by extracting correct correspondence between two images captured from different point of views. There are two research parts in stereo matching: similarity measure between correspondence points and optimization technique for dence disparity estimation. The crux of stereo matching problem in similarity measure perspective is how to deal with the inferent points ambiguity that results from the ambiguous local appearances of image points. Similarity measures in stereo matching are classified as feature-based, intensity-based or non-parametric measure. And most similarity measures in the literatures are based on pixel intensity comparison. When images are taken at different illumination conditions or different sensors used, it is very unlikely that the corresponding pixels would have the same intensity creating false correspondences if it is only based on intensity matching functions alone. Especially illumination variations between input images can cause serious degrade in the performance of stereo matching algorithms. In this situation, mutual information-based method is powerful. However, it is still ambiguous or erroneous in considering local illumination variations between images. Therefore, similarity measure to these radiometric variations are demanded and become inevitable for stereo matching. Optimization method in stereo matching can be classified into two categories: local and global optimization methods, and most state-of-the-art algorithms fall into global optimization method. Global optimization methods can greatly suppress the matching ambiguities caused by various factors such as occluded and textureless regions. However, They are usually computationally expensive due to the slow-converging optimization process. In this paper, it was proposed that a stereo matching similarity measure based on entropy and census transform and an optimization technique using dynamic programming to estimate disparity efficiently based on multi-resolution method. Proposed similarity measure is composed of entropy, Haar wavelet feature vector, and modified Census transform. In general, mutual information similarity measure based on entropy about stereo images and disparity map is a popular and powerful similarity measure which is robust to complex intensity transformation. However, it is still ambiguous or erroneous with local radiometric variations, since it only accounts for global variation between images, and does not contain spatial information. Haar wavelet response can express frequency properties of image regions and is robust to various intensity changes and bias. Therefore, entropy was utilized with Haar wavelet feature vector as geometric measure. Modified Census transform was used as another spatial similarity measure. Census transform is a well-known non-parametric measure. And it is powerful to textureless and disparity discontinuity region and robust to noisy environment. A combination of entropy with Haar wavelet feature vector and modified Census transform as similarity measure was proposed to find correspondence. It is invariant to local radiometric variations and global illumination changes, so it can be applied to find correspondence for images which undergo local as well as global radiometric variations. Proposed optimization method is a new disparity estimation technique based on dynamic programming. A method using dynamic programming with 8-direction energy aggregation to estimate accurate disparity map was applied. Using 8-direction energy aggregation, accurate disparities can be found at disparity discontinuous region and suppress a streaking phenomenon in disparity map. Finally, the multi-resolution scheme was proposed to increase efficiency while processing and disparity estimation method. A Gaussian pyramid which prevent the ailasing at low-resolution image pyramid levels was used. And the multi-resolution scheme was proposed to perform matching at every levels to find accurate disparity. This method can perform matching efficiently and make accurate disparity map. And proposed method was validated with experimental results on stereo images.제 1 μž₯ μ„œλ‘ ...........................................1 1.1 연ꡬ λͺ©μ  및 λ°°κ²½...............................1 1.2κ΄€λ ¨ 연ꡬ........................................3 1.3연ꡬ λ‚΄μš©........................................6 1.4λ…Όλ¬Έμ˜ ꡬ성......................................7 제 2 μž₯ μŠ€ν…Œλ ˆμ˜€ μ‹œκ°κ³Ό μŠ€ν…Œλ ˆμ˜€ μ •ν•©..................8 2.1 μŠ€ν…Œλ ˆμ˜€ μ‹œκ°...................................8 2.2μŠ€ν…Œλ ˆμ˜€μ •ν•©....................................10 2.2.1 μœ μ‚¬λ„ 척도.................................10 2.2.2 μ΅œμ ν™” 방법.................................15 2.3 ν™˜κ²½ 변화에 κ°•μΈν•œ μœ μ‚¬λ„ 척도.................18 2.3.1 νŠΉμ§• 기반 μœ μ‚¬λ„ 척도.......................18 2.3.2 λͺ…암도 기반 μœ μ‚¬λ„ 척도.....................19 2.3.3 λΉ„λͺ¨μˆ˜ μœ μ‚¬λ„ 척도..........................22 제 3 μž₯ μ—”νŠΈλ‘œν”Ό 및 Census 기반의 μœ μ‚¬λ„ 척도.........24 3.1 μ—”νŠΈλ‘œν”Ό 기반의 μœ μ‚¬λ„ 척도....................26 3.1.1 μ—”νŠΈλ‘œν”Ό....................................26 3.1.2 μ—”νŠΈλ‘œν”Όλ₯Ό μ΄μš©ν•œ MI μœ μ‚¬λ„ 척도............27 3.2 μ œμ•ˆν•œ Haar 웨이블렛 νŠΉμ§•μ„ κ²°ν•©ν•œ μ—”νŠΈλ‘œν”Ό μœ μ‚¬λ„μ²™λ„.................................................28 3.2.1 ν™”μ†Œ λ‹¨μœ„ μ—”νŠΈλ‘œν”Ό.........................28 3.2.2 Haar웨이블렛 νŠΉμ§•μ„ κ²°ν•©ν•œ μ—”νŠΈλ‘œν”Ό........35 3.3 μ œμ•ˆν•œCensus λ³€ν™˜ 기반의 μœ μ‚¬λ„ 척도..........46 3.3.1 Census λ³€ν™˜................................46 3.3.2 μ œμ•ˆν•œ Census λ³€ν™˜μ„ μ΄μš©ν•œ μœ μ‚¬λ„ 척도....49 제 4 μž₯ 8λ°©ν–₯ 동적 κ³„νšλ²•μ„ μ΄μš©ν•œ λ³€μœ„μΆ”μ •..........53 4.1 동적 κ³„νšλ²•...................................53 4.2 μ œμ•ˆν•œ 8λ°©ν–₯ 동적 κ³„νšλ²•......................57 제 5 μž₯ 닀해상도 기반의 μŠ€ν…Œλ ˆμ˜€ μ •ν•©................67 5.1 κ°€μš°μ‹œμ•ˆ μ˜μƒ ν”ΌλΌλ―Έλ“œ........................67 5.2 μ œμ•ˆν•œ 닀해상도 기반 μŠ€ν…Œλ ˆμ˜€ μ •ν•©............71 제 6 μž₯ μ‹€ν—˜ 및 κ³ μ°°.................................77 6.1 μ •ν•© μ„±λŠ₯ 평가 방법...........................77 6.2 μŠ€ν…Œλ ˆμ˜€ μ •ν•© μ‹€ν—˜............................79 6.2.1 RDS μ˜μƒ μ‹€ν—˜..............................79 6.2.2 ν™˜κ²½ λ³€ν™”κ°€ μ—†λŠ” ν‘œμ€€ μ˜μƒ μ‹€ν—˜............84 6.2.3 ν™˜κ²½ λ³€ν™”κ°€ λ°œμƒν•œ ν‘œμ€€ μ˜μƒ μ‹€ν—˜..........92 6.2.4 μ‹€μ œ μ˜μƒ μ‹€ν—˜............................110 6.3 계산 속도....................................118 6.4 μ œμ•ˆν•œ λ°©λ²•μ˜ μ •ν•© μ„±λŠ₯에 λŒ€ν•œ κ³ μ°°..........121 제 7 μž₯ κ²° λ‘ .....................................123 μ°Έκ³  λ¬Έν—Œ...........................................12

    Connected Attribute Filtering Based on Contour Smoothness

    Get PDF
    corecore