375,830 research outputs found

    Commercial Flight Crew Decision-Making during Low-Visibility Approach Operations Using Fused Synthetic/Enhanced Vision Systems

    Get PDF
    NASA is investigating revolutionary crew-vehicle interface technologies that strive to proactively overcome aircraft safety barriers that would otherwise constrain the full realization of the next-generation air transportation system. A fixed-based piloted simulation experiment was conducted to evaluate the complementary use of Synthetic and Enhanced Vision technologies. Specific focus was placed on new techniques for integration and/or fusion of Enhanced and Synthetic Vision and its impact within a two-crew flight deck on the crew's decision-making process during low-visibility approach and landing operations. Overall, the experimental data showed that significant improvements in situation awareness, without concomitant increases in workload and display clutter, could be provided by the integration and/or fusion of synthetic and enhanced vision technologies for the pilot-flying and the pilot-not-flying. During non-normal operations, the ability of the crew to handle substantial navigational errors and runway incursions were neither improved nor adversely impacted by the display concepts. The addition of Enhanced Vision may not, unto itself, provide an improvement in runway incursion detection without being specifically tailored for this application. Existing enhanced vision system procedures were effectively used in the crew decision-making process during approach and missed approach operations but having to forcibly transition from an excellent FLIR image to natural vision by 100 ft above field level was awkward for the pilot-flying

    Long-term Robust Tracking whith on Failure Recovery

    Get PDF
    This article aims at a new algorithm for tracking moving objects in the long term. We have tried to overcome some potential difficulties, first by a comparative study of the measuring methods of the difference and the similarity between the template and the source image. In the second part, an improvement of the best method allows us to follow the target in a robust way. This method also allows us to effectively overcome the problems of geometric deformation, partial occlusion and recovery after the target leaves the field of vision. The originality of our algorithm is based on a new model, which does not depend on a probabilistic process and does not require a data based detection in advance. Experimental results on several difficult video sequences have proven performance advantages over many recent trackers. The developed algorithm can be employed in several applications such as video surveillance, active vision or industrial visual servoing

    Gaussian Processes with Context-Supported Priors for Active Object Localization

    Full text link
    We devise an algorithm using a Bayesian optimization framework in conjunction with contextual visual data for the efficient localization of objects in still images. Recent research has demonstrated substantial progress in object localization and related tasks for computer vision. However, many current state-of-the-art object localization procedures still suffer from inaccuracy and inefficiency, in addition to failing to provide a principled and interpretable system amenable to high-level vision tasks. We address these issues with the current research. Our method encompasses an active search procedure that uses contextual data to generate initial bounding-box proposals for a target object. We train a convolutional neural network to approximate an offset distance from the target object. Next, we use a Gaussian Process to model this offset response signal over the search space of the target. We then employ a Bayesian active search for accurate localization of the target. In experiments, we compare our approach to a state-of-theart bounding-box regression method for a challenging pedestrian localization task. Our method exhibits a substantial improvement over this baseline regression method.Comment: 10 pages, 4 figure

    Res2Net: A New Multi-scale Backbone Architecture

    Full text link
    Representing features at multiple scales is of great importance for numerous vision tasks. Recent advances in backbone convolutional neural networks (CNNs) continually demonstrate stronger multi-scale representation ability, leading to consistent performance gains on a wide range of applications. However, most existing methods represent the multi-scale features in a layer-wise manner. In this paper, we propose a novel building block for CNNs, namely Res2Net, by constructing hierarchical residual-like connections within one single residual block. The Res2Net represents multi-scale features at a granular level and increases the range of receptive fields for each network layer. The proposed Res2Net block can be plugged into the state-of-the-art backbone CNN models, e.g., ResNet, ResNeXt, and DLA. We evaluate the Res2Net block on all these models and demonstrate consistent performance gains over baseline models on widely-used datasets, e.g., CIFAR-100 and ImageNet. Further ablation studies and experimental results on representative computer vision tasks, i.e., object detection, class activation mapping, and salient object detection, further verify the superiority of the Res2Net over the state-of-the-art baseline methods. The source code and trained models are available on https://mmcheng.net/res2net/.Comment: 11 pages, 7 figure

    Refactoring, reengineering and evolution: paths to Geant4 uncertainty quantification and performance improvement

    Full text link
    Ongoing investigations for the improvement of Geant4 accuracy and computational performance resulting by refactoring and reengineering parts of the code are discussed. Issues in refactoring that are specific to the domain of physics simulation are identified and their impact is elucidated. Preliminary quantitative results are reported.Comment: To be published in the Proc. CHEP (Computing in High Energy Physics) 201

    The effect of experience and of dots\u2019 density and duration on the detection of coherent motion in dogs

    Get PDF
    Knowledge about the mechanisms underlying canine vision is far from being exhaustive, especially that concerning post- retinal elaboration. One aspect that has received little attention is motion perception, and in spite of the common belief that dogs are extremely apt at detecting moving stimuli, there is no scientific support for such an assumption. In fact, we recently showed that dogs have higher thresholds than humans for coherent motion detection (Kanizsar et al. in Sci Rep UK 7:11259, 2017). This term refers to the ability of the visual system to perceive several units moving in the same direction, as one coherently moving global unit. Coherent motion perception is commonly investigated using random dot displays, containing variable proportions of coherently moving dots. Here, we investigated the relative contribution of local and global integra- tion mechanisms for coherent motion perception, and changes in detection thresholds as a result of repeated exposure to the experimental stimuli. Dogs who had been involved in the previous study were given a conditioned discrimination task, in which we systematically manipulated dot density and duration and, eventually, re-assessed our subjects\u2019 threshold after extensive exposure to the stimuli. Decreasing dot duration impacted on dogs\u2019 accuracy in detecting coherent motion only at very low duration values, revealing the efficacy of local integration mechanisms. Density impacted on dogs\u2019 accuracy in a linear fashion, indicating less efficient global integration. There was limited evidence of improvement in the re-assessment but, with an average threshold at re-assessment of 29%, dogs\u2019 ability to detect coherent motion remains much poorer than that of humans

    Lifting schemes for joint coding of stereoscopic pairs of satellite images

    Get PDF
    electronic version (5 pp.)International audienceStereo data compression is an important issue for the new generation of vision systems. In this paper, we are interested in lossless coding methods for stereo images allowing progressive reconstruction. Most of the existing approaches account for the mutual similarities between the left and the right images. More precisely, the disparity compensation process consists in predicting the right image from the left one based on the disparity map. Then, the disparity map, the reference image, and the residual image are encoded. In this work, we propose a novel approach based on the concept of vector lifting scheme. Its main feature is that it does not generate one residual image but two compact multiresolution representations of the left and the right views, driven by the underlying disparity map. Experimental results show a significant improvement using this technique compared with conventional methods

    S-OHEM: Stratified Online Hard Example Mining for Object Detection

    Full text link
    One of the major challenges in object detection is to propose detectors with highly accurate localization of objects. The online sampling of high-loss region proposals (hard examples) uses the multitask loss with equal weight settings across all loss types (e.g, classification and localization, rigid and non-rigid categories) and ignores the influence of different loss distributions throughout the training process, which we find essential to the training efficacy. In this paper, we present the Stratified Online Hard Example Mining (S-OHEM) algorithm for training higher efficiency and accuracy detectors. S-OHEM exploits OHEM with stratified sampling, a widely-adopted sampling technique, to choose the training examples according to this influence during hard example mining, and thus enhance the performance of object detectors. We show through systematic experiments that S-OHEM yields an average precision (AP) improvement of 0.5% on rigid categories of PASCAL VOC 2007 for both the IoU threshold of 0.6 and 0.7. For KITTI 2012, both results of the same metric are 1.6%. Regarding the mean average precision (mAP), a relative increase of 0.3% and 0.5% (1% and 0.5%) is observed for VOC07 (KITTI12) using the same set of IoU threshold. Also, S-OHEM is easy to integrate with existing region-based detectors and is capable of acting with post-recognition level regressors.Comment: 9 pages, 3 figures, accepted by CCCV 201
    corecore