2 research outputs found

    Extended Observation Particle Filter with SVD Template Generation Implemented for GPU

    Get PDF
    This work presents a novel template updating strategy based on singular value decomposition (SVD) together with an expansion and extension of previous work combining observations across temporally adjacent frames to implement a likelihood function that provides improvement to velocity refinement in a particle filter tracker. SVD as a novel approach to template generation is used to take advantage of the intuitive notion that the largest singular value corresponds to the highest correlate across template candidates which should more adequately represent the target appearance while rejecting noise and other distractions for use in a correlation based scoring system such as the proposed likelihood function extended across temporally adjacent frames. The tracker is implemented to accelerate computationally expensive operations by moving them to the GPU for processing. This proposed expanded likelihood function provides an improvement of 11.2%-12.5% to particle degeneracy as compared to the previous method in the "Augmented State Vector" approach across the composite of videos. This improvement to particle degeneracy provides for a lower requirement in the number of actual particles necessary for implementation of a particle filter tracker and thus a lower computational requirement while simultaneously providing similar performance. The proposed SVD template generation provides 23.8% increase in time before track loss when comparing the best case in each category of update-by-score and update-by-SVD across the composite of videos. While not bench-marked in this work for quantitative comparison, the use of the GPU with Tensorflow-GPU and Python has allowed the large data set needed for analysis to be obtained in days instead of the months that would have been required in my original proof-of-concept work that was targeted solely for CPU in Matlab

    Object detection in dual-band infrared

    Get PDF
    Dual-Band Infrared (DBIR) offers the advantage of combining Mid-Wave Infrared (MWIR) and Long-Wave Infrared (LWIR) within a single field-of-view (FoV). This provides additional information for each spectral band. DBIR camera systems find applications in both military and civilian contexts. This work introduces a novel labeled DBIR dataset that includes civilian vehicles, aircraft, birds, and people. The dataset is designed for utilization in object detection and tracking algorithms. It comprises 233 objects with tracks spanning up to 1,300 frames, encompassing images in both MW and LW. This research reviews pertinent literature related to object detection, object detection in the infrared spectrum, and data fusion. Two sets of experiments were conducted using this DBIR dataset: Motion Detection and CNNbased object detection. For motion detection, a parallel implementation of the Visual Background Extractor (ViBe) was developed, employing ConnectedComponents analysis to generate bounding boxes. To assess these bounding boxes, Intersection-over-Union (IoU) calculations were performed. The results demonstrate that DBIR enhances the IoU of bounding boxes in 6.11% of cases within sequences where the camera’s field of view remains stationary. A size analysis reveals ViBe’s effectiveness in detecting small and dim objects within this dataset. A subsequent experiment employed You Only Look Once (YOLO) versions 4 and 7 to conduct inference on this dataset, following image preprocessing. The inference models were trained using visible spectrum MS COCO data. The findings confirm that YOLOv4/7 effectively detect objects within the infrared spectrum in this dataset. An assessment of these CNNs’ performance relative to the size of the detected object highlights the significance of object size in detection capabilities. Notably, DBIR substantially enhances detection capabilities in both YOLOv4 and YOLOv7; however, in the latter case, the number of False Positive detections increases. Consequently, while DBIR improves the recall of YOLOv4/7, the introduction of DBIR information reduces the precision of YOLOv7. This study also demonstrates the complementary nature of ViBe and YOLO in their detection capabilities based on object size in this data set. Though this is known prior art, an approach using these two approaches in a hybridized configuration is discussed. ViBe excels in detecting small, distant objects, while YOLO excels in detecting larger, closer objects. The research underscores that DBIR offers multiple advantages over MW or LW alone in modern computer vision algorithms, warranting further research investment
    corecore