2,653 research outputs found

    The AAU Multimodal Annotation Toolboxes: Annotating Objects in Images and Videos

    Get PDF
    This tech report gives an introduction to two annotation toolboxes that enable the creation of pixel and polygon-based masks as well as bounding boxes around objects of interest. Both toolboxes support the annotation of sequential images in the RGB and thermal modalities. Each annotated object is assigned a classification tag, a unique ID, and one or more optional meta data tags. The toolboxes are written in C++ with the OpenCV and Qt libraries and are operated by using the visual interface and the extensive range of keyboard shortcuts. Pre-built binaries are available for Windows and MacOS and the tools can be built from source under Linux as well. So far, tens of thousands of frames have been annotated using the toolboxes.Comment: 6 pages, 10 figure

    Effect of Annotation Errors on Drone Detection with YOLOv3

    Full text link
    Following the recent advances in deep networks, object detection and tracking algorithms with deep learning backbones have been improved significantly; however, this rapid development resulted in the necessity of large amounts of annotated labels. Even if the details of such semi-automatic annotation processes for most of these datasets are not known precisely, especially for the video annotations, some automated labeling processes are usually employed. Unfortunately, such approaches might result with erroneous annotations. In this work, different types of annotation errors for object detection problem are simulated and the performance of a popular state-of-the-art object detector, YOLOv3, with erroneous annotations during training and testing stages is examined. Moreover, some inevitable annotation errors in CVPR-2020 Anti-UAV Challenge dataset is also examined in this manner, while proposing a solution to correct such annotation errors of this valuable data set.Comment: Best Paper Award at The 1st Anti-UAV Workshop & Challenge - CVPR Workshops, 202

    Implementation of Unmanned aerial vehicles (UAVs) for assessment of transportation infrastructure - Phase II

    Get PDF
    Technological advances in unmanned aerial vehicle (UAV) technologies continue to enable these tools to become easier to use, more economical, and applicable for transportation-related operations, maintenance, and asset management while also increasing safety and decreasing cost. This Phase 2 project continued to test and evaluate five main UAV platforms with a combination of optical, thermal, and lidar sensors to determine how to implement them into MDOT workflows. Field demonstrations were completed at bridges, a construction site, road corridors, and along highways with data being processed and analyzed using customized algorithms and tools. Additionally, a cost-benefit analysis was conducted, comparing manual and UAV-based inspection methods. The project team also gave a series of technical demonstrations and conference presentations, enabling outreach to interested audiences who gained understanding of the potential implementation of this technology and the advanced research that MDOT is moving to implementation. The outreach efforts and research activities performed under this contract demonstrated how implementing UAV technologies into MDOT workflows can provide many benefits to MDOT and the motoring public; such as advantages in improved cost-effectiveness, operational management, and timely maintenance of Michigan’s transportation infrastructure

    Rain Removal in Traffic Surveillance: Does it Matter?

    Get PDF
    Varying weather conditions, including rainfall and snowfall, are generally regarded as a challenge for computer vision algorithms. One proposed solution to the challenges induced by rain and snowfall is to artificially remove the rain from images or video using rain removal algorithms. It is the promise of these algorithms that the rain-removed image frames will improve the performance of subsequent segmentation and tracking algorithms. However, rain removal algorithms are typically evaluated on their ability to remove synthetic rain on a small subset of images. Currently, their behavior is unknown on real-world videos when integrated with a typical computer vision pipeline. In this paper, we review the existing rain removal algorithms and propose a new dataset that consists of 22 traffic surveillance sequences under a broad variety of weather conditions that all include either rain or snowfall. We propose a new evaluation protocol that evaluates the rain removal algorithms on their ability to improve the performance of subsequent segmentation, instance segmentation, and feature tracking algorithms under rain and snow. If successful, the de-rained frames of a rain removal algorithm should improve segmentation performance and increase the number of accurately tracked features. The results show that a recent single-frame-based rain removal algorithm increases the segmentation performance by 19.7% on our proposed dataset, but it eventually decreases the feature tracking performance and showed mixed results with recent instance segmentation methods. However, the best video-based rain removal algorithm improves the feature tracking accuracy by 7.72%.Comment: Published in IEEE Transactions on Intelligent Transportation System

    Anomaly detection in fleet service vehicles: improving object segmentation

    Get PDF
    Dissertação de mestrado integrado em Engenharia InformáticaThe present dissertation is inserted in a BOSCH project in which the global focus is au tonomous driving. The project is divided in multiple phases, being the main focus of this dissertation object detection and segmentation inside fleet service vehicles. The objective is to detect/segment objects and dirt left inside a vehicle, warning the commuter if they forgot an object inside or the administrators if the vehicles need to be cleaned. To train the models, BOSCH provided an initial dataset containing a small set of annotated images. This dataset contains pictures of a vehicles cockpit with many diverse objects and dirt. One of the goals for BOSCH is to increment this dataset with more images. Hence, in this project several state of the art segmentation methods were thoroughly studied and analysed, with two of them being selected for further exploration: DeepExtremeCut and FgSegNet v2. The main objective is to see to what extent can these methods be used in a semi automatic process to segment more images, thereby increasing the initial dataset. DeepExtremeCut works by using a framework in which, after the model is trained, it allows us to click on four extreme points in the desired object, producing the segmentation. This method produced reliable segmentations, however it requires human intervention both for the initial segmentation and verification of the output. Hence, it was not regarded as a good solution for a future augmentation of the BOSCH dataset. Regarding FgSegNet v2, this later method does not require any initial annotation of the input images. Under this approach only a final verification and possible rectification is required. Therefore, this method meets the requirements defined by BOSCH for a dataset expansion solution. An ablation study is also presented for FgSegNet v2, analysing its three stages: (i) Encoder, (ii) Feature Pooling Module and (iii) Decoder. The result of this study is a proposal of a variation of the aforementioned method called Mod FgSegNet. It was also compared with state of the art methods in public datasets. Three datasets are used for testing: CDNet2014, SBI2015 and CityScapes. In CDNet2014 we got an overall improvement when compared to the state of the art, particularly in the LowFrameRate subset. Regarding SBI2015 the overall results are lower in comparison with the top state of art, while in CityScapes some promising results are presented.A presente dissertação está inserida num projeto da BOSCH, em que o foco global é a condução autónoma. Este projeto foi dividido em múltiplas fases, sendo o foco principal desta dissertação e a detecção e a segmentação de objetos. O objetivo é detectar / segmentar objetos e lixo deixados no interior de um veículo, avisando o passageiro se ele se esqueceu de algum objeto no interior ou os administradores se os veículos precisam de ser limpos. Para lidar com este tópico, vários métodos de segmentação do estado da arte foram exaustivamente estudados e analisados, nos quais dois deles em particular foram explorados, ou seja, o DeepExtremeCut e o FgSegNet. Algumas melhorias foram feitas no desempenho deste último, permitindo um potencial aumento semiautomático no tamanho de um dataset fornecido pela BOSCH, uma vez que não possuía imagens suficientes. Este dataset contém fotos da cabine de um veículo com diferentes objetos e lixo. O DeepExtremeCut funciona através do uso de uma framework no qual, após o treino do modelo, permite clicar em quatro pontos extremos do objeto desejado, produzindo a segmentação. Este método produz segmentações confiáveis, embora não corresponda a uma segmentação automática, visto que existe a necessidade de selecionar todos os objetos, ainda pode ser útil quando a segmentação automática de um método diferente não estiver a funcionar em casos particulares. Em relação ao FgSegNet v2, e apresentado um estudo de ablação, sendo feita uma análise das suas três etapas: (i) Encoder, (ii) Feate Pooling Module e (iii) Decoder. O resultado deste estudo e uma proposta de variação do referido método no dataset da BOSCH, de forma a utilizá-lo potencialmente em vários projetos dentro da empresa, chamado Mod_FgSegNet. Também foi comparado com métodos do estado de arte em datasets públicos. Os três datasets usados para teste são: CDNet 2014, SBI2015 e CityScapes. No CDNet2014, obtivemos uma melhoria geral em comparação com o estado da arte, principalmente no subconjunto LowFrameRate. Em relação ao ao SBI2015, os resultados gerais foram inferiores em comparação com o estado da arte de ponta, enquanto que no CityScapes alguns resultados promissores foram apresentados
    corecore