19 research outputs found

    ShuffleDet: Real-Time Vehicle Detection Network in On-board Embedded UAV Imagery

    Get PDF
    On-board real-time vehicle detection is of great significance for UAVs and other embedded mobile platforms. We propose a computationally inexpensive detection network for vehicle detection in UAV imagery which we call ShuffleDet. In order to enhance the speed-wise performance, we construct our method primarily using channel shuffling and grouped convolutions. We apply inception modules and deformable modules to consider the size and geometric shape of the vehicles. ShuffleDet is evaluated on CARPK and PUCPR+ datasets and compared against the state-of-the-art real-time object detection networks. ShuffleDet achieves 3.8 GFLOPs while it provides competitive performance on test sets of both datasets. We show that our algorithm achieves real-time performance by running at the speed of 14 frames per second on NVIDIA Jetson TX2 showing high potential for this method for real-time processing in UAVs.Comment: Accepted in ECCV 2018, UAVision 201

    SISTEM INFORMASI PARKIR MENGGUNAKAN TEKNIK OBJECT TRACKING DAN OBJECT COUNTING

    Get PDF
    Fasilitas parkir merupakan hal yang sangat penting untuk para pengguna kendaraan, khususnya kendaraan beroda empat. Di sebagian besar negara, mobil adalah moda transportasi yang dominan, bahkan diperkirakan pada tahun 2050 mendatang benua Asia akan mengalami pertumbuhan penggunaan mobil pribadi sebesar 40%. Dalam memaksimalkan efisiensi penggunaan lahan parkir, diperlukan sistem informasi parkir yang memberikan informasi ketersediaan tempat parkir untuk mempermudah pengendara menggunakan fasilitas parkir. Namun demikian, permasalahan terkait hunian parkir masih sering terjadi, seperti dari tidak adanya informasi lahan parkir yang tersedia, lambatnya informasi ketersediaan lahan parkir dan minimnya informasi terkait hunian parkir. Ditambah lagi dengan peningkatan kepemilikan mobil, menyebabkan kurangnya area parkir mobil karena tidak seimbangnya antara ketersediaan parkir dan kebutuhan parkir. Berbagai sistem berbasis sensor maupun vision dirancang untuk mengatasi masalah tersebut, namun masih terdapat kekurangan dari kinerja sistem dalam hal akurasi dan kecepatan pemrosesan. Untuk mengatasi permasalahan tersebut, penelitian ini mengusulkan sistem menggunakan model YOLOv8 dengan teknik object tracking serta object counting. Teknik object tracking digunakan untuk meningkatkan akurasi dan kestabilan deteksi dalam setiap frame khususnya dalam video. Kemudian object counting dimanfaatkan dengan membuat zona deteksi untuk meningkatkan keakuratan penghitungan ketersediaan lahan parkir. Informasi yang dihasilkan dari proses deteksi, tracking, dan counting disimpan pada database lokal dan ditampilkan pada aplikasi berbasis web sehingga membantu mengetahui informasi ketersediaan parkir. Pada penelitian ini didapatkan hasil bahwa sistem yang dirancang memiliki kinerja baik dan dapat menandingi sistem yang dibangun oleh peneliti lainnya dengan akurasi rata-rata sebesar 98,67%, latensi sekitar 20 milidetik, dan 50 frame per second. ; Parking facilities are very important for vehicle users, especially four-wheeled vehicles. In most countries, cars are the dominant mode of transportation, it is estimated that by 2050 the Asian continent will experience a growth in private car use of 40%. In maximizing the efficiency of the use of parking space, a parking information system is needed, prodividng information about the availability of parking spaces to make it easier for drivers to use parking facilities. However, problems related to parking occupancy still occur frequently, such as the unavailability of parking space information, the speed of parking space information retrieved, and the lack of information about parking occupancy. Moreover, the increase in car ownership causes a shortage of car parking areas due to an imbalance between parking availability and parking needs. Various sensor and vision-based systems are designed to overcome these problems, but there are still deficiencies in system performance in terms of accuracy and processing speed. To overcome these problems, this study proposes a system using the YOLOv8 model with object tracking and object counting techniques. Object tracking techniques are used to improve the accuracy and stability of detection in each frame, especially in video. Then object counting is utilized by creating detection zones to improve the accuracy of calculating parking space availability. The information generated from the detection, tracking and counting processes is stored in a local database and displayed on a web-based application that helps determine parking availability information. In this study, the results showed that the designed system had good performance and could compete with systems built by other researchers with an average accuracy of 98.67%, a latency of around 20 milliseconds, and 50 frames per second

    Counting and Locating High-Density Objects Using Convolutional Neural Network

    Full text link
    This paper presents a Convolutional Neural Network (CNN) approach for counting and locating objects in high-density imagery. To the best of our knowledge, this is the first object counting and locating method based on a feature map enhancement and a Multi-Stage Refinement of the confidence map. The proposed method was evaluated in two counting datasets: tree and car. For the tree dataset, our method returned a mean absolute error (MAE) of 2.05, a root-mean-squared error (RMSE) of 2.87 and a coefficient of determination (R2^2) of 0.986. For the car dataset (CARPK and PUCPR+), our method was superior to state-of-the-art methods. In the these datasets, our approach achieved an MAE of 4.45 and 3.16, an RMSE of 6.18 and 4.39, and an R2^2 of 0.975 and 0.999, respectively. The proposed method is suitable for dealing with high object-density, returning a state-of-the-art performance for counting and locating objects.Comment: 15 pages, 10 figures, 8 table

    MOR-UAV: A Benchmark Dataset and Baselines for Moving Object Recognition in UAV Videos

    Full text link
    Visual data collected from Unmanned Aerial Vehicles (UAVs) has opened a new frontier of computer vision that requires automated analysis of aerial images/videos. However, the existing UAV datasets primarily focus on object detection. An object detector does not differentiate between the moving and non-moving objects. Given a real-time UAV video stream, how can we both localize and classify the moving objects, i.e. perform moving object recognition (MOR)? The MOR is one of the essential tasks to support various UAV vision-based applications including aerial surveillance, search and rescue, event recognition, urban and rural scene understanding.To the best of our knowledge, no labeled dataset is available for MOR evaluation in UAV videos. Therefore, in this paper, we introduce MOR-UAV, a large-scale video dataset for MOR in aerial videos. We achieve this by labeling axis-aligned bounding boxes for moving objects which requires less computational resources than producing pixel-level estimates. We annotate 89,783 moving object instances collected from 30 UAV videos, consisting of 10,948 frames in various scenarios such as weather conditions, occlusion, changing flying altitude and multiple camera views. We assigned the labels for two categories of vehicles (car and heavy vehicle). Furthermore, we propose a deep unified framework MOR-UAVNet for MOR in UAV videos. Since, this is a first attempt for MOR in UAV videos, we present 16 baseline results based on the proposed framework over the MOR-UAV dataset through quantitative and qualitative experiments. We also analyze the motion-salient regions in the network through multiple layer visualizations. The MOR-UAVNet works online at inference as it requires only few past frames. Moreover, it doesn't require predefined target initialization from user. Experiments also demonstrate that the MOR-UAV dataset is quite challenging
    corecore