8 research outputs found

    LIDAR-Camera Fusion for Road Detection Using Fully Convolutional Neural Networks

    Full text link
    In this work, a deep learning approach has been developed to carry out road detection by fusing LIDAR point clouds and camera images. An unstructured and sparse point cloud is first projected onto the camera image plane and then upsampled to obtain a set of dense 2D images encoding spatial information. Several fully convolutional neural networks (FCNs) are then trained to carry out road detection, either by using data from a single sensor, or by using three fusion strategies: early, late, and the newly proposed cross fusion. Whereas in the former two fusion approaches, the integration of multimodal information is carried out at a predefined depth level, the cross fusion FCN is designed to directly learn from data where to integrate information; this is accomplished by using trainable cross connections between the LIDAR and the camera processing branches. To further highlight the benefits of using a multimodal system for road detection, a data set consisting of visually challenging scenes was extracted from driving sequences of the KITTI raw data set. It was then demonstrated that, as expected, a purely camera-based FCN severely underperforms on this data set. A multimodal system, on the other hand, is still able to provide high accuracy. Finally, the proposed cross fusion FCN was evaluated on the KITTI road benchmark where it achieved excellent performance, with a MaxF score of 96.03%, ranking it among the top-performing approaches

    MFMAN-YOLO: A Method for Detecting Pole-like Obstacles in Complex Environment

    Full text link
    In real-world traffic, there are various uncertainties and complexities in road and weather conditions. To solve the problem that the feature information of pole-like obstacles in complex environments is easily lost, resulting in low detection accuracy and low real-time performance, a multi-scale hybrid attention mechanism detection algorithm is proposed in this paper. First, the optimal transport function Monge-Kantorovich (MK) is incorporated not only to solve the problem of overlapping multiple prediction frames with optimal matching but also the MK function can be regularized to prevent model over-fitting; then, the features at different scales are up-sampled separately according to the optimized efficient multi-scale feature pyramid. Finally, the extraction of multi-scale feature space channel information is enhanced in complex environments based on the hybrid attention mechanism, which suppresses the irrelevant complex environment background information and focuses the feature information of pole-like obstacles. Meanwhile, this paper conducts real road test experiments in a variety of complex environments. The experimental results show that the detection precision, recall, and average precision of the method are 94.7%, 93.1%, and 97.4%, respectively, and the detection frame rate is 400 f/s. This research method can detect pole-like obstacles in a complex road environment in real time and accurately, which further promotes innovation and progress in the field of automatic driving.Comment: 11 page

    Multi-task Learning for Visual Perception in Automated Driving

    Full text link
    Every year, 1.2 million people die, and up to 50 million people are injured in accidents worldwide. Automated driving can significantly reduce that number. Automated driving also has several economic and societal benefits that include convenient and efficient transportation, enhanced mobility for the disabled and elderly population, etc. Visual perception is the ability to perceive the environment, which is a critical component in decision-making that builds safer automated driving. Recent progress in computer vision and deep learning paired with high-quality sensors like cameras and LiDARs fueled mature visual perception solutions. The main bottleneck for these solutions is the limited processing power available to build real-time applications. This bottleneck often leads to a trade-off between performance and run-time efficiency. To address these bottlenecks, we focus on: 1) building optimized architectures for different visual perception tasks like semantic segmentation, panoptic segmentation, etc. using convolutional neural networks that have high performance and low computational complexity, 2) using multi-task learning to overcome computational bottlenecks by sharing the initial convolutional layers between different tasks while developing advanced learning strategies that achieve balanced learning between tasks.PhDCollege of Engineering & Computer ScienceUniversity of Michigan-Dearbornhttp://deepblue.lib.umich.edu/bitstream/2027.42/167355/1/Sumanth Chennupati Final Dissertation.pd

    A Systematic Review on Object Localisation Methods in Images

    Full text link
    [EN] Currently, many applications require a precise localization of the objects that appear in an image, to later process them. This is the case of visual inspection in the industry, computer-aided clinical diagnostic systems, the obstacle detection in vehicles or in robots, among others. However, several factors such as the quality of the image and the appearance of the objects to be detected make this automatic location difficult. In this article, we carry out a systematic revision of the main methods used to locate objects by considering since the methods based on sliding windows, as the detector proposed by Viola and Jones, until the current methods that use deep learning networks, such as Faster-RCNN or Mask-RCNN. For each proposal, we describe the relevant details, considering their advantages and disadvantages, as well as the main applications of these methods in various areas. This paper aims to provide a clean and condensed review of the state of the art of these techniques, their usefulness and their implementations in order to facilitate their knowledge and use by any researcher that requires locating objects in digital images. We conclude this work by summarizing the main ideas presented and discussing the future trends of these methods.[ES] Actualmente, muchas aplicaciones requieren localizar de forma precisa los objetos que aparecen en una imagen, para su posterior procesamiento. Este es el caso de la inspección visual en la industria, los sistemas de diagnóstico clínico asistido por computador, la detección de obstáculos en vehículos o en robots, entre otros. Sin embargo, diversos factores como la calidad de la imagen y la apariencia de los objetos a detectar, dificultan la localización automática. En este artículo realizamos una revisión sistemática de los principales métodos utilizados para localizar objetos, considerando desde los métodos basados en ventanas deslizantes, como el detector propuesto por Viola y Jones, hasta los métodos actuales que usan redes de aprendizaje profundo, tales como Faster-RCNNo Mask-RCNN. Para cada propuesta, describimos los detalles relevantes, considerando sus ventajas y desventajas, así como sus aplicaciones en diversas áreas. El artículo pretende proporcionar una revisión ordenada y condensada del estado del arte de estas técnicas, su utilidad y sus implementaciones a fin de facilitar su conocimiento y uso por cualquier investigador que requiera localizar objetos en imágenes digitales. Concluimos este trabajo resumiendo las ideas presentadas y discutiendo líneas de trabajo futuro.Este trabajo ha sido financiado parcialmente por diferentes instituciones. Deisy Chaves cuenta con una beca “Estudios de Doctorado en Colombia 2013” de COLCIENCIAS. Surajit Saikia cuenta con una beca de la Junta de Castilla y León con referencia EDU/529/2017. También queremos agradecer el apoyo de INCIBE (Instituto Nacional de Ciberseguridad) mediante la Adenda 22 al convenio con la Universidad de León.Chaves, D.; Saikia, S.; Fernández-Robles, L.; Alegre, E.; Trujillo, M. (2018). Una Revisión Sistemática de Métodos para Localizar Automáticamente Objetos en Imágenes. Revista Iberoamericana de Automática e Informática industrial. 15(3):231-242. https://doi.org/10.4995/riai.2018.10229OJS231242153Akselrod-Ballin, A., Karlinsky, L., Alpert, S., Hasoul, S., Ben-Ari, R., Barkan, E., 2016. A region based convolutional network for tumor detection and classification in breast mammography. In: Deep Learning and Data Labe-ling for Medical Applications. pp. 197-205.Alexe, B., Deselaers, T., Ferrari, V., 2010. What is an object? In: CVPR. pp.73-80.Ammour, N., Alhichri, H., Bazi, Y., Benjdira, B., Alajlan, N., Zuair, M., 2017.Deep learning approach for car detection in uav imagery. Remote Sens. 9 (4). DOI:10.3390/rs9040312Boser, B. E., Guyon, I. M., Vapnik, V. N., 1992. A training algorithm for opti-mal margin classifiers. In: COLT. pp. 144-152.Brazil, G., Yin, X., Liu, X., 2017. Illuminating pedestrians via simultaneous detection & segmentation. CoRR abs/1706.08564.Cai, Z., Fan, Q., Feris, R. S., Vasconcelos, N., 2016. A unified multi-scale deep convolutional neural network for fast object detection. CoRRabs/1607.07155.Cao, X., Gong, G., Liu, M.,Qi, J., 2016. Foreign object debris detection on air-field pavement using region based convolution neural network. In: DICTA. pp. 1-6. DOI:10.1109/DICTA.2016.7797045Cao, X., Wang, P., Meng, C., Bai, X., Gong, G., Liu, M., Qi, J., 2018. Region based cnn for foreign object debris detection on airfield pavement. Sensors18 (3). DOI:10.3390/s18030737Chen, J., Liu, Z., Wang, H., Núñez, A., Han, Z., 2018. Automatic defect detection of fasteners on the catenary support device using deep convolutional neural network. IEEE T Instrum Meas 67 (2), 257-269. DOI:10.1109/TIM.2017.2775345Cireʂan, D. C., Giusti, A., Gambardella, L. M., Schmidhuber, J., 2013. Mitosis detection in breast cancer histology images with deep neural networks. In: MICCAI. pp. 411-418.Coifman, B., McCord, M., Mishalani, R. G., Iswalt, M., Ji, Y., 2006. Roadway traffic monitoring from an unmanned aerial vehicle. IEE Proceedings - Intelligent Transport Systems 153 (1),11-20. DOI:10.1049/ip-its:20055014Dai, J., Li, Y., He, K., Sun, J., 2016. R-FCN: object detection via region-based fully convolutional networks. CoRR abs/1605.06409.Dalal, N., Triggs, B., June2005. Histograms of oriented gradients for human detection. In: CVPR. Vol. 1. pp. 886-893 vol. 1. DOI:10.1109/CVPR.2005.177Deng, L., 2014. A tutorial survey of architectures, algorithms, and applications for deep learning. APSIPA Transactions on Signal and Information Processing 3, e2.Deng, L., Yu, D., 2014. Deep learning: Methods and applications. Foundations and Trends in Signal Processing 7 (3-4), 197-387.Dollár, P., Tu, Z., Perona, P., Belongie, S. J., 2009. Integral channel features. In: BMVC. pp. 1-11.Dollar, P., Zitnick, L., 2013. Structured forests for fast edge detection. In: ICCV. pp. 1841-1848.Donoser, M., Bischof, H., 2006. Efficient maximally stable extremal region (mser) tracking. In: CVPR. pp. 553-560. DOI:10.1109/CVPR.2006.107Du, X., El-Khamy, M., Lee, J., Davis, L., 2017. Fused dnn: A deep neural net-work fusion approach to fast and robust pedestrian detection. In: WACV. pp.953-961. DOI:10.1109/WACV.2017.111Dženan, Z., Aleš, V., Jan, E., Daniel, H., Christopher, N., Andreas, K., 2014. Robust detection and segmentation for diagnosis of vertebral diseases using routine mr images. Computer Graphics Forum 33 (6), 190-204. DOI:10.1111/cgf.12343Felzenszwalb, P. F., Girshick, R. B., McAllester, D., Ramanan, D., 2010. Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 32 (9), 1627-1645. DOI:10.1109/TPAMI.2009.167Felzenszwalb, P. F., Huttenlocher, D. P., 2004. Efficient graph-based image segmentation. IJCV 59 (2), 167-181. DOI:10.1023/B:VISI.0000022288.19776.77Ferguson, M., Ak, R., Lee, Y. T. T., Law, K. H., 2017. Automatic localization of casting defects with convolutional neural networks. In: IEEE International Conference on Big Data. pp. 1726-1735. DOI:10.1109/BigData.2017.8258115Fernández-Robles, L., Azzopardi, G., Alegre, E., Petkov, N., 2017a. Machine-vision-based identification of broken inserts in edge profile milling heads. Robot Comput Integr Manuf 44, 276 - 283. DOI:https://doi.org/10.1016/j.rcim.2016.10.004Fernández-Robles, L., Azzopardi, G., Alegre, E., Petkov, N., Castejón-Limas ,M., 2017b. Identification of milling inserts in situ based on a versatile machine vision system. JMSY 45, 48 - 57. DOI: https://doi.org/10.1016/j.jmsy.2017.08.002Freund, Y., Schapire, R. E., 1999. A short introduction to boosting. In: IJCAI. pp. 1401-1406.García-Ordás, M. T., Alegre, E., González-Castro, V., Alaiz-Rodríguez, R.,2017. A computer vision approach to analyze and classify tool wear level in milling processes using shape descriptors and machine learning techniques. Int J Adv Manuf Technol 90 (5), 1947-1961. DOI:10.1007/s00170-016-9541-0García-Olalla, O., Alegre, E., Fernández-Robles, L., Fidalgo, E., Saikia, S., 2018. Textile retrieval based on image content from cdc and webcam cameras in indoor environments. Sensors 18 (5). DOI:10.3390/s18051329Garnett, N., Silberstein, S., Oron, S., Fetaya, E., Verner, U., Ayash, A., Goldner,V., Cohen, R., Horn, K., Levi, D., 2017. Real-time category-based and general obstacle detection for autonomous driving. In: ICCVW. pp. 198-205. DOI:10.1109/ICCVW.2017.32Girshick, R. B., 2015. Fast R-CNN. CoRR abs/1504.08083.Girshick, R. B., Donahue, J., Darrell, T., Malik, J., 2013. Rich feature hierarchies for accurate object detection and semantic segmentation. CoRRabs/1311.2524.He, B., Xiao, D., Hu, Q., Jia, F., 2018. Automatic magnetic resonance image prostate segmentation based on adaptive feature learning probability boos-ting tree initialization and cnn-asm refinement. IEEE Access 6, 2005-2015.He, K., Gkioxari, G., Doll ́ar, P., Girshick, R. B., 2017. Mask R-CNN. CoRRabs/1703.06870.He, K., Zhang, X., Ren, S., Sun, J., 2016. Deep residual learning for image recognition. In: CVPR. pp. 770-778.Heo, Y. J., Lee, D., Kang, J., Lee, K., Chung, W. K., 2017. Real-time Image Processing for Microscopy-based Label-free Imaging Flow Cytometry in a Microfluidic Chip. Scientific Reports 7 (1), 11651. DOI:10.1038/s41598-017-11534-0Hosang, J., Benenson, R., Doll ́ar, P., Schiele, B., 2016. What makes for effective detection proposals? IEEE Trans. Pattern Anal. Mach. Intell. 38 (4),814-830. DOI:10.1109/TPAMI.2015.2465908Jiamin, L., David, W., Le, L., Zhuoshi, W., Lauren, K., B., T. E., Berkman,S., A., P. N., M., S. R., 2017. Detection and diagnosis of colitis on computed tomography using deep convolutional neural networks. Medical Physics44 (9), 4630-4642. DOI:10.1002/mp.12399Jung, F., Kirschner, M., Wesarg, S., 2013. A generic approach to organ detection using 3d haar-like features. In: Bildverarbeitung für die Medizin 2013.pp. 320-325.Kisilev, P., Sason, E., Barkan, E., Hashoul, S., 2016. Medical image description nusing multi-task-loss cnn. In: Deep Learning and Data Labeling for Medical Applications. pp. 121-129.Krizhevsky, A., Sutskever, I., Hinton, G. E., 2012. Imagenet classification with deep convolutional neural networks. In: Adv Neural Inf Process Syst. pp. 1097-1105.Lampert, C. H., Blaschko, M. B., Hofmann, T., 2008. Beyond sliding windows: Object localization by efficient subwindow search. In: CVPR. pp. 1-8. DOI:10.1109/CVPR.2008.4587586Lecun, Y., Bengio, Y., Hinton, G., 2015. Deep learning. Nature 521, 436-444.Lee, C. J., Tseng, T. H., Huang, B. J., Jun-Weihsieh, Tsai, C. M., 2015. Obstacle detection and avoidance via cascade classifier for wheeled mobile robot. In: ICMLC. Vol. 1. pp. 403-407. DOI:10.1109/ICMLC.2015.7340955Lee, J., Wang, J., Crandall, D., Šabanovic, S., Fox, G., 2017. Real-time, cloud-based object detection for unmanned aerial vehicles. In: IRC. pp. 36-43. DOI:10.1109/IRC.2017.77Levi, D., Garnett, N., Fetaya, E., September 2015a. Stixelnet: A deep convolutional network for obstacle detection and road segmentation. In: BMVC. pp. 109.1-109.12. DOI:10.5244/C.29.109Levi, D., Garnett, N., Fetaya, E., 2015b. Stixelnet: A deep convolutional network for obstacle detection and road segmentation. In: BMVC. pp. 109.1-109.12. DOI:10.5244/C.29.109Li, J., Liang, X., Shen, S., Xu, T., Feng, J., Yan, S., 2018. Scale-aware fast r-cnn for pedestrian detection. IEEE Trans Multimedia 20 (4), 985-996. DOI:10.1109/TMM.2017.2759508Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., Berg, A. C.,2016. Ssd: Single shot multibox detector. In: ECCV. pp. 21-37.Luo, S., Lu, H., Xiao, J., Yu, Q., Zheng, Z., 2017. Robot detection and localization based on deep learning. In: CAC. pp. 7091-7095.Ma, Y., Jiang, Z., Zhang, H., Xie, F., Zheng, Y., Shi, H., 2017. Proposing regions from histopathological whole slide image for retrieval using selective search. In: ISBI. pp. 156-159. DOI:10.1109/ISBI.2017.7950491Mery, D., Rio, V., Zscherpel, U., Mondrag ́on, G., Lillo, I., Zuccar, I., Lobel,H., Carrasco, M., 2015. Gdxray: The database of x-ray images for nondestructive testing. Journal of Nondestructive Evaluation 34 (4), 42. DOI:10.1007/s10921-015-0315-7Park, J.-K., Kwon, B.-K., Park, J.-H., Kang, D.-J., 2016. Machine learning-based imaging system for surface defect inspection. IJPEM-GT 3 (3), 303-310. DOI:10.1007/s40684-016-0039-xRedmon, J., Divvala, S. K., Girshick, R. B., Farhadi, A., 2015. You only look once: Unified, real-time object detection. CoRR abs/1506.02640.Ren, S., He, K., Girshick, R. B., Sun, J., 2015. Faster R-CNN: towards real-time object detection with region proposal networks. CoRR abs/1506.01497.Říha, K., Mašek, J., Burget, R., Beneš, R., Závodná, E., 2013. Novel method for localization of common carotid artery transverse section in ultrasound images using modified viola-jones detector. Ultrasound Med Biol 39 (10),1887 - 1902. DOI:10.1016/j.ultrasmedbio.2013.04.013Sa, R., Owens, W., Wiegand, R., Studin, M., Capoferri, D., Barooha, K.,Greaux, A., Rattray, R., Hutton, A., Cintineo, J., Chaudhary, V., 2017. Intervertebral disc detection in x-ray images using faster r-cnn. In: EMBC. pp. 564-567. DOI:10.1109/EMBC.2017.8036887Saikia, S., Fidalgo, E., Alegre, E., Fernández-Robles, L., 2017. Object detection for crime scene evidence analysis using deep learning. In: ICIAP. pp.14-24.Sepúlveda, G. V., Torriti, M. T.,Calero, M. F., 2017. Sistema de detección de señales de tráfico para la localización de intersecciones viales y frenado anticipado. Revista Iberoamericana de Automática e Informática Industrial14 (2), 152-162. DOI:10.1016/j.riai.2016.09.010Shah, V. R., Maru, S. V., Jhaveri, R. H., 2018. An obstacle detection scheme for vehicles in an intelligent transportation system. IJCNIS 8 (10), 23-28. DOI:10.5815/ijcnis.2016.10.03Shi, Y., Li, Y., Wei, X., Zhou, Y., 2017. A faster-rcnn based chemical fiber paper tube defect detection method. In: International Conference on Enterprise Systems. pp. 173-177. DOI:10.1109/ES.2017.35Simonyan, K., Zisserman, A., 2014. Very deep convolutional networks for large-scale image recognition. CoRR abs/1409.1556.Szegedy, C., Ioe, S., Vanhoucke, V., Alemi, A. A., 2017. Inception-v4, inception-resnet and the impact of residual connections on learning. In: AAAI. pp. 4278-4284.Tang, T., Zhou, S., Deng, Z., Zou, H., Lei, L., 2017. Vehicle detection in aerial images based on region convolutional neural networks and hard negative example mining. Sensors 17 (2). DOI:10.3390/s17020336Tek, F., 2013. Mitosis detection using generic features and an ensemble of cascade adaboosts. J Pathol Inform 4 (1), 12. DOI:10.4103/2153-3539.112697Uijlings, J. R. R., van de Sande, K. E. A., Gevers, T., Smeulders, A. W. M. ,2013. Selective search for object recognition. IJCV 104 (2), 154-171.Viola, P., Jones, M. J., May 2004. Robust real-time face detection. IJCV 57 (2), 137-154 .DOI:10.1023/B:VISI.0000013087.49260.fbWang, S., Cheng, J., Liu, H., Tang, M., 2018. Pcn: Part and context information for pedestrian detection with cnns. CoRR abs/1804.04483.Xu, Y., Yu, G., Wang, Y., Ma, Y., 2017a. Car detection from low-altitude uav imagery with the faster r-cnn. JAT 2017. DOI:https://doi.org/10.1155/2017/2823617Xu, Y., Yu, G., Wang, Y., Wu, X., Ma, Y., 2016. A hybrid vehicle detection method based on viola-jones and hog+svm from uav images. Sensors 16 (8). DOI:10.3390/s16081325Xu, Y., Yu, G., Wu, X., Wang, Y., Ma, Y., 2017b. An enhanced viola-jones vehicle detection method from unmanned aerial vehicles imagery. IEEE trans Intell Transp Syst 18 (7), 1845-1856. DOI:10.1109/TITS.2016.2617202Yang, S., Fang, B., Tang, W., Wu, X., Qian, J., Yang, W., 2017. Faster r-cnn based microscopic cell detection. In: SPAC. pp. 345-350. DOI:10.1109/SPAC.2017.8304302Yi, X., Song, G., Derong, T., Dong, G., Liang, S., Yuqiong, W., 2018. Fast road obstacle detection method based on maximally stable extremal regions. IJARS 15 (1), 1-10. DOI:10.1177/1729881418759118Zeiler, M. D., Fergus, R., 2014. Visualizing and understanding convolutional networks. In: ECCV. pp. 818-833.Zhang, L., Lin, L., Liang, X., He, K., 2016. Is faster r-cnn doing well for pedestrian detection? In: ECCV. pp. 443-457.Zhong, J., Lei, T., Yao, G., 2017. Robust vehicle detection in aerial images based on cascaded convolutional neural networks. Sensors 17 (12). DOI:10.3390/s17122720Zitnick, L., Dollar, P., 2014. Edge boxes: Locating object proposals from edges. In: ECCV. pp. 391-405
    corecore