28 research outputs found

    Smart environment monitoring through micro unmanned aerial vehicles

    Get PDF
    In recent years, the improvements of small-scale Unmanned Aerial Vehicles (UAVs) in terms of flight time, automatic control, and remote transmission are promoting the development of a wide range of practical applications. In aerial video surveillance, the monitoring of broad areas still has many challenges due to the achievement of different tasks in real-time, including mosaicking, change detection, and object detection. In this thesis work, a small-scale UAV based vision system to maintain regular surveillance over target areas is proposed. The system works in two modes. The first mode allows to monitor an area of interest by performing several flights. During the first flight, it creates an incremental geo-referenced mosaic of an area of interest and classifies all the known elements (e.g., persons) found on the ground by an improved Faster R-CNN architecture previously trained. In subsequent reconnaissance flights, the system searches for any changes (e.g., disappearance of persons) that may occur in the mosaic by a histogram equalization and RGB-Local Binary Pattern (RGB-LBP) based algorithm. If present, the mosaic is updated. The second mode, allows to perform a real-time classification by using, again, our improved Faster R-CNN model, useful for time-critical operations. Thanks to different design features, the system works in real-time and performs mosaicking and change detection tasks at low-altitude, thus allowing the classification even of small objects. The proposed system was tested by using the whole set of challenging video sequences contained in the UAV Mosaicking and Change Detection (UMCD) dataset and other public datasets. The evaluation of the system by well-known performance metrics has shown remarkable results in terms of mosaic creation and updating, as well as in terms of change detection and object detection

    Sensing and Signal Processing in Smart Healthcare

    Get PDF
    In the last decade, we have witnessed the rapid development of electronic technologies that are transforming our daily lives. Such technologies are often integrated with various sensors that facilitate the collection of human motion and physiological data and are equipped with wireless communication modules such as Bluetooth, radio frequency identification, and near-field communication. In smart healthcare applications, designing ergonomic and intuitive human–computer interfaces is crucial because a system that is not easy to use will create a huge obstacle to adoption and may significantly reduce the efficacy of the solution. Signal and data processing is another important consideration in smart healthcare applications because it must ensure high accuracy with a high level of confidence in order for the applications to be useful for clinicians in making diagnosis and treatment decisions. This Special Issue is a collection of 10 articles selected from a total of 26 contributions. These contributions span the areas of signal processing and smart healthcare systems mostly contributed by authors from Europe, including Italy, Spain, France, Portugal, Romania, Sweden, and Netherlands. Authors from China, Korea, Taiwan, Indonesia, and Ecuador are also included

    Face Liveness Detection under Processed Image Attacks

    Get PDF
    Face recognition is a mature and reliable technology for identifying people. Due to high-definition cameras and supporting devices, it is considered the fastest and the least intrusive biometric recognition modality. Nevertheless, effective spoofing attempts on face recognition systems were found to be possible. As a result, various anti-spoofing algorithms were developed to counteract these attacks. They are commonly referred in the literature a liveness detection tests. In this research we highlight the effectiveness of some simple, direct spoofing attacks, and test one of the current robust liveness detection algorithms, i.e. the logistic regression based face liveness detection from a single image, proposed by the Tan et al. in 2010, against malicious attacks using processed imposter images. In particular, we study experimentally the effect of common image processing operations such as sharpening and smoothing, as well as corruption with salt and pepper noise, on the face liveness detection algorithm, and we find that it is especially vulnerable against spoofing attempts using processed imposter images. We design and present a new facial database, the Durham Face Database, which is the first, to the best of our knowledge, to have client, imposter as well as processed imposter images. Finally, we evaluate our claim on the effectiveness of proposed imposter image attacks using transfer learning on Convolutional Neural Networks. We verify that such attacks are more difficult to detect even when using high-end, expensive machine learning techniques

    Piloted aircraft simulation concepts and overview

    Get PDF
    An overview of piloted aircraft simulation is presented that reflects the viewpoint of an aeronautical technologist. The intent is to acquaint potential users with some of the basic concepts and issues that characterize piloted simulation. Application to the development of aircraft are highlighted, but some aspects of training simulators are covered. A historical review is given together with a description of some current simulators. Simulator usages, advantages, and limitations are discussed and human perception qualities important to simulation are related. An assessment of current simulation is presented that addresses validity, fidelity, and deficiencies. Future prospects are discussed and technology projections are made

    3D Gaze Estimation from Remote RGB-D Sensors

    Get PDF
    The development of systems able to retrieve and characterise the state of humans is important for many applications and fields of study. In particular, as a display of attention and interest, gaze is a fundamental cue in understanding people activities, behaviors, intentions, state of mind and personality. Moreover, gaze plays a major role in the communication process, like for showing attention to the speaker, indicating who is addressed or averting gaze to keep the floor. Therefore, many applications within the fields of human-human, human-robot and human-computer interaction could benefit from gaze sensing. However, despite significant advances during more than three decades of research, current gaze estimation technologies can not address the conditions often required within these fields, such as remote sensing, unconstrained user movements and minimum user calibration. Furthermore, to reduce cost, it is preferable to rely on consumer sensors, but this usually leads to low resolution and low contrast images that current techniques can hardly cope with. In this thesis we investigate the problem of automatic gaze estimation under head pose variations, low resolution sensing and different levels of user calibration, including the uncalibrated case. We propose to build a non-intrusive gaze estimation system based on remote consumer RGB-D sensors. In this context, we propose algorithmic solutions which overcome many of the limitations of previous systems. We thus address the main aspects of this problem: 3D head pose tracking, 3D gaze estimation, and gaze based application modeling. First, we develop an accurate model-based 3D head pose tracking system which adapts to the participant without requiring explicit actions. Second, to achieve a head pose invariant gaze estimation, we propose a method to correct the eye image appearance variations due to head pose. We then investigate on two different methodologies to infer the 3D gaze direction. The first one builds upon machine learning regression techniques. In this context, we propose strategies to improve their generalization, in particular, to handle different people. The second methodology is a new paradigm we propose and call geometric generative gaze estimation. This novel approach combines the benefits of geometric eye modeling (normally restricted to high resolution images due to the difficulty of feature extraction) with a stochastic segmentation process (adapted to low-resolution) within a Bayesian model allowing the decoupling of user specific geometry and session specific appearance parameters, along with the introduction of priors, which are appropriate for adaptation relying on small amounts of data. The aforementioned gaze estimation methods are validated through extensive experiments in a comprehensive database which we collected and made publicly available. Finally, we study the problem of automatic gaze coding in natural dyadic and group human interactions. The system builds upon the thesis contributions to handle unconstrained head movements and the lack of user calibration. It further exploits the 3D tracking of participants and their gaze to conduct a 3D geometric analysis within a multi-camera setup. Experiments on real and natural interactions demonstrate the system is highly accuracy. Overall, the methods developed in this dissertation are suitable for many applications, involving large diversity in terms of setup configuration, user calibration and mobility

    Integrated Condition Assessment of Subway Networks Using Computer Vision and Nondestructive Evaluation Techniques

    Get PDF
    Subway networks play a key role in the smart mobility of millions of commuters in major metropolises. The facilities of these networks constantly deteriorate, which may compromise the integrity and durability of concrete structures. The ASCE 2017 Report Card revealed that the condition of public transit infrastructure in the U.S. is rated D-; hence a rehabilitation backlog of $90 billion is estimated to improve transit status to good conditions. Moreover, the Canadian Urban Transit Association (CUTA) reported 56.6 billion CAD in infrastructure needs for the period 2014-2018. The inspection and assessment of metro structures are predominantly conducted on the basis of Visual Inspection (VI) techniques, which are known to be time-consuming, costly, and qualitative in nature. The ultimate goal of this research is to develop an integrated condition assessment model for subway networks based on image processing, Artificial Intelligence (AI), and Non-Destructive Evaluation (NDE) techniques. Multiple image processing algorithms are created to enhance the crucial clues associated with RGB images and detect surface distresses. A complementary scheme is structured to channel the resulted information to Artificial Neural Networks (ANNs) and Regression Analysis (RA) techniques. The ANN model comprises sequential processors that automatically detect and quantify moisture marks (MM) defects. The RA model predicts spalling/scaling depth and simulates the de-facto scene by developing a hybrid algorithm and interactive 3D presentation. In addition, a comparative analysis is performed to select the most appropriate NDE technique for subway inspection. This technique is applied to probe the structure and measure the subsurface defects. Also, a novel model for the detection of air voids and water voids is proposed. The Fuzzy Inference System (FIS), Adaptive Neuro-Fuzzy Inference System (ANFIS), and Monte Carlo Simulation (MCS) are streamlined through successive operations to create the integrated condition assessment model. To exemplify and validate the proposed methodology, a myriad of images and profiles are collected from Montréal Metro systems. The results ascertain the efficacy of the developed detection algorithms. The attained recall, precision, and accuracy for MM detection algorithm are 93.2%, 96.1%, and 91.5% respectively. Whereas for spalling detection algorithm, are 91.7%, 94.8%, and 89.3% respectively. The mean and standard deviation of error percentage in MM region extraction are 12.2% and 7.9% respectively. While for spalling region extraction, they account for 11% and 7.1% respectively. Subsequent to selecting the Ground Penetrating Radar (GPR) for subway inspection, attenuation maps are generated by both the amplitude analysis and image-based analysis. Thus, the deteriorated zones and corrosiveness indices for subway elements are automatically computed. The ANN and RA models are validated versus statistical tests and key performance metrics that indicated the average validity of 96% and 93% respectively. The air/water voids model is validated through coring samples, camera images, infrared thermography and 3D laser scanning techniques. The validation outcomes reflected a strong correlation between the different results. A sensitivity analysis is conducted showing the influence of the studied subway elements on the overall subway condition. The element condition index using neuro-fuzzy technique indicated different conditions in Montréal subway systems, ranging from sound concrete to very poor, represented by 74.8 and 35.1 respectively. The fuzzy consolidator extrapolated the subway condition index of 61.6, which reveals a fair condition for Montréal Metro network. This research developed an automated tool, expected to improve the quality of decision making, as it can assist transportation agencies in identifying critical deficiencies, and by focusing constrained funding on most deserving assets

    A first step toward cognitive remediation of voices: a case study.

    Get PDF
    Several studies have shown that source-monitoring errors are related to verbal hallucinations in schizophrenia. An exploratory pilot study has been carried out to investigate the possibility of training patients in how to avoid errors in source-monitoring. One patient with paranoid schizophrenia and persistent thought insertions was trained for 6 hours to use mnemonic techniques to compensate specific deficits in source-monitoring. Results show that the patient was able to improve his performance and maintain the acquired progress at a 1-month follow-up assessment. These preliminary results are interesting for developing a larger controlled study of cognitive remediation of source-monitoring deficits

    Journal of Telecommunications and Information Technology, 2010, nr 4

    Get PDF
    kwartalni

    Internet of Underwater Things and Big Marine Data Analytics -- A Comprehensive Survey

    Full text link
    The Internet of Underwater Things (IoUT) is an emerging communication ecosystem developed for connecting underwater objects in maritime and underwater environments. The IoUT technology is intricately linked with intelligent boats and ships, smart shores and oceans, automatic marine transportations, positioning and navigation, underwater exploration, disaster prediction and prevention, as well as with intelligent monitoring and security. The IoUT has an influence at various scales ranging from a small scientific observatory, to a midsized harbor, and to covering global oceanic trade. The network architecture of IoUT is intrinsically heterogeneous and should be sufficiently resilient to operate in harsh environments. This creates major challenges in terms of underwater communications, whilst relying on limited energy resources. Additionally, the volume, velocity, and variety of data produced by sensors, hydrophones, and cameras in IoUT is enormous, giving rise to the concept of Big Marine Data (BMD), which has its own processing challenges. Hence, conventional data processing techniques will falter, and bespoke Machine Learning (ML) solutions have to be employed for automatically learning the specific BMD behavior and features facilitating knowledge extraction and decision support. The motivation of this paper is to comprehensively survey the IoUT, BMD, and their synthesis. It also aims for exploring the nexus of BMD with ML. We set out from underwater data collection and then discuss the family of IoUT data communication techniques with an emphasis on the state-of-the-art research challenges. We then review the suite of ML solutions suitable for BMD handling and analytics. We treat the subject deductively from an educational perspective, critically appraising the material surveyed.Comment: 54 pages, 11 figures, 19 tables, IEEE Communications Surveys & Tutorials, peer-reviewed academic journa
    corecore