19 research outputs found

    Traffic and road sign recognition

    Get PDF
    This thesis presents a system to recognise and classify road and traffic signs for the purpose of developing an inventory of them which could assist the highway engineers' tasks of updating and maintaining them. It uses images taken by a camera from a moving vehicle. The system is based on three major stages: colour segmentation, recognition, and classification.Four colour segmentation algorithms are developed and tested. They are a shadow and highlight invariant, a dynamic threshold, a modification of de la Escalera's algorithm and a Fuzzy colour segmentation algorithm. All algorithms are tested using hundreds of images and the shadow-highlight invariant algorithm is eventually chosen as the best performer. This is because it is immune to shadows and highlights. It is also robust as it was tested in different lighting conditions, weather conditions, and times of the day.Approximately 97% successful segmentation rate was achieved using this algorithm. Recognition of traffic signs is carried out using a fuzzy shape recogniser. Based on four shape measures - the rectangularity, triangularity, ellipticity, and octagonality, fuzzy rules were developed to determine the shape of the sign. Among these shape measures octangonality has been introduced in this research. The final decision of the recogniser is based on the combination of both the colour and shape of the sign. The recogniser was tested in a variety of testing conditions giving an overall performance of approximately 88%.Classification was undertaken using a Support Vector Machine (SVM) classifier. The classification is carried out in two stages: rim's shape classification followed by the classification of interior of the sign. The classifier was trained and tested using binary images in addition to five different types of moments which are Geometric moments, Zernike moments, Legendre moments, Orthogonal Fourier-Mellin Moments, and Binary Haar features. The performance of the SVM was tested using different features, kernels, SVM types, SVM parameters, and moment's orders. The average classification rate achieved is about 97%. Binary images show the best testing results followed by Legendre moments. Linear kernel gives the best testing results followed by RBF. C-SVM shows very good performance, but v-SVM gives better results in some case

    On the Recognition of Emotion from Physiological Data

    Get PDF
    This work encompasses several objectives, but is primarily concerned with an experiment where 33 participants were shown 32 slides in order to create ‗weakly induced emotions‘. Recordings of the participants‘ physiological state were taken as well as a self report of their emotional state. We then used an assortment of classifiers to predict emotional state from the recorded physiological signals, a process known as Physiological Pattern Recognition (PPR). We investigated techniques for recording, processing and extracting features from six different physiological signals: Electrocardiogram (ECG), Blood Volume Pulse (BVP), Galvanic Skin Response (GSR), Electromyography (EMG), for the corrugator muscle, skin temperature for the finger and respiratory rate. Improvements to the state of PPR emotion detection were made by allowing for 9 different weakly induced emotional states to be detected at nearly 65% accuracy. This is an improvement in the number of states readily detectable. The work presents many investigations into numerical feature extraction from physiological signals and has a chapter dedicated to collating and trialing facial electromyography techniques. There is also a hardware device we created to collect participant self reported emotional states which showed several improvements to experimental procedure

    Perception and Acquisition of Natural Authentic English Speech for Chinese Learners Using DIT\u27s Speech Technologies

    Get PDF
    Given that Chinese language learners are greatly influenced by their mother-tongue, which is a tone language rather than an intonation language, learning and coping with authentic English speech seems more difficult than for learners of other languages. The focus of the current research is, on the basis of analysis of the nature of spoken English and spoken Chinese, to help Chinese learners derive benefit from ICT technologies developed by the Technological University Dublin (DIT). The thesis concentrates on investigating the application of speech technologies in bridging the gap between students’ internalised, idealised formulations and natural, authentic English speech. Part of the testing carried out by the present author demonstrates the acceptability of a slow-down algorithm in facilitating Chinese learners of English in re-producing formulaic language. This algorithm is useful because it can slow down audio files to any desired speed between 100% and 40% without distortion, so as to allow language learners to pay attention to the real, rapid flow of ‘messy’ speech and follow the intonation patterns contained in them. The rationale for and the application of natural, dialogic native-to-native English speech to language learning is also explored. The Chinese language learners involved in this study are exposed to authentic, native speech patterns by providing them access to real, informal dialogue in various contexts. In the course of this analysis, the influence of speed of delivery and pitch range on the categorisation of formulaic language is also investigated. The study investigates the potential of the speech tools available to the present author as an effective EFL learning facility, especially for speakers of tone languages, and their role in helping language learners achieve confluent interaction in an English L1 environment

    Geospatial Computing: Architectures and Algorithms for Mapping Applications

    Get PDF
    Beginning with the MapTube website (1), which was launched in 2007 for crowd-sourcing maps, this project investigates approaches to exploratory Geographic Information Systems (GIS) using web-based mapping, or ‘web GIS’. Users can log in to upload their own maps and overlay different layers of GIS data sets. This work looks into the theory behind how web-based mapping systems function and whether their performance can be modelled and predicted. One of the important questions when dealing with different geospatial data sets is how they relate to one another. Internet data stores provide another source of information, which can be exploited if more generic geospatial data mining techniques are developed. The identification of similarities between thousands of maps is a GIS technique that can give structure to the overall fabric of the data, once the problems of scalability and comparisons between different geographies are solved. After running MapTube for nine years to crowd-source data, this would mark a natural progression from visualisation of individual maps to wider questions about what additional knowledge can be discovered from the data collected. In the new ‘data science’ age, the introduction of real-time data sets introduces a new challenge for web-based mapping applications. The mapping of real-time geospatial systems is technically challenging, but has the potential to show inter-dependencies as they emerge in the time series. Combined geospatial and temporal data mining of realtime sources can provide archives of transport and environmental data from which to accurately model the systems under investigation. By using techniques from machine learning, the models can be built directly from the real-time data stream. These models can then be used for analysis and experimentation, being derived directly from city data. This then leads to an analysis of the behaviours of the interacting systems. (1) The MapTube website: http://www.maptube.org

    Perception and recognition of computer-enhanced facial attributes and abstracted prototypes

    Get PDF
    The influence of the human facial image was surveyed and the nature of its many interpretations were examined. The role of distinctiveness was considered particularly relevant as it accounted for many of the impressions of character and identity ascribed to individuals. The notion of structural differences with respect to some selective essence of normality is especially important as it allows a wide range of complex facial types to be considered and understood in an objective manner. A software tool was developed which permitted the manipulation of facial images. Quantitative distortions of digital images were examined using perceptual and recognition memory paradigms. Seven experiments investigated the role of distinctiveness in memory for faces using synthesised caricatures. The results showed that caricatures, both photographic and line-drawing, improved recognition speed and accuracy, indicating that both veridical and distinctiveness information are coded for familiar faces in long-term memory. The impact of feature metrics on perceptual estimates of facial age was examined using 'age-caricatured' images and were found to be in relative accordance with the 'intended' computed age. Further modifying the semantics permitted the differences between individual faces to be visualised in terms of facial structure and skin texture patterns. Transformations of identity between two, or more, faces established the necessary matrices which can offer an understanding of facial expression in a categorical manner and the inherent interactions. A procedural extension allowed generation of composite images in which all features are perfectly aligned. Prototypical facial types specified in this manner enabled high-level manipulations to be made of gender and attractiveness; two experiments corroborated previously speculative material and thus gave credence to the prototype model. In summary, psychological assessment of computer-manipulated facial images demonstrated the validity of the objective techniques and highlighted particular parameters which contribute to our perception and recognition of the individual and of underlying facial types

    Noise and morphogenesis: Uncertainty, randomness and control

    Get PDF
    This thesis presents a processual ontology of noise by virtue of which morphogenesis (in its most general understanding as the processes by which order/form is created) must be instantiated. Noise is here outlined as the far from equilibrium environment out of which metastable temporary ‘solutions’ can emerge as the system transitions through the pre-individual state space. While frequently addressed by humanities and arts studies on the basis of its supposed disruptive character (often in terms of aesthetics), this thesis aims to thoroughly examine noise’s conceptual potencies. To explore and amplify the epistemic consequences not merely of the ineliminability of noise but of its originative power as well as within the course of the elimination of givenness by epistemology. This philosophical work is informed by many different fields of contemporary science (namely: statistical physics, information theory, probability theory, 4E cognition, synthetic biology, nonlinear dynamics, complexity science and computer science) in order to assess and highlight the problems of the metascientific and ideological foundations of diverse projects of prediction and control of uncertainty. From algorithmic surveillance back to cybernetics and how these rendered noise “informationally heretical”. This conveys an analysis of how contemporary prediction technologies are dramatically transforming our relationship with the future and with uncertainty in a great number of our social structures. It is a philosophico-critical anthropology of data ontology and a critique of reductive pan-info-computationalism. Additionally, two practical examples of noise characterised as an enabling constraint for the functioning of complex adaptive systems are presented. These are at once biophysical and cognitive, : 1) interaction-dominance constituted by ‘pink noise’ and 2) noise as a source of variability that cells may exploit in (synthetic) biology. Finally, noise is posited as an intractable active ontological randomness that limits the scope of determinism and that goes beyond unpredictability in any epistemological sense due to the insuperability of the situation in which epistemology finds itself following the critique of the given

    Information and Communication Technologies in Tourism 2022

    Get PDF
    This open access book presents the proceedings of the International Federation for IT and Travel & Tourism (IFITT)’s 29th Annual International eTourism Conference, which assembles the latest research presented at the ENTER2022 conference, which will be held on January 11–14, 2022. The book provides an extensive overview of how information and communication technologies can be used to develop tourism and hospitality. It covers the latest research on various topics within the field, including augmented and virtual reality, website development, social media use, e-learning, big data, analytics, and recommendation systems. The readers will gain insights and ideas on how information and communication technologies can be used in tourism and hospitality. Academics working in the eTourism field, as well as students and practitioners, will find up-to-date information on the status of research
    corecore