19 research outputs found
Recommended from our members
Secure digital documents using Steganography and QR Code
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University LondonWith the increasing use of the Internet several problems have arisen regarding the processing of electronic documents. These include content filtering, content retrieval/search. Moreover, document security has taken a centre stage including copyright protection, broadcast monitoring etc. There is an acute need of an effective tool which can find the identity, location and the time when the document was created so that it can be determined whether or not the contents of the document were tampered with after creation. Owing the sensitivity of the large amounts of data which is processed on a daily basis, verifying the authenticity and integrity of a document is more important now than it ever was. Unsurprisingly document authenticity verification has become the centre of attention in the world of research. Consequently, this research is concerned with creating a tool which deals with the above problem. This research proposes the use of a Quick Response Code as a message carrier for Text Key-print. The Text Key-print is a novel method which employs the basic element of the language (i.e. Characters of the alphabet) in order to achieve authenticity of electronic documents through the transformation of its physical structure into a logical structured relationship. The resultant dimensional matrix is then converted into a binary stream and encapsulated with a serial number or URL inside a Quick response Code (QR code) to form a digital fingerprint mark. For hiding a QR code, two image steganography techniques were developed based upon the spatial and the transform domains. In the spatial domain, three methods were proposed and implemented based on the least significant bit insertion technique and the use of pseudorandom number generator to scatter the message into a set of arbitrary pixels. These methods utilise the three colour channels in the images based on the RGB model based in order to embed one, two or three bits per the eight bit channel which results in three different hiding capacities. The second technique is an adaptive approach in transforming domain where a threshold value is calculated under a predefined location for embedding in order to identify the embedding strength of the embedding technique. The quality of the generated stego images was evaluated using both objective (PSNR) and Subjective (DSCQS) methods to ensure the reliability of our proposed methods. The experimental results revealed that PSNR is not a strong indicator of the perceived stego image quality, but not a bad interpreter also of the actual quality of stego images. Since the visual difference between the cover and the stego image must be absolutely imperceptible to the human visual system, it was logically convenient to ask human observers with different qualifications and experience in the field of image processing to evaluate the perceived quality of the cover and the stego image. Thus, the subjective responses were analysed using statistical measurements to describe the distribution of the scores given by the assessors. Thus, the proposed scheme presents an alternative approach to protect digital documents rather than the traditional techniques of digital signature and watermarking
Traffic and road sign recognition
This thesis presents a system to recognise and classify road and traffic signs for the purpose of developing an inventory of them which could assist the highway engineers' tasks of updating and maintaining them. It uses images taken by a camera from a moving vehicle. The system is based on three major stages: colour segmentation, recognition, and classification.Four colour segmentation algorithms are developed and tested. They are a shadow and highlight invariant, a dynamic threshold, a modification of de la Escalera's algorithm and a Fuzzy colour segmentation algorithm. All algorithms are tested using hundreds of images and the shadow-highlight invariant algorithm is eventually chosen as the best performer. This is because it is immune to shadows and highlights. It is also robust as it was tested in different lighting conditions, weather conditions, and times of the day.Approximately 97% successful segmentation rate was achieved using this algorithm. Recognition of traffic signs is carried out using a fuzzy shape recogniser. Based on four shape measures - the rectangularity, triangularity, ellipticity, and octagonality, fuzzy rules were developed to determine the shape of the sign. Among these shape measures octangonality has been introduced in this research. The final decision of the recogniser is based on the combination of both the colour and shape of the sign. The recogniser was tested in a variety of testing conditions giving an overall performance of approximately 88%.Classification was undertaken using a Support Vector Machine (SVM) classifier. The classification is carried out in two stages: rim's shape classification followed by the classification of interior of the sign. The classifier was trained and tested using binary images in addition to five different types of moments which are Geometric moments, Zernike moments, Legendre moments, Orthogonal Fourier-Mellin Moments, and Binary Haar features. The performance of the SVM was tested using different features, kernels, SVM types, SVM parameters, and moment's orders. The average classification rate achieved is about 97%. Binary images show the best testing results followed by Legendre moments. Linear kernel gives the best testing results followed by RBF. C-SVM shows very good performance, but v-SVM gives better results in some case
On the Recognition of Emotion from Physiological Data
This work encompasses several objectives, but is primarily concerned with an experiment where 33 participants were shown 32 slides in order to create âweakly induced emotionsâ. Recordings of the participantsâ physiological state were taken as well as a self report of their emotional state. We then used an assortment of classifiers to predict emotional state from the recorded physiological signals, a process known as Physiological Pattern Recognition (PPR). We investigated techniques for recording, processing and extracting features from six different physiological signals: Electrocardiogram (ECG), Blood Volume Pulse (BVP), Galvanic Skin Response (GSR), Electromyography (EMG), for the corrugator muscle, skin temperature for the finger and respiratory rate. Improvements to the state of PPR emotion detection were made by allowing for 9 different weakly induced emotional states to be detected at nearly 65% accuracy. This is an improvement in the number of states readily detectable. The work presents many investigations into numerical feature extraction from physiological signals and has a chapter dedicated to collating and trialing facial electromyography techniques. There is also a hardware device we created to collect participant self reported emotional states which showed several improvements to experimental procedure
Perception and Acquisition of Natural Authentic English Speech for Chinese Learners Using DIT\u27s Speech Technologies
Given that Chinese language learners are greatly influenced by their mother-tongue, which is a tone language rather than an intonation language, learning and coping with authentic English speech seems more difficult than for learners of other languages. The focus of the current research is, on the basis of analysis of the nature of spoken English and spoken Chinese, to help Chinese learners derive benefit from ICT technologies developed by the Technological University Dublin (DIT). The thesis concentrates on investigating the application of speech technologies in bridging the gap between studentsâ internalised, idealised formulations and natural, authentic English speech. Part of the testing carried out by the present author demonstrates the acceptability of a slow-down algorithm in facilitating Chinese learners of English in re-producing formulaic language. This algorithm is useful because it can slow down audio files to any desired speed between 100% and 40% without distortion, so as to allow language learners to pay attention to the real, rapid flow of âmessyâ speech and follow the intonation patterns contained in them. The rationale for and the application of natural, dialogic native-to-native English speech to language learning is also explored. The Chinese language learners involved in this study are exposed to authentic, native speech patterns by providing them access to real, informal dialogue in various contexts. In the course of this analysis, the influence of speed of delivery and pitch range on the categorisation of formulaic language is also investigated. The study investigates the potential of the speech tools available to the present author as an effective EFL learning facility, especially for speakers of tone languages, and their role in helping language learners achieve confluent interaction in an English L1 environment
Geospatial Computing: Architectures and Algorithms for Mapping Applications
Beginning with the MapTube website (1), which was launched in 2007 for crowd-sourcing maps, this project investigates approaches to exploratory Geographic Information Systems (GIS) using web-based mapping, or âweb GISâ. Users can log in to upload their own maps and overlay different layers of GIS data sets. This work looks into the theory behind how web-based mapping systems function and whether their performance can be modelled and predicted. One of the important questions when dealing with different geospatial data sets is how they relate to one another. Internet data stores provide another source of information, which can be exploited if more generic geospatial data mining techniques are developed. The identification of similarities between thousands of maps is a GIS technique that can give structure to the overall fabric of the data, once the problems of scalability and comparisons between different geographies are solved. After running MapTube for nine years to crowd-source data, this would mark a natural progression from visualisation of individual maps to wider questions about what additional knowledge can be discovered from the data collected. In the new âdata scienceâ age, the introduction of real-time data sets introduces a new challenge for web-based mapping applications. The mapping of real-time geospatial systems is technically challenging, but has the potential to show inter-dependencies as they emerge in the time series. Combined geospatial and temporal data mining of realtime sources can provide archives of transport and environmental data from which to accurately model the systems under investigation. By using techniques from machine learning, the models can be built directly from the real-time data stream. These models can then be used for analysis and experimentation, being derived directly from city data. This then leads to an analysis of the behaviours of the interacting systems. (1) The MapTube website: http://www.maptube.org
Perception and recognition of computer-enhanced facial attributes and abstracted prototypes
The influence of the human facial image was surveyed and the nature of its many interpretations were examined. The role of distinctiveness was considered particularly relevant as it accounted for many of the impressions of character and identity ascribed to individuals. The notion of structural differences with respect to some selective essence of normality is especially important as it allows a wide range of complex facial types to be considered and understood in an objective manner. A software tool was developed which permitted the manipulation of facial images. Quantitative distortions of digital images were examined using perceptual and recognition memory paradigms. Seven experiments investigated the role of distinctiveness in memory for faces using synthesised caricatures. The results showed that caricatures, both photographic and line-drawing, improved recognition speed and accuracy, indicating that both veridical and distinctiveness information are coded for familiar faces in long-term memory. The impact of feature metrics on perceptual estimates of facial age was examined using 'age-caricatured' images and were found to be in relative accordance with the 'intended' computed age. Further modifying the semantics permitted the differences between individual faces to be visualised in terms of facial structure and skin texture patterns. Transformations of identity between two, or more, faces established the necessary matrices which can offer an understanding of facial expression in a categorical manner and the inherent interactions. A procedural extension allowed generation of composite images in which all features are perfectly aligned. Prototypical facial types specified in this manner enabled high-level manipulations to be made of gender and attractiveness; two experiments corroborated previously speculative material and thus gave credence to the prototype model. In summary, psychological assessment of computer-manipulated facial images demonstrated the validity of the objective techniques and highlighted particular parameters which contribute to our perception and recognition of the individual and of underlying facial types
Noise and morphogenesis: Uncertainty, randomness and control
This thesis presents a processual ontology of noise by virtue of which morphogenesis (in its most general understanding as the processes by which order/form is created) must be instantiated. Noise is here outlined as the far from equilibrium environment out of which metastable temporary âsolutionsâ can emerge as the system transitions through the pre-individual state space. While frequently addressed by humanities and arts studies on the basis of its supposed disruptive character (often in terms of aesthetics), this thesis aims to thoroughly examine noiseâs conceptual potencies. To explore and amplify the epistemic consequences not merely of the ineliminability of noise but of its originative power as well as within the course of the elimination of givenness by epistemology. This philosophical work is informed by many different fields of contemporary science (namely: statistical physics, information theory, probability theory, 4E cognition, synthetic biology, nonlinear dynamics, complexity science and computer science) in order to assess and highlight the problems of the metascientific and ideological foundations of diverse projects of prediction and control of uncertainty. From algorithmic surveillance back to cybernetics and how these rendered noise âinformationally hereticalâ. This conveys an analysis of how contemporary prediction technologies are dramatically transforming our relationship with the future and with uncertainty in a great number of our social structures. It is a philosophico-critical anthropology of data ontology and a critique of reductive pan-info-computationalism. Additionally, two practical examples of noise characterised as an enabling constraint for the functioning of complex adaptive systems are presented. These are at once biophysical and cognitive, : 1) interaction-dominance constituted by âpink noiseâ and 2) noise as a source of variability that cells may exploit in (synthetic) biology. Finally, noise is posited as an intractable active ontological randomness that limits the scope of determinism and that goes beyond unpredictability in any epistemological sense due to the insuperability of the situation in which epistemology finds itself following the critique of the given
Information and Communication Technologies in Tourism 2022
This open access book presents the proceedings of the International Federation for IT and Travel & Tourism (IFITT)âs 29th Annual International eTourism Conference, which assembles the latest research presented at the ENTER2022 conference, which will be held on January 11â14, 2022. The book provides an extensive overview of how information and communication technologies can be used to develop tourism and hospitality. It covers the latest research on various topics within the field, including augmented and virtual reality, website development, social media use, e-learning, big data, analytics, and recommendation systems. The readers will gain insights and ideas on how information and communication technologies can be used in tourism and hospitality. Academics working in the eTourism field, as well as students and practitioners, will find up-to-date information on the status of research