25 research outputs found

    Fast complexified quaternion Fourier transform

    Full text link
    A discrete complexified quaternion Fourier transform is introduced. This is a generalization of the discrete quaternion Fourier transform to the case where either or both of the signal/image and the transform kernel are complex quaternion-valued. It is shown how to compute the transform using four standard complex Fourier transforms and the properties of the transform are briefly discussed

    A maximum uncertainty LDA-based approach for limited sample size problems – with application to Face Recognition

    Get PDF
    A critical issue of applying Linear Discriminant Analysis (LDA) is both the singularity and instability of the within-class scatter matrix. In practice, particularly in image recognition applications such as face recognition, there are often a large number of pixels or pre-processed features available, but the total number of training patterns is limited and commonly less than the dimension of the feature space. In this paper, a new LDA-based method is proposed. It is based on a straighforward stabilisation approach for the within-class scatter matrix. In order to evaluate its effectiveness, experiments on face recognition using the well-known ORL and FERET face databases were carried out and compared with other LDA-based methods. The results indicate that our method improves the LDA classification performance when the within-class scatter matrix is not only singular but also poorly estimated, with or without a Principal Component Analysis intermediate step and using less linear discriminant features

    Artificial Intelligence in Hungary - The First 20 Years = Mesterséges intelligencia Magyarországon - az első 20 év

    Get PDF
    A magyarországi mesterséges intelligencia kutatások történetéről 1996-ban készített áttekintés 2006-ban korszerűsített változata, bőséges irodalom jegyzékkel

    Exploiting Attitude Sensing in Vision-Based Navigation for an Airship

    Get PDF
    An Attitude Heading Reference System (AHRS) is used to compensate for rotational motion, facilitating vision-based navigation above smooth terrain by generating virtual images to simulate pure translation movement. The AHRS combines inertial and earth field magnetic sensors to provide absolute orientation measurements, and our recently developed calibration routine determines the rotation between the frames of reference of the AHRS and the monocular camera. In this way, the rotation is compensated, and the remaining translational motion is recovered by directly finding a rigid transformation to register corresponding scene coordinates. With a horizontal ground plane, the pure translation model performs more accurately than image-only approaches, and this is evidenced by recovering the trajectory of our airship UAV and comparing with GPS data. Visual odometry is also fused with the GPS, and ground plane maps are generated from the estimated vehicle poses and used to evaluate the results. Finally, loop closure is detected by looking for a previous image of the same area, and an open source SLAM package based in 3D graph optimization is employed to correct the visual odometry drift. The accuracy of the height estimation is also evaluated against ground truth in a controlled environment

    A generic face processing framework: technologies, analyses and applications.

    Get PDF
    Jang Kim-fung.Thesis (M.Phil.)--Chinese University of Hong Kong, 2003.Includes bibliographical references (leaves 108-124).Abstracts in English and Chinese.Abstract --- p.iAcknowledgement --- p.iiiChapter 1 --- Introduction --- p.1Chapter 1.1 --- Background --- p.1Chapter 1.2 --- Introduction about Face Processing Framework --- p.4Chapter 1.2.1 --- Basic architecture --- p.4Chapter 1.2.2 --- Face detection --- p.5Chapter 1.2.3 --- Face tracking --- p.6Chapter 1.2.4 --- Face recognition --- p.6Chapter 1.3 --- The scope and contributions of the thesis --- p.7Chapter 1.4 --- The outline of the thesis --- p.8Chapter 2 --- Facial Feature Representation --- p.10Chapter 2.1 --- Facial feature analysis --- p.10Chapter 2.1.1 --- Pixel information --- p.11Chapter 2.1.2 --- Geometry information --- p.13Chapter 2.2 --- Extracting and coding of facial feature --- p.14Chapter 2.2.1 --- Face recognition --- p.15Chapter 2.2.2 --- Facial expression classification --- p.38Chapter 2.2.3 --- Other related work --- p.44Chapter 2.3 --- Discussion about facial feature --- p.48Chapter 2.3.1 --- Performance evaluation for face recognition --- p.49Chapter 2.3.2 --- Evolution of the face recognition --- p.52Chapter 2.3.3 --- Evaluation of two state-of-the-art face recog- nition methods --- p.53Chapter 2.4 --- Problem for current situation --- p.58Chapter 3 --- Face Detection Algorithms and Committee Ma- chine --- p.61Chapter 3.1 --- Introduction about face detection --- p.62Chapter 3.2 --- Face Detection Committee Machine --- p.64Chapter 3.2.1 --- Review of three approaches for committee machine --- p.65Chapter 3.2.2 --- The approach of FDCM --- p.68Chapter 3.3 --- Evaluation --- p.70Chapter 4 --- Facial Feature Localization --- p.73Chapter 4.1 --- Algorithm for gray-scale image: template match- ing and separability filter --- p.73Chapter 4.1.1 --- Position of face and eye region --- p.74Chapter 4.1.2 --- Position of irises --- p.75Chapter 4.1.3 --- Position of lip --- p.79Chapter 4.2 --- Algorithm for color image: eyemap and separa- bility filter --- p.81Chapter 4.2.1 --- Position of eye candidates --- p.81Chapter 4.2.2 --- Position of mouth candidates --- p.83Chapter 4.2.3 --- Selection of face candidates by cost function --- p.84Chapter 4.3 --- Evaluation --- p.85Chapter 4.3.1 --- Algorithm for gray-scale image --- p.86Chapter 4.3.2 --- Algorithm for color image --- p.88Chapter 5 --- Face Processing System --- p.92Chapter 5.1 --- System architecture and limitations --- p.92Chapter 5.2 --- Pre-processing module --- p.93Chapter 5.2.1 --- Ellipse color model --- p.94Chapter 5.3 --- Face detection module --- p.96Chapter 5.3.1 --- Choosing the classifier --- p.96Chapter 5.3.2 --- Verifying the candidate region --- p.97Chapter 5.4 --- Face tracking module --- p.99Chapter 5.4.1 --- Condensation algorithm --- p.99Chapter 5.4.2 --- Tracking the region using Hue color model --- p.101Chapter 5.5 --- Face recognition module --- p.102Chapter 5.5.1 --- Normalization --- p.102Chapter 5.5.2 --- Recognition --- p.103Chapter 5.6 --- Applications --- p.104Chapter 6 --- Conclusion --- p.106Bibliography --- p.10

    Polynomial approximation using particle swarm optimization of lineal enhanced neural networks with no hidden layers.

    Full text link
    This paper presents some ideas about a new neural network architecture that can be compared to a Taylor analysis when dealing with patterns. Such architecture is based on lineal activation functions with an axo-axonic architecture. A biological axo-axonic connection between two neurons is defined as the weight in a connection in given by the output of another third neuron. This idea can be implemented in the so called Enhanced Neural Networks in which two Multilayer Perceptrons are used; the first one will output the weights that the second MLP uses to computed the desired output. This kind of neural network has universal approximation properties even with lineal activation functions. There exists a clear difference between cooperative and competitive strategies. The former ones are based on the swarm colonies, in which all individuals share its knowledge about the goal in order to pass such information to other individuals to get optimum solution. The latter ones are based on genetic models, that is, individuals can die and new individuals are created combining information of alive one; or are based on molecular/celular behaviour passing information from one structure to another. A swarm-based model is applied to obtain the Neural Network, training the net with a Particle Swarm algorithm

    COMPUTATIONAL MODELLING OF HUMAN AESTHETIC PREFERENCES IN THE VISUAL DOMAIN: A BRAIN-INSPIRED APPROACH

    Get PDF
    Following the rise of neuroaesthetics as a research domain, computational aesthetics has also known a regain in popularity over the past decade with many works using novel computer vision and machine learning techniques to evaluate the aesthetic value of visual information. This thesis presents a new approach where low-level features inspired from the human visual system are extracted from images to train a machine learning-based system to classify visual information depending on its aesthetics, regardless of the type of visual media. Extensive tests are developed to highlight strengths and weaknesses of such low-level features while establishing good practices in the domain of study of computational aesthetics. The aesthetic classification system is not only tested on the most widely used dataset of photographs, called AVA, on which it is trained initially, but also on other photographic datasets to evaluate the robustness of the learnt aesthetic preferences over other rating communities. The system is then assessed in terms of aesthetic classification on other types of visual media to investigate whether the learnt aesthetic preferences represent photography rules or more general aesthetic rules. The skill transfer from aesthetic classification of photos to videos demonstrates a satisfying correct classification rate of videos without any prior training on the test set created by Tzelepis et al. Moreover, the initial photograph classifier can also be used on feature films to investigate the classifier’s learnt visual preferences, due to films providing a large number of frames easily labellable. The study on aesthetic classification of videos concludes with a case study on the work by an online content creator. The classifier recognised a significantly greater percentage of aesthetically high frames in videos filmed in studios than on-the-go. The results obtained across datasets containing videos of diverse natures manifest the extent of the system’s aesthetic knowledge. To conclude, the evolution of low-level visual features is studied in popular culture such as in paintings and brand logos. The work attempts to link aesthetic preferences during contemplation tasks such as aesthetic rating of photographs with preferred low-level visual features in art creation. It questions whether favoured visual features usage varies over the life of a painter, implicitly showing a relationship with artistic expertise. Findings display significant changes in use of universally preferred features over influential vi abstract painters’ careers such an increase in cardinal lines and the colour blue; changes that were not observed in landscape painters. Regarding brand logos, only a few features evolved in a significant manner, most of them being colour-related features. Despite the incredible amount of data available online, phenomena developing over an entire life are still complicated to study. These computational experiments show that simple approaches focusing on the fundamentals instead of high-level measures allow to analyse artists’ visual preferences, as well as extract a community’s visual preferences from photos or videos while limiting impact from cultural and personal experiences
    corecore