2,051 research outputs found

    Face Authentication using Speed Fractal Technique

    Get PDF
    In this paper, a new fractal based recognition method, Face Authentication using Speed Fractal Technique (FAST), is presented. The main contribution is the good compromise between memory requirements, execution time and recognition ratio. FAST is based on Iterated Function Systems (IFS) theory, largely studied in still image compression and indexing, but not yet widely used for face recognition. Indeed, Fractals are well known to be invariant to a large set of global transformations. FAST is robust with respect to meaningful variations in facial expression and to the small changes of illumination and pose. Another advantage of the FAST strategy consists in the speed up that it introduces. The typical slowness of fractal image compression is avoided by exploiting only the indexing phase, which requires time O(D log (D)), where D is the size of the domain pool. Lastly, the FAST algorithm compares well to a large set of other recognition methods, as underlined in the experimental results

    A survey of parallel algorithms for fractal image compression

    Get PDF
    This paper presents a short survey of the key research work that has been undertaken in the application of parallel algorithms for Fractal image compression. The interest in fractal image compression techniques stems from their ability to achieve high compression ratios whilst maintaining a very high quality in the reconstructed image. The main drawback of this compression method is the very high computational cost that is associated with the encoding phase. Consequently, there has been significant interest in exploiting parallel computing architectures in order to speed up this phase, whilst still maintaining the advantageous features of the approach. This paper presents a brief introduction to fractal image compression, including the iterated function system theory upon which it is based, and then reviews the different techniques that have been, and can be, applied in order to parallelize the compression algorithm

    Giving eyes to ICT!, or How does a computer recognize a cow?

    Get PDF
    Het door Schouten en andere onderzoekers op het CWI ontwikkelde systeem berust op het beschrijven van beelden met behulp van fractale meetkunde. De menselijke waarneming blijkt mede daardoor zo efficiënt omdat zij sterk werkt met gelijkenissen. Het ligt dus voor de hand het te zoeken in wiskundige methoden die dat ook doen. Schouten heeft daarom beeldcodering met behulp van 'fractals' onderzocht. Fractals zijn zelfgelijkende meetkundige figuren, opgebouwd door herhaalde transformatie (iteratie) van een eenvoudig basispatroon, dat zich daardoor op steeds kleinere schalen vertakt. Op elk niveau van detaillering lijkt een fractal op zichzelf (Droste-effect). Met fractals kan men vrij eenvoudig bedrieglijk echte natuurvoorstellingen maken. Fractale beeldcodering gaat ervan uit dat het omgekeerde ook geldt: een beeld effectief opslaan in de vorm van de basispatronen van een klein aantal fractals, samen met het voorschrift hoe het oorspronkelijke beeld daaruit te reconstrueren. Het op het CWI in samenwerking met onderzoekers uit Leuven ontwikkelde systeem is mede gebaseerd op deze methode. ISBN 906196502

    The Application of Fractal Concept to Content-Based Image Retrieval

    Get PDF

    The Use of Quadtree Range Domain Partitioning with Fast Double Moment Descriptors to Enhance FIC of Colored Image

    Get PDF
    In this paper, an enhanced fractal image compression system (FIC) is proposed; it is based on using both symmetry prediction and blocks indexing to speed up the blocks matching process. The proposed FIC uses quad tree as variable range block partitioning mechanism. two criteria’s for guiding the partitioning decision are used: The first one uses sobel-based edge magnitude, whereas the second uses the contrast of block. A new set of moment descriptors are introduced, they differ from the previously used descriptors by their ability to emphasize the weights of different parts of each block. The effectiveness of all possible combinations of double moments descriptors has been investigated. Furthermore, a fast computation mechanism is introduced to compute the moments attended to improve the overall computation cost. the results of applied tests on the system for the cases “variable and fixed range” block partitioning mechanism indicated that the variable partitioning scheme can produce better results than fixed partitioning one (that is, 4 × 4 block) in term of compression ratio, faster than and PSNR does not significantly decreased

    Combining Fractal Coding and Orthogonal Linear Transforms

    Get PDF

    Feature Extraction Using Fractal Codes

    Get PDF
    Fast and successful searching for an object in a multimedia database is a highly desirable functionality. Several approaches to content based retrieval for multimedia databases can be found in the literature [9,10,12,14,17]. The approach we consider is feature extraction. A feature can be seen as a way to present simple information like the texture, color and spatial information of an image, or the pitch, frequency of a sound etc. In this paper we present a method for feature extraction on texture and spatial similarity, using fractal coding techniques. Our method is based upon the observation that the coefficients describing the fractal code of an image, contain very useful information about the structural content of the image. We apply simple statistics on information produced by fractal image coding. The statistics reveal features and require a small amount of storage. Several invariances are a consequence of the used methods: size, global contrast, orientation

    Fractal transforms and feature invariance

    Get PDF
    In this paper, fractal transforms are employed with the aim of image recognition. It is known that such transforms are highly sensitive to distortions like a small shift of an image. However, by using features based on statistics kept during the actual decomposition we can derive features from fractal transforms, which are invariant to perturbations like rotation, translation, folding or contrast scaling. Further, we introduce a feature invariance measure, which reveals the degree of invariance of a feature with respect to a database. The features and the way their invariance is measured, appear well suited for the application to images of textures

    Image databases, scale and fractal transforms

    Get PDF

    Efficient similarity search in high-dimensional data spaces

    Get PDF
    Similarity search in high-dimensional data spaces is a popular paradigm for many modern database applications, such as content based image retrieval, time series analysis in financial and marketing databases, and data mining. Objects are represented as high-dimensional points or vectors based on their important features. Object similarity is then measured by the distance between feature vectors and similarity search is implemented via range queries or k-Nearest Neighbor (k-NN) queries. Implementing k-NN queries via a sequential scan of large tables of feature vectors is computationally expensive. Building multi-dimensional indexes on the feature vectors for k-NN search also tends to be unsatisfactory when the dimensionality is high. This is due to the poor index performance caused by the dimensionality curse. Dimensionality reduction using the Singular Value Decomposition method is the approach adopted in this study to deal with high-dimensional data. Noting that for many real-world datasets, data distribution tends to be heterogeneous, dimensionality reduction on the entire dataset may cause a significant loss of information. More efficient representation is sought by clustering the data into homogeneous subsets of points, and applying dimensionality reduction to each cluster respectively, i.e., utilizing local rather than global dimensionality reduction. The thesis deals with the improvement of the efficiency of query processing associated with local dimensionality reduction methods, such as the Clustering and Singular Value Decomposition (CSVD) and the Local Dimensionality Reduction (LDR) methods. Variations in the implementation of CSVD are considered and the two methods are compared from the viewpoint of the compression ratio, CPU time, and retrieval efficiency. An exact k-NN algorithm is presented for local dimensionality reduction methods by extending an existing multi-step k-NN search algorithm, which is designed for global dimensionality reduction. Experimental results show that the new method requires less CPU time than the approximate method proposed original for CSVD at a comparable level of accuracy. Optimal subspace dimensionality reduction has the intent of minimizing total query cost. The problem is complicated in that each cluster can retain a different number of dimensions. A hybrid method is presented, combining the best features of the CSVD and LDR methods, to find optimal subspace dimensionalities for clusters generated by local dimensionality reduction methods. The experiments show that the proposed method works well for both real-world datasets and synthetic datasets
    corecore