138 research outputs found

    Learning a world model and planning with a self-organizing, dynamic neural system

    Full text link
    We present a connectionist architecture that can learn a model of the relations between perceptions and actions and use this model for behavior planning. State representations are learned with a growing self-organizing layer which is directly coupled to a perception and a motor layer. Knowledge about possible state transitions is encoded in the lateral connectivity. Motor signals modulate this lateral connectivity and a dynamic field on the layer organizes a planning process. All mechanisms are local and adaptation is based on Hebbian ideas. The model is continuous in the action, perception, and time domain.Comment: 9 pages, see http://www.marc-toussaint.net

    Feature-driven Emergence of Model Graphs for Object Recognition and Categorization

    Get PDF
    An important requirement for the expression of cognitive structures is the ability to form mental objects by rapidly binding together constituent parts. In this sense, one may conceive the brain\u27s data structure to have the form of graphs whose nodes are labeled with elementary features. These provide a versatile data format with the additional ability to render the structure of any mental object. Because of the multitude of possible object variations the graphs are required to be dynamic. Upon presentation of an image a so-called model graph should rapidly emerge by binding together memorized subgraphs derived from earlier learning examples driven by the image features. In this model, the richness and flexibility of the mind is made possible by a combinatorical game of immense complexity. Consequently, the emergence of model graphs is a laborious task which, in computer vision, has most often been disregarded in favor of employing model graphs tailored to specific object categories like, for instance, faces in frontal pose. Recognition or categorization of arbitrary objects, however, demands dynamic graphs. In this work we propose a form of graph dynamics, which proceeds in two steps. In the first step component classifiers, which decide whether a feature is present in an image, are learned from training images. For processing arbitrary objects, features are small localized grid graphs, so-called parquet graphs, whose nodes are attributed with Gabor amplitudes. Through combination of these classifiers into a linear discriminant that conforms to Linsker\u27s infomax principle a weighted majority voting scheme is implemented. It allows for preselection of salient learning examples, so-called model candidates, and likewise for preselection of categories the object in the presented image supposably belongs to. Each model candidate is verified in a second step using a variant of elastic graph matching, a standard correspondence-based technique for face and object recognition. To further differentiate between model candidates with similar features it is asserted that the features be in similar spatial arrangement for the model to be selected. Model graphs are constructed dynamically by assembling model features into larger graphs according to their spatial arrangement. From the viewpoint of pattern recognition, the presented technique is a combination of a discriminative (feature-based) and a generative (correspondence-based) classifier while the majority voting scheme implemented in the feature-based part is an extension of existing multiple feature subset methods. We report the results of experiments on standard databases for object recognition and categorization. The method achieved high recognition rates on identity, object category, pose, and illumination type. Unlike many other models the presented technique can also cope with varying background, multiple objects, and partial occlusion

    Behaviourally meaningful representations from normalisation and context-guided denoising

    Get PDF
    Many existing independent component analysis algorithms include a preprocessing stage where the inputs are sphered. This amounts to normalising the data such that all correlations between the variables are removed. In this work, I show that sphering allows very weak contextual modulation to steer the development of meaningful features. Context-biased competition has been proposed as a model of covert attention and I propose that sphering-like normalisation also allows weaker top-down bias to guide attention

    Binding and Normalization of Binary Sparse Distributed Representations by Context-Dependent Thinning

    Get PDF
    Distributed representations were often criticized as inappropriate for encoding of data with a complex structure. However Plate's Holographic Reduced Representations and Kanerva's Binary Spatter Codes are recent schemes that allow on-the-fly encoding of nested compositional structures by real-valued or dense binary vectors of fixed dimensionality. In this paper we consider procedures of the Context-Dependent Thinning which were developed for representation of complex hierarchical items in the architecture of Associative-Projective Neural Networks. These procedures provide binding of items represented by sparse binary codevectors (with low probability of 1s). Such an encoding is biologically plausible and allows a high storage capacity of distributed associative memory where the codevectors may be stored. In contrast to known binding procedures, Context-Dependent Thinning preserves the same low density (or sparseness) of the bound codevector for varied number of component codevectors. Besides, a bound codevector is not only similar to another one with similar component codevectors (as in other schemes), but it is also similar to the component codevectors themselves. This allows the similarity of structures to be estimated just by the overlap of their codevectors, without retrieval of the component codevectors. This also allows an easy retrieval of the component codevectors. Examples of algorithmic and neural-network implementations of the thinning procedures are considered. We also present representation examples for various types of nested structured data (propositions using role-filler and predicate-arguments representation schemes, trees, directed acyclic graphs) using sparse codevectors of fixed dimension. Such representations may provide a fruitful alternative to the symbolic representations of traditional AI, as well as to the localist and microfeature-based connectionist representations

    Поиск аналогов с помощью распределенных представлений

    Get PDF
    Рассматриваются модели первой стадии рассуждений по аналогии: поиск в памяти аналога некоторой ситуации. Ситуации и аналоги иерархически структурированы и могут включать отношения высших порядков, что усложняет поиск при использовании традиционных подходов. Приводятся схемы распределенного представления аналогов в виде многомерных бинарных векторов. Показано, что степень сходства ситуаций можно оценить по величине скалярного произведения представляющих их векторов. Это создает основу для моделирования процессов поиска аналогов людьми и для более эффективного поиска в базах знаний.Розглядаються моделі першої стадії міркувань за аналогією: пошук у пам’яті аналогів деякої ситуації. Ситуації та аналоги ієрархічно структуровані і можуть включати відношення вищих порядків, що ускладнює пошук за умов використання традиційних підходів. Наводяться схеми розподіленого представлення аналогів у вигляді багатовимірних бінарних векторів. Показано, що ступінь схожості ситуацій можна відобразити за допомогою величини скалярного добутку векторів, які їх представляють. Це створює підґрунтя для моделювання процесів пошуку аналогів людьми та для більш ефективного пошуку у базах знань.Models of the first stage in analogical reasoning, analogical access, are considered. Schemes for distributed representations of hierarchically structured episodes as multidimensional binary vectors are presented. Such schemes reflect similarity of analogical episodes by the scalar product of vectors representing them. This simplifies searching for the most similar analogs and allows modeling of preferences demonstrated by humans in analogical access tasks

    GTM through time

    Get PDF
    The standard GTM (generative topographic mapping) algorithm assumes that the data on which it is trained consists of independent, identically distributed (iid) vectors. For time series, however, the iid assumption is a poor approximation. In this paper we show how the GTM algorithm can be extended to model time series by incorporating it as the emission density in a hidden Markov model. Since GTM has discrete hidden states we are able to find a tractable EM algorithm, based on the forward-backward algorithm, to train the model. We illustrate the performance of GTM through time using flight recorder data from a helicopter

    Phoneme-retrieval; voice recognition; vowels recognition

    Full text link
    A phoneme-retrieval technique is proposed, which is due to the particular way of the construction of the network. An initial set of neurons is given. The number of these neurons is approximately equal to the number of typical structures of the data. For example if the network is built for voice retrieval then the number of neurons must be equal to the number of characteristic phonemes of the alphabet of the language spoken by the social group to which the particular person belongs. Usually this task is very complicated and the network can depend critically on the samples used for the learning. If the network is built for image retrieval then it works only if the data to be retrieved belong to a particular set of images. If the network is built for voice recognition it works only for some particular set of words. A typical example is the words used for the flight of airplanes. For example a command like the "airplane should make a turn of 120 degrees towards the east" can be easily recognized by the network if a suitable learning procedure is used.Comment: 10 page

    Applying Deep Learning and Machine Learning Algorithms for The Identification of Medicinal Plant Leaves Based on Their Spectral Characteristics

    Get PDF
    The study and consideration of medicinal plants have been ongoing throughout history due to their significant role in maintaining the well-being of mammals. Although identifying medicinal plants can be a valuable skill, it is often time-consuming, tedious, and requires the expertise of a specialist. The project works on the technique of image processing, which identifies the various medicinal plants. There has been a strong emphasis on improving efficiency through the application of technology, with a focus on incorporating digital image processing and pattern recognition techniques. To ensure accurate plant identification, proposals involving the application of computer vision neural network techniques have been advanced. This approach involves neural network models such as CNN, SVM, KNN, and Navie Bay for identifying the medical plants based on their respective features. After the validation step, the project provides a classification of 92.3 precision and 90.56 F1 score
    corecore