19,695 research outputs found

    Meta-stable memory in an artificial immune network

    Get PDF
    Abstract. This paper describes an artificial immune system algorithm which implements a fairly close analogue of the memory mechanism proposed by Jerne(1) (usually known as the Immune Network Theory). The algorithm demonstrates the ability of these types of network to produce meta-stable structures representing populated regions of the antigen space. The networks produced retain their structure indefinitely and capture inherent structure within the sets of antigens used to train them. Results from running the algorithm on a variety of data sets are presented and shown to be stable over long time periods and wide ranges of parameters. The potential of the algorithm as a tool for multivariate data analysis is also explored.

    "Going back to our roots": second generation biocomputing

    Full text link
    Researchers in the field of biocomputing have, for many years, successfully "harvested and exploited" the natural world for inspiration in developing systems that are robust, adaptable and capable of generating novel and even "creative" solutions to human-defined problems. However, in this position paper we argue that the time has now come for a reassessment of how we exploit biology to generate new computational systems. Previous solutions (the "first generation" of biocomputing techniques), whilst reasonably effective, are crude analogues of actual biological systems. We believe that a new, inherently inter-disciplinary approach is needed for the development of the emerging "second generation" of bio-inspired methods. This new modus operandi will require much closer interaction between the engineering and life sciences communities, as well as a bidirectional flow of concepts, applications and expertise. We support our argument by examining, in this new light, three existing areas of biocomputing (genetic programming, artificial immune systems and evolvable hardware), as well as an emerging area (natural genetic engineering) which may provide useful pointers as to the way forward.Comment: Submitted to the International Journal of Unconventional Computin

    Capturing Regular Human Activity through a Learning Context Memory

    Get PDF
    A learning context memory consisting of two main parts is presented. The first part performs lossy data compression, keeping the amount of stored data at a minimum by combining similar context attributes ā€” the compression rate for the presented GPS data is 150:1 on average. The resulting data is stored in an appropriate data structure highlighting the level of compression. Elements with a high level of compression are used in the second part to form the start and end points of episodes capturing common activity consisting of consecutive events. The context memory is used to investigate how little context data can be stored containing still enough information to capture regular human activity

    Randomly Evolving Idiotypic Networks: Structural Properties and Architecture

    Full text link
    We consider a minimalistic dynamic model of the idiotypic network of B-lymphocytes. A network node represents a population of B-lymphocytes of the same specificity (idiotype), which is encoded by a bitstring. The links of the network connect nodes with complementary and nearly complementary bitstrings, allowing for a few mismatches. A node is occupied if a lymphocyte clone of the corresponding idiotype exists, otherwise it is empty. There is a continuous influx of new B-lymphocytes of random idiotype from the bone marrow. B-lymphocytes are stimulated by cross-linking their receptors with complementary structures. If there are too many complementary structures, steric hindrance prevents cross-linking. Stimulated cells proliferate and secrete antibodies of the same idiotype as their receptors, unstimulated lymphocytes die. Depending on few parameters, the autonomous system evolves randomly towards patterns of highly organized architecture, where the nodes can be classified into groups according to their statistical properties. We observe and describe analytically the building principles of these patterns, which allow to calculate number and size of the node groups and the number of links between them. The architecture of all patterns observed so far in simulations can be explained this way. A tool for real-time pattern identification is proposed.Comment: 19 pages, 15 figures, 4 table

    Primordial Evolution in the Finitary Process Soup

    Full text link
    A general and basic model of primordial evolution--a soup of reacting finitary and discrete processes--is employed to identify and analyze fundamental mechanisms that generate and maintain complex structures in prebiotic systems. The processes--Ļµ\epsilon-machines as defined in computational mechanics--and their interaction networks both provide well defined notions of structure. This enables us to quantitatively demonstrate hierarchical self-organization in the soup in terms of complexity. We found that replicating processes evolve the strategy of successively building higher levels of organization by autocatalysis. Moreover, this is facilitated by local components that have low structural complexity, but high generality. In effect, the finitary process soup spontaneously evolves a selection pressure that favors such components. In light of the finitary process soup's generality, these results suggest a fundamental law of hierarchical systems: global complexity requires local simplicity.Comment: 7 pages, 10 figures; http://cse.ucdavis.edu/~cmg/compmech/pubs/pefps.ht
    • ā€¦
    corecore