2,440 research outputs found
Symbolic Image Matching by Simulated Annealing
International audienceIn this paper we suggest an optimisation approach to visual matching. We assume that the information available in an image may be conveniently represented symbolically in a relational graph. We concentrate on the problem of matching two such graphs. First we derive a cost function associated with graph matching and more precisely associated with subgraph isomorphism and with maximum subgraph matching. This cost function is well suited for optimization methods such as simulated annealing. We show how the graph matching problem is easily cast into a simulated annealing algorithm. Finally we show some preliminary experimental results and discuss the utility of this graph matching method in computer vision in general
Structural graph matching using the EM algorithm and singular value decomposition
This paper describes an efficient algorithm for inexact graph matching. The method is purely structural, that is, it uses only the edge or connectivity structure of the graph and does not draw on node or edge attributes. We make two contributions: 1) commencing from a probability distribution for matching errors, we show how the problem of graph matching can be posed as maximum-likelihood estimation using the apparatus of the EM algorithm; and 2) we cast the recovery of correspondence matches between the graph nodes in a matrix framework. This allows one to efficiently recover correspondence matches using the singular value decomposition. We experiment with the method on both real-world and synthetic data. Here, we demonstrate that the method offers comparable performance to more computationally demanding method
Malware Classification based on Call Graph Clustering
Each day, anti-virus companies receive tens of thousands samples of
potentially harmful executables. Many of the malicious samples are variations
of previously encountered malware, created by their authors to evade
pattern-based detection. Dealing with these large amounts of data requires
robust, automatic detection approaches. This paper studies malware
classification based on call graph clustering. By representing malware samples
as call graphs, it is possible to abstract certain variations away, and enable
the detection of structural similarities between samples. The ability to
cluster similar samples together will make more generic detection techniques
possible, thereby targeting the commonalities of the samples within a cluster.
To compare call graphs mutually, we compute pairwise graph similarity scores
via graph matchings which approximately minimize the graph edit distance. Next,
to facilitate the discovery of similar malware samples, we employ several
clustering algorithms, including k-medoids and DBSCAN. Clustering experiments
are conducted on a collection of real malware samples, and the results are
evaluated against manual classifications provided by human malware analysts.
Experiments show that it is indeed possible to accurately detect malware
families via call graph clustering. We anticipate that in the future, call
graphs can be used to analyse the emergence of new malware families, and
ultimately to automate implementation of generic detection schemes.Comment: This research has been supported by TEKES - the Finnish Funding
Agency for Technology and Innovation as part of its ICT SHOK Future Internet
research programme, grant 40212/0
Chemical structure matching using correlation matrix memories
This paper describes the application of the Relaxation By Elimination (RBE) method to matching the 3D structure of molecules in chemical databases within the frame work of binary correlation matrix memories. The paper illustrates that, when combined with distributed representations, the method maps well onto these networks, allowing high performance implementation in parallel systems. It outlines the motivation, the neural architecture, the RBE method and presents some results of matching small molecules against a database of 100,000 models
Bridging ACT-R and Project Malmo, developing models of behavior in complex environments
Cognitive architectures such as ACT-R provide a system for simulating the mind and human behavior. On their own they model decision making of an isolated agent. However, applying a cognitive architecture to a complex environment yields more interesting results about how people make decisions in more realistic scenarios. Furthermore, cognitive architectures enable researchers to study human behavior in dangerous tasks which cannot be tested because they would harm participants. Nonetheless, these architectures aren’t commonly applied to such environments as they don’t come with one. It is left to the researcher to develop a task environment for their model. The difficulty in creating one prevents cognitive architectures from being utilized in more advanced studies. This project aims to address that issue by building a bridge between ACT-R and Project Malmo, an artificial general intelligence test suite. The bridge facilitates easy integration of new missions by allowing researchers to specify how to create the world and update it without worrying about the overhead of Malmo. Furthermore, this study analyses how well ACT-R’s utility learning system will adapt in a complex environment. The Adaptive Gain Theory was implemented to improve how the system adapts by using task engagement, derived from measures of utility, to dynamically modify noise. The system was tested using a modified Symbolic Maze task. Tests revealed the parameters of the Adaptive Gain mechanism need to be refined to have a greater impact on model performance. Nonetheless, the bridge provides an interface for ACT-R to be used to study decision making in a complex environment. Improving the bridge will enable more advanced experiments to be conducted whilst improving the Adaptive Gain Theory implementation will move us one step closer to understanding everyday intelligent behavior
Surrogate time series
Before we apply nonlinear techniques, for example those inspired by chaos
theory, to dynamical phenomena occurring in nature, it is necessary to first
ask if the use of such advanced techniques is justified "by the data". While
many processes in nature seem very unlikely a priori to be linear, the possible
nonlinear nature might not be evident in specific aspects of their dynamics.
The method of surrogate data has become a very popular tool to address such a
question. However, while it was meant to provide a statistically rigorous,
foolproof framework, some limitations and caveats have shown up in its
practical use. In this paper, recent efforts to understand the caveats, avoid
the pitfalls, and to overcome some of the limitations, are reviewed and
augmented by new material. In particular, we will discuss specific as well as
more general approaches to constrained randomisation, providing a full range of
examples. New algorithms will be introduced for unevenly sampled and
multivariate data and for surrogate spike trains. The main limitation, which
lies in the interpretability of the test results, will be illustrated through
instructive case studies. We will also discuss some implementational aspects of
the realisation of these methods in the TISEAN
(http://www.mpipks-dresden.mpg.de/~tisean) software package.Comment: 28 pages, 23 figures, software at
http://www.mpipks-dresden.mpg.de/~tisea
- …