1,399 research outputs found
Designing fuzzy rule based classifier using self-organizing feature map for analysis of multispectral satellite images
We propose a novel scheme for designing fuzzy rule based classifier. An SOFM
based method is used for generating a set of prototypes which is used to
generate a set of fuzzy rules. Each rule represents a region in the feature
space that we call the context of the rule. The rules are tuned with respect to
their context. We justified that the reasoning scheme may be different in
different context leading to context sensitive inferencing. To realize context
sensitive inferencing we used a softmin operator with a tunable parameter. The
proposed scheme is tested on several multispectral satellite image data sets
and the performance is found to be much better than the results reported in the
literature.Comment: 23 pages, 7 figure
Collaborative Layer-wise Discriminative Learning in Deep Neural Networks
Intermediate features at different layers of a deep neural network are known
to be discriminative for visual patterns of different complexities. However,
most existing works ignore such cross-layer heterogeneities when classifying
samples of different complexities. For example, if a training sample has
already been correctly classified at a specific layer with high confidence, we
argue that it is unnecessary to enforce rest layers to classify this sample
correctly and a better strategy is to encourage those layers to focus on other
samples.
In this paper, we propose a layer-wise discriminative learning method to
enhance the discriminative capability of a deep network by allowing its layers
to work collaboratively for classification. Towards this target, we introduce
multiple classifiers on top of multiple layers. Each classifier not only tries
to correctly classify the features from its input layer, but also coordinates
with other classifiers to jointly maximize the final classification
performance. Guided by the other companion classifiers, each classifier learns
to concentrate on certain training examples and boosts the overall performance.
Allowing for end-to-end training, our method can be conveniently embedded into
state-of-the-art deep networks. Experiments with multiple popular deep
networks, including Network in Network, GoogLeNet and VGGNet, on scale-various
object classification benchmarks, including CIFAR100, MNIST and ImageNet, and
scene classification benchmarks, including MIT67, SUN397 and Places205,
demonstrate the effectiveness of our method. In addition, we also analyze the
relationship between the proposed method and classical conditional random
fields models.Comment: To appear in ECCV 2016. Maybe subject to minor changes before
camera-ready versio
Relative blocking in posets
Poset-theoretic generalizations of set-theoretic committee constructions are
presented. The structure of the corresponding subposets is described. Sequences
of irreducible fractions associated to the principal order ideals of finite
bounded posets are considered and those related to the Boolean lattices are
explored; it is shown that such sequences inherit all the familiar properties
of the Farey sequences.Comment: 29 pages. Corrected version of original publication which is
available at http://www.springerlink.com, see Corrigendu
Modeling concept drift: A probabilistic graphical model based approach
An often used approach for detecting and adapting to concept drift when doing classi cation is to treat the data as i.i.d. and use changes in classi cation accuracy as an indication of concept drift. In this paper, we take a different perspective and propose a framework, based on probabilistic graphical models, that explicitly represents concept drift using latent variables. To ensure effcient inference and learning, we resort to a variational Bayes inference scheme. As a proof of concept, we demonstrate and analyze the proposed framework using synthetic data sets as well as a real fi nancial data set from a Spanish bank
AI Researchers, Video Games Are Your Friends!
If you are an artificial intelligence researcher, you should look to video
games as ideal testbeds for the work you do. If you are a video game developer,
you should look to AI for the technology that makes completely new types of
games possible. This chapter lays out the case for both of these propositions.
It asks the question "what can video games do for AI", and discusses how in
particular general video game playing is the ideal testbed for artificial
general intelligence research. It then asks the question "what can AI do for
video games", and lays out a vision for what video games might look like if we
had significantly more advanced AI at our disposal. The chapter is based on my
keynote at IJCCI 2015, and is written in an attempt to be accessible to a broad
audience.Comment: in Studies in Computational Intelligence Studies in Computational
Intelligence, Volume 669 2017. Springe
A Compromise between Neutrino Masses and Collider Signatures in the Type-II Seesaw Model
A natural extension of the standard gauge
model to accommodate massive neutrinos is to introduce one Higgs triplet and
three right-handed Majorana neutrinos, leading to a neutrino mass
matrix which contains three sub-matrices ,
and . We show that three light Majorana neutrinos (i.e., the mass
eigenstates of , and ) are exactly massless in this
model, if and only if
exactly holds. This no-go theorem implies that small but non-vanishing neutrino
masses may result from a significant but incomplete cancellation between
and terms in the Type-II
seesaw formula, provided three right-handed Majorana neutrinos are of TeV and experimentally detectable at the LHC. We propose three simple
Type-II seesaw scenarios with the flavor symmetry to
interpret the observed neutrino mass spectrum and neutrino mixing pattern. Such
a TeV-scale neutrino model can be tested in two complementary ways: (1)
searching for possible collider signatures of lepton number violation induced
by the right-handed Majorana neutrinos and doubly-charged Higgs particles; and
(2) searching for possible consequences of unitarity violation of the neutrino mixing matrix in the future long-baseline neutrino oscillation
experiments.Comment: RevTeX 19 pages, no figure
Identifying discrete behavioural types: A re-analysis of public goods game contributions by hierarchical clustering
We propose a framework for identifying discrete behavioural types in experimental data. We re-analyse data from six previous studies of public goods voluntary contributions games. Using hierarchical clustering analysis, we construct a typology of behaviour based on a simi- larity measure between strategies. We identify four types with distinct sterotypical behaviours, which together account for about 90% of participants. Compared to previous approaches, our method produces a classification in which different types are more clearly distinguished in terms of strategic behaviour and the resulting economic implications
Unsupervised Bayesian linear unmixing of gene expression microarrays
Background: This paper introduces a new constrained model and the corresponding algorithm, called unsupervised Bayesian linear unmixing (uBLU), to identify biological signatures from high dimensional assays like gene expression microarrays. The basis for uBLU is a Bayesian model for the data samples which are represented as an additive mixture of random positive gene signatures, called factors, with random positive mixing coefficients, called factor scores, that specify the relative contribution of each signature to a specific sample. The particularity of the proposed method is that uBLU constrains the factor loadings to be non-negative and the factor scores to be probability distributions over the factors. Furthermore, it also provides estimates of the number of factors. A Gibbs sampling strategy is adopted here to generate random samples according to the posterior distribution of the factors, factor scores, and number of factors. These samples are then used to estimate all the unknown parameters. Results: Firstly, the proposed uBLU method is applied to several simulated datasets with known ground truth and compared with previous factor decomposition methods, such as principal component analysis (PCA), non negative matrix factorization (NMF), Bayesian factor regression modeling (BFRM), and the gradient-based algorithm for general matrix factorization (GB-GMF). Secondly, we illustrate the application of uBLU on a real time-evolving gene expression dataset from a recent viral challenge study in which individuals have been inoculated with influenza A/H3N2/Wisconsin. We show that the uBLU method significantly outperforms the other methods on the simulated and real data sets considered here. Conclusions: The results obtained on synthetic and real data illustrate the accuracy of the proposed uBLU method when compared to other factor decomposition methods from the literature (PCA, NMF, BFRM, and GB-GMF). The uBLU method identifies an inflammatory component closely associated with clinical symptom scores collected during the study. Using a constrained model allows recovery of all the inflammatory genes in a single factor
Automatic Network Fingerprinting through Single-Node Motifs
Complex networks have been characterised by their specific connectivity
patterns (network motifs), but their building blocks can also be identified and
described by node-motifs---a combination of local network features. One
technique to identify single node-motifs has been presented by Costa et al. (L.
D. F. Costa, F. A. Rodrigues, C. C. Hilgetag, and M. Kaiser, Europhys. Lett.,
87, 1, 2009). Here, we first suggest improvements to the method including how
its parameters can be determined automatically. Such automatic routines make
high-throughput studies of many networks feasible. Second, the new routines are
validated in different network-series. Third, we provide an example of how the
method can be used to analyse network time-series. In conclusion, we provide a
robust method for systematically discovering and classifying characteristic
nodes of a network. In contrast to classical motif analysis, our approach can
identify individual components (here: nodes) that are specific to a network.
Such special nodes, as hubs before, might be found to play critical roles in
real-world networks.Comment: 16 pages (4 figures) plus supporting information 8 pages (5 figures
Evaluation of machine-learning methods for ligand-based virtual screening
Machine-learning methods can be used for virtual screening by analysing the structural characteristics of molecules of known (in)activity, and we here discuss the use of kernel discrimination and naive Bayesian classifier (NBC) methods for this purpose. We report a kernel method that allows the processing of molecules represented by binary, integer and real-valued descriptors, and show that it is little different in screening performance from a previously described kernel that had been developed specifically for the analysis of binary fingerprint representations of molecular structure. We then evaluate the performance of an NBC when the training-set contains only a very few active molecules. In such cases, a simpler approach based on group fusion would appear to provide superior screening performance, especially when structurally heterogeneous datasets are to be processed
- …
