249 research outputs found
Automorphism group of the subspace inclusion graph of a vector space
In a recent paper [Comm. Algebra, 44(2016) 4724-4731], Das introduced the
graph , called subspace inclusion graph on a finite
dimensional vector space , where the vertex set is the collection
of nontrivial proper subspaces of and two vertices are adjacent if
one is properly contained in another. Das studied the diameter, girth, clique
number, and chromatic number of when the base field
is arbitrary, and he also studied some other properties of
when the base field is finite. In this paper, the
automorphisms of are determined when the base field
is finite.Comment: 10 page
General Backpropagation Algorithm for Training Second-order Neural Networks
The artificial neural network is a popular framework in machine learning. To
empower individual neurons, we recently suggested that the current type of
neurons could be upgraded to 2nd order counterparts, in which the linear
operation between inputs to a neuron and the associated weights is replaced
with a nonlinear quadratic operation. A single 2nd order neurons already has a
strong nonlinear modeling ability, such as implementing basic fuzzy logic
operations. In this paper, we develop a general backpropagation (BP) algorithm
to train the network consisting of 2nd-order neurons. The numerical studies are
performed to verify of the generalized BP algorithm.Comment: 5 pages, 7 figures, 19 reference
Shamap: Shape-based Manifold Learning
For manifold learning, it is assumed that high-dimensional sample/data points
are embedded on a low-dimensional manifold. Usually, distances among samples
are computed to capture an underlying data structure. Here we propose a metric
according to angular changes along a geodesic line, thereby reflecting the
underlying shape-oriented information or a topological similarity between high-
and low-dimensional representations of a data cloud. Our results demonstrate
the feasibility and merits of the proposed dimensionality reduction scheme
Fuzzy Logic Interpretation of Quadratic Networks
Over past several years, deep learning has achieved huge successes in various
applications. However, such a data-driven approach is often criticized for lack
of interpretability. Recently, we proposed artificial quadratic neural networks
consisting of second-order neurons in potentially many layers. In each
second-order neuron, a quadratic function is used in the place of the inner
product in a traditional neuron, and then undergoes a nonlinear activation.
With a single second-order neuron, any fuzzy logic operation, such as XOR, can
be implemented. In this sense, any deep network constructed with quadratic
neurons can be interpreted as a deep fuzzy logic system. Since traditional
neural networks and second-order counterparts can represent each other and
fuzzy logic operations are naturally implemented in second-order neural
networks, it is plausible to explain how a deep neural network works with a
second-order network as the system model. In this paper, we generalize and
categorize fuzzy logic operations implementable with individual second-order
neurons, and then perform statistical/information theoretic analyses of
exemplary quadratic neural networks.Comment: 10 pages and 9 figure
Slim, Sparse, and Shortcut Networks
Over the recent years, deep learning has become the mainstream data-driven
approach to solve many real-world problems in many important areas. Among the
successful network architectures, shortcut connections are well established to
take the outputs of earlier layers as additional inputs to later layers, which
have produced excellent results. Despite the extraordinary power, there remain
important questions on the underlying mechanism and associated functionalities
regarding shortcuts. For example, why are the shortcuts powerful? How to tune
the shortcut topology to optimize the efficiency and capacity of the network
model? Along this direction, here we first demonstrate a topology of shortcut
connections that can make a one-neuron-wide deep network approximate any
univariate function. Then, we present a novel width-bounded universal
approximator in contrast to depth-bounded universal approximators. Next we
demonstrate a family of theoretically equivalent networks, corroborated by the
concerning statistical significance experiments, and their graph spectral
characterization, thereby associating the representation ability of neural
network with their graph spectral properties. Furthermore, we shed light on the
effect of concatenation shortcuts on the margin-based multi-class
generalization bound of deep networks. Encouraged by the positive results from
the bounds analysis, we instantiate a slim, sparse, and shortcut network
(S3-Net) and the experimental results demonstrate that the S3-Net can achieve
better learning performance than the densely connected networks and other
state-of-the-art models on some well-known benchmarks
Exploiting Color Name Space for Salient Object Detection
In this paper, we will investigate the contribution of color names for the
task of salient object detection. An input image is first converted to color
name space, which is consisted of 11 probabilistic channels. By exploiting a
surroundedness cue, we obtain a saliency map through a linear combination of a
set of sequential attention maps. To overcome the limitation of only using the
surroundedness cue, two global cues with respect to color names are invoked to
guide the computation of a weighted saliency map. Finally, we integrate the
above two saliency maps into a unified framework to generate the final result.
In addition, an improved post-processing procedure is introduced to effectively
suppress image backgrounds while uniformly highlight salient objects.
Experimental results show that the proposed model produces more accurate
saliency maps and performs well against twenty-one saliency models in terms of
three evaluation metrics on three public data sets.Comment: http://www.loujing.com/cns-sod
A New Type of Neurons for Machine Learning
In machine learning, the use of an artificial neural network is the
mainstream approach. Such a network consists of layers of neurons. These
neurons are of the same type characterized by the two features: (1) an inner
product of an input vector and a matching weighting vector of trainable
parameters and (2) a nonlinear excitation function. Here we investigate the
possibility of replacing the inner product with a quadratic function of the
input vector, thereby upgrading the 1st order neuron to the 2nd order neuron,
empowering individual neurons, and facilitating the optimization of neural
networks. Also, numerical examples are provided to illustrate the feasibility
and merits of the 2nd order neurons. Finally, further topics are discussed.Comment: 5 pages, 8 figures, 11 reference
On Interpretability of Artificial Neural Networks: A Survey
Deep learning as represented by the artificial deep neural networks (DNNs)
has achieved great success in many important areas that deal with text, images,
videos, graphs, and so on. However, the black-box nature of DNNs has become one
of the primary obstacles for their wide acceptance in mission-critical
applications such as medical diagnosis and therapy. Due to the huge potential
of deep learning, interpreting neural networks has recently attracted much
research attention. In this paper, based on our comprehensive taxonomy, we
systematically review recent studies in understanding the mechanism of neural
networks, describe applications of interpretability especially in medicine, and
discuss future directions of interpretability research, such as in relation to
fuzzy logic and brain science
Soft-Autoencoder and Its Wavelet Shrinkage Interpretation
Recently, deep learning becomes the main focus of machine learning research
and has greatly impacted many fields. However, deep learning is criticized for
lack of interpretability. As a successful unsupervised model in deep learning,
the autoencoder embraces a wide spectrum of applications, yet it suffers from
the model opaqueness as well. In this paper, we propose a new type of
convolutional autoencoders, termed as Soft-Autoencoder (Soft-AE), in which the
activation functions of encoding layers are implemented with adaptable
soft-thresholding units while decoding layers are realized with linear units.
Consequently, Soft-AE can be naturally interpreted as a learned cascaded
wavelet shrinkage system. Our denoising experiments demonstrate that Soft-AE
not only is interpretable but also offers a competitive performance relative to
its counterparts. Furthermore, we propose a generalized linear unit (GeLU) and
its truncated variant (tGeLU) to allow autoencoder for more tasks from
denoising to deblurring
Quadratic Autoencoder (Q-AE) for Low-dose CT Denoising
Inspired by complexity and diversity of biological neurons, our group
proposed quadratic neurons by replacing the inner product in current artificial
neurons with a quadratic operation on input data, thereby enhancing the
capability of an individual neuron. Along this direction, we are motivated to
evaluate the power of quadratic neurons in popular network architectures,
simulating human-like learning in the form of quadratic-neuron-based deep
learning. Our prior theoretical studies have shown important merits of
quadratic neurons and networks in representation, efficiency, and
interpretability. In this paper, we use quadratic neurons to construct an
encoder-decoder structure, referred as the quadratic autoencoder, and apply it
to low-dose CT denoising. The experimental results on the Mayo low-dose CT
dataset demonstrate the utility of quadratic autoencoder in terms of image
denoising and model efficiency. To our best knowledge, this is the first time
that the deep learning approach is implemented with a new type of neurons and
demonstrates a significant potential in the medical imaging field
- …