13,983 research outputs found
Picasso: A Modular Framework for Visualizing the Learning Process of Neural Network Image Classifiers
Picasso is a free open-source (Eclipse Public License) web application
written in Python for rendering standard visualizations useful for analyzing
convolutional neural networks. Picasso ships with occlusion maps and saliency
maps, two visualizations which help reveal issues that evaluation metrics like
loss and accuracy might hide: for example, learning a proxy classification
task. Picasso works with the Tensorflow deep learning framework, and Keras
(when the model can be loaded into the Tensorflow backend). Picasso can be used
with minimal configuration by deep learning researchers and engineers alike
across various neural network architectures. Adding new visualizations is
simple: the user can specify their visualization code and HTML template
separately from the application code.Comment: 9 pages, submission to the Journal of Open Research Software,
github.com/merantix/picass
Molding the Knowledge in Modular Neural Networks
Problem description. The learning of monolithic neural networks becomes harder with growing network size. Likewise the knowledge obtained while learning becomes harder to extract. Such disadvantages are caused by a lack of internal structure, that by its presence would reduce the degrees of freedom in evolving to a training target. A suitable internal structure with respect to modular network construction as well as to nodal discrimination is required. Details on the grouping and selection of nodes can sometimes be concluded from the characteristics of the application area; otherwise a comprehensive search within the solution space is necessary
Beware of the Small-World neuroscientist!
The SW has undeniably been one of the most popular network descriptors in the
neuroscience literature. Two main reasons for its lasting popularity are its
apparent ease of computation and the intuitions it is thought to provide on how
networked systems operate. Over the last few years, some pitfalls of the SW
construct and, more generally, of network summary measures, have widely been
acknowledged
Research Priorities for Robust and Beneficial Artificial Intelligence
Success in the quest for artificial intelligence has the potential to bring
unprecedented benefits to humanity, and it is therefore worthwhile to
investigate how to maximize these benefits while avoiding potential pitfalls.
This article gives numerous examples (which should by no means be construed as
an exhaustive list) of such worthwhile research aimed at ensuring that AI
remains robust and beneficial.Comment: This article gives examples of the type of research advocated by the
open letter for robust & beneficial AI at
http://futureoflife.org/ai-open-lette
Artificial Neural Network Methodology for Modelling and Forecasting Maize Crop Yield
A particular type of ĂąâŹĆArtificial neural network (ANN)ĂąâŹ, viz. Multilayered feedforward artificial neural network (MLFANN) has been described. To train such a network, two types of learning algorithms, namely Gradient descent algorithm (GDA) and Conjugate gradient descent algorithm (CGDA), have been discussed. The methodology has been illustrated by considering maize crop yield data as response variable and total human labour, farm power, fertilizer consumption, and pesticide consumption as predictors. The data have been taken from a recently concluded National Agricultural Technology Project of Division of Agricultural Economics, I.A.R.I., New Delhi. To train the neural network, relevant computer programs have been written in MATLAB software package using Neural network toolbox. It has been found that a three-layered MLFANN with (11,16) units in the two hidden layers performs best in terms of having minimum mean square errors (MSE) for training, validation, and test sets. Superiority of this MLFANN over multiple linear regression (MLR) analysis has also been demonstrated for the maize data considered in the study. It is hoped that, in future, research workers would start applying not only MLFANN but also some of the other more advanced ANN models, like âRadial basis function neural networkâ, and âGeneralized regression neural networkâ in their studies.Crop Production/Industries,
Convolutional neural networks: a magic bullet for gravitational-wave detection?
In the last few years, machine learning techniques, in particular
convolutional neural networks, have been investigated as a method to replace or
complement traditional matched filtering techniques that are used to detect the
gravitational-wave signature of merging black holes. However, to date, these
methods have not yet been successfully applied to the analysis of long
stretches of data recorded by the Advanced LIGO and Virgo gravitational-wave
observatories. In this work, we critically examine the use of convolutional
neural networks as a tool to search for merging black holes. We identify the
strengths and limitations of this approach, highlight some common pitfalls in
translating between machine learning and gravitational-wave astronomy, and
discuss the interdisciplinary challenges. In particular, we explain in detail
why convolutional neural networks alone cannot be used to claim a statistically
significant gravitational-wave detection. However, we demonstrate how they can
still be used to rapidly flag the times of potential signals in the data for a
more detailed follow-up. Our convolutional neural network architecture as well
as the proposed performance metrics are better suited for this task than a
standard binary classifications scheme. A detailed evaluation of our approach
on Advanced LIGO data demonstrates the potential of such systems as trigger
generators. Finally, we sound a note of caution by constructing adversarial
examples, which showcase interesting "failure modes" of our model, where inputs
with no visible resemblance to real gravitational-wave signals are identified
as such by the network with high confidence.Comment: First two authors contributed equally; appeared at Phys. Rev.
Neural Networks with Non-Uniform Embedding and Explicit Validation Phase to Assess Granger Causality
A challenging problem when studying a dynamical system is to find the
interdependencies among its individual components. Several algorithms have been
proposed to detect directed dynamical influences between time series. Two of
the most used approaches are a model-free one (transfer entropy) and a
model-based one (Granger causality). Several pitfalls are related to the
presence or absence of assumptions in modeling the relevant features of the
data. We tried to overcome those pitfalls using a neural network approach in
which a model is built without any a priori assumptions. In this sense this
method can be seen as a bridge between model-free and model-based approaches.
The experiments performed will show that the method presented in this work can
detect the correct dynamical information flows occurring in a system of time
series. Additionally we adopt a non-uniform embedding framework according to
which only the past states that actually help the prediction are entered into
the model, improving the prediction and avoiding the risk of overfitting. This
method also leads to a further improvement with respect to traditional Granger
causality approaches when redundant variables (i.e. variables sharing the same
information about the future of the system) are involved. Neural networks are
also able to recognize dynamics in data sets completely different from the ones
used during the training phase
Learning Rigid Image Registration - Utilizing Convolutional Neural Networks for Medical Image Registration
Many traditional computer vision tasks, such as segmentation, have seen large step-changes in accuracy and/or speed with the application of Convolutional Neural Networks (CNNs). Image registration, the alignment of two or more images to a common space, is a fundamental step in many medical imaging workflows. In this paper we investigate whether these techniques can also bring tangible benefits to the registration task. We describe and evaluate the use of convolutional neural networks (CNNs) for both mono- and multi- modality registration and compare their performance to more traditional schemes, namely multi-scale, iterative registration. This paper also investigates incorporating inverse consistency of the learned spatial transformations to impose additional constraints on the network during training and investigate any benefit in accuracy during detection. The approaches are validated with a series of artificial mono-modal registration tasks utilizing T1-weighted MR brain i mages from the Open Access Series of Imaging Studies (OASIS) study and IXI brain development dataset and a series of real multi-modality registration tasks using T1-weighted and T2-weighted MR brain images from the 2015 Ischemia Stroke Lesion segmentation (ISLES) challenge. The results demonstrate that CNNs give excellent performance for both mono- and multi- modality head and neck registration compared to the baseline method with significantly fewer outliers and lower mean errors
- âŠ