65,224 research outputs found
Interpreting Deep Visual Representations via Network Dissection
The success of recent deep convolutional neural networks (CNNs) depends on
learning hidden representations that can summarize the important factors of
variation behind the data. However, CNNs often criticized as being black boxes
that lack interpretability, since they have millions of unexplained model
parameters. In this work, we describe Network Dissection, a method that
interprets networks by providing labels for the units of their deep visual
representations. The proposed method quantifies the interpretability of CNN
representations by evaluating the alignment between individual hidden units and
a set of visual semantic concepts. By identifying the best alignments, units
are given human interpretable labels across a range of objects, parts, scenes,
textures, materials, and colors. The method reveals that deep representations
are more transparent and interpretable than expected: we find that
representations are significantly more interpretable than they would be under a
random equivalently powerful basis. We apply the method to interpret and
compare the latent representations of various network architectures trained to
solve different supervised and self-supervised training tasks. We then examine
factors affecting the network interpretability such as the number of the
training iterations, regularizations, different initializations, and the
network depth and width. Finally we show that the interpreted units can be used
to provide explicit explanations of a prediction given by a CNN for an image.
Our results highlight that interpretability is an important property of deep
neural networks that provides new insights into their hierarchical structure.Comment: *B. Zhou and D. Bau contributed equally to this work. 15 pages, 27
figure
Automatic Programming of Cellular Automata and Artificial Neural Networks Guided by Philosophy
Many computer models such as cellular automata and artificial neural networks
have been developed and successfully applied. However, in some cases, these
models might be restrictive on the possible solutions or their solutions might
be difficult to interpret. To overcome this problem, we outline a new approach,
the so-called allagmatic method, that automatically programs and executes
models with as little limitations as possible while maintaining human
interpretability. Earlier we described a metamodel and its building blocks
according to the philosophical concepts of structure (spatial dimension) and
operation (temporal dimension). They are entity, milieu, and update function
that together abstractly describe cellular automata, artificial neural
networks, and possibly any kind of computer model. By automatically combining
these building blocks in an evolutionary computation, interpretability might be
increased by the relationship to the metamodel, and models might be translated
into more interpretable models via the metamodel. We propose generic and
object-oriented programming to implement the entities and their milieus as
dynamic and generic arrays and the update function as a method. We show two
experiments where a simple cellular automaton and an artificial neural network
are automatically programmed, compiled, and executed. A target state is
successfully evolved and learned in the cellular automaton and artificial
neural network, respectively. We conclude that the allagmatic method can create
and execute cellular automaton and artificial neural network models in an
automated manner with the guidance of philosophy.Comment: 12 pages, 1 figur
- …
