13 research outputs found
Towards Automated Circuit Discovery for Mechanistic Interpretability
Recent work in mechanistic interpretability has reverse-engineered nontrivial
behaviors of transformer models. These contributions required considerable
effort and researcher intuition, which makes it difficult to apply the same
methods to understand the complex behavior that current models display. At
their core however, the workflow for these discoveries is surprisingly similar.
Researchers create a data set and metric that elicit the desired model
behavior, subdivide the network into appropriate abstract units, replace
activations of those units to identify which are involved in the behavior, and
then interpret the functions that these units implement. By varying the data
set, metric, and units under investigation, researchers can understand the
functionality of each neural network region and the circuits they compose. This
work proposes a novel algorithm, Automatic Circuit DisCovery (ACDC), to
automate the identification of the important units in the network. Given a
model's computational graph, ACDC finds subgraphs that explain a behavior of
the model. ACDC was able to reproduce a previously identified circuit for
Python docstrings in a small transformer, identifying 6/7 important attention
heads that compose up to 3 layers deep, while including 91% fewer the
connections
Jardins per a la salut
Facultat de FarmĂ cia, Universitat de Barcelona. Ensenyament: Grau de FarmĂ cia. Assignatura: BotĂ nica farmacĂšutica. Curs: 2014-2015. Coordinadors: Joan Simon, CĂšsar BlanchĂ© i Maria Bosch.Els materials que aquĂ es presenten sĂłn el recull de les fitxes botĂ niques de 128 espĂšcies presents en el JardĂ Ferran Soldevila de lâEdifici HistĂČric de la UB. Els treballs han estat realitzats manera individual per part dels estudiants dels grups M-3 i T-1 de lâassignatura BotĂ nica FarmacĂšutica durant els mesos de febrer a maig del curs 2014-15 com a resultat final del Projecte dâInnovaciĂł Docent «Jardins per a la salut: aprenentatge servei a BotĂ nica farmacĂšutica» (codi 2014PID-UB/054). Tots els treballs sâhan dut a terme a travĂ©s de la plataforma de GoogleDocs i han estat tutoritzats pels professors de lâassignatura. Lâobjectiu principal de lâactivitat ha estat fomentar lâaprenentatge autĂČnom i col·laboratiu en BotĂ nica farmacĂšutica. TambĂ© sâha pretĂšs motivar els estudiants a travĂ©s del retorn de part del seu esforç a la societat a travĂ©s dâuna experiĂšncia dâAprenentatge-Servei, deixant disponible finalment el treball dels estudiants per a poder ser consultable a travĂ©s dâuna Web pĂșblica amb la possibilitat de poder-ho fer in-situ en el propi jardĂ mitjançant codis QR amb un smartphone
Recommended from our members
Priors in finite and infinite Bayesian convolutional neural networks
Bayesian neural networks (BNNs) have undergone many changes since the seminal work of Neal [Nea96]. Advances in approximate inference and the use of GPUs have scaled BNNs to larger data sets, and much higher layer and parameter counts. Yet, the priors used for BNN parameters have remained essentially the same. The isotropic Gaussian prior introduced by Neal, where each element of the weights and biases is drawn independently from a Gaussian, is still used almost everywhere.
This thesis seeks to undo the neglect in the development of priors for BNNs, especially convolutional BNNs, using a two-pronged approach. First, I theoretically examine the effect of the Gaussian isotropic prior on the distribution over functions of a deep BNN prior. I show that, as the number of channels of a convolutional BNN goes to infinity, its output converges in distribution to a Gaussian process (GP). Thus, we can draw rough conclusions about the function-space of finite BNNs by looking at the mean and covariance of their limiting GPs.
The limiting GP itself performs surprisingly well at image classification, suggesting that knowledge encoded in the convolutional neural network (CNN) architecture, as opposed to the learned features, plays a larger role than previously thought.
Examining the derived CNN kernel shows that, if the weights are independent, the output of the limiting GP loses translation equivariance. This is an important inductive bias for learning from images. We can prevent this loss by introducing spatial correlations in the weight prior of a Bayesian CNN, which still results in a GP in the infinite width limit.
The second prong is an empirical methodology for identifying new priors for BNNs. Since BNNs are often considered to underfit, I examine the empirical distribution of weights learned using stochastic gradient descent (SGD). The resulting weight distributions tend to have heavier tails than a Gaussian, and display strong spatial correlations in CNNs.
I incorporate the found features into BNN priors, and test the performance of the resulting posterior. The spatially correlated priors, recommended by both prongs, greatly increase the classification performance of Bayesian CNNs. However, they do not at all reduce the cold-posterior effect (CPE), which indicates model misspecification or inference failure in BNNs. Heavy-tailed priors somewhat reduce the CPE in fully connected neural networks.
Ultimately, it is unlikely that the remaining misspecification is all in the prior. Nevertheless, I have found better priors for Bayesian CNNs. I have provided empirical methods that can be used to further improve BNN priors
Solving Montezuma's Revenge with Planning and Reinforcement Learning
Treball de fi de grau en informĂ ticaTutor: Anders JonssonTraditionally, methods for solving Sequential Decision Processes (SDPs) have not
worked well with those that feature sparse feedback. Both planning and reinforcement
learning, methods for solving SDPs, have trouble with it.
With the rise to prominence of the Arcade Learning Environment (ALE) in the
broader research community of sequential decision processes, one SDP featuring
sparse feedback has become familiar: the Atari game Montezumaâs Revenge. In this
particular game, the great amount of knowledge the human player already possesses,
and uses to find rewards, cannot be bridged by blindly exploring in a realistic time.
We apply planning and reinforcement learning approaches, combined with domain
knowledge, to enable an agent to obtain better scores in this game.
We hope that these domain-specific algorithms can inspire better approaches to solve
SDPs with sparse feedback in general
BNNpriors: A library for Bayesian neural network inference with different prior distributions
Bayesian neural networks have shown great promise in many applications where calibrated uncertainty estimates are crucial and can often also lead to a higher predictive performance. However, it remains challenging to choose a good prior distribution over their weights. While isotropic Gaussian priors are often chosen in practice due to their simplicity, they do not reflect our true prior beliefs well and can lead to suboptimal performance. Our new library, BNNpriors, enables state-of-the-art Markov Chain Monte Carlo inference on Bayesian neural networks with a wide range of predefined priors, including heavy-tailed ones, hierarchical ones, and mixture priors. Moreover, it follows a modular approach that eases the design and implementation of new custom priors. It has facilitated foundational discoveries on the nature of the cold posterior effect in Bayesian neural networks and will hopefully catalyze future research as well as practical applications in this area.ISSN:2665-963
TuringLang/AdvancedHMC.jl: v0.6.0
<h2>AdvancedHMC v0.6.0</h2>
<p><a href="https://github.com/TuringLang/AdvancedHMC.jl/compare/v0.5.5...v0.6.0">Diff since v0.5.5</a></p>
<p><strong>Merged pull requests:</strong></p>
<ul>
<li>fix: arg order (#349) (@xukai92)</li>
<li>CompatHelper: bump compat for AbstractMCMC to 5, (keep existing compat) (#352) (@github-actions[bot])</li>
<li>Deprecate <code>init_params</code> which is no longer in AbstractMCMC (#353) (@torfjelde)</li>
<li>CompatHelper: add new compat entry for Statistics at version 1, (keep existing compat) (#354) (@github-actions[bot])</li>
<li>Removed deprecation of init_params + bump minor version (#355) (@torfjelde)</li>
<li>Fix some tests. (#356) (@yebai)</li>
<li>Fix docs CI (#357) (@yebai)</li>
</ul>
<p><strong>Closed issues:</strong></p>
<ul>
<li>Doc string error for NUTS (#346)</li>
</ul>