4,043 research outputs found
Estimating effective connectivity in linear brain network models
Contemporary neuroscience has embraced network science to study the complex
and self-organized structure of the human brain; one of the main outstanding
issues is that of inferring from measure data, chiefly functional Magnetic
Resonance Imaging (fMRI), the so-called effective connectivity in brain
networks, that is the existing interactions among neuronal populations. This
inverse problem is complicated by the fact that the BOLD (Blood Oxygenation
Level Dependent) signal measured by fMRI represent a dynamic and nonlinear
transformation (the hemodynamic response) of neuronal activity. In this paper,
we consider resting state (rs) fMRI data; building upon a linear population
model of the BOLD signal and a stochastic linear DCM model, the model
parameters are estimated through an EM-type iterative procedure, which
alternately estimates the neuronal activity by means of the Rauch-Tung-Striebel
(RTS) smoother, updates the connections among neuronal states and refines the
parameters of the hemodynamic model; sparsity in the interconnection structure
is favoured using an iteratively reweighting scheme. Experimental results using
rs-fMRI data are shown demonstrating the effectiveness of our approach and
comparison with state of the art routines (SPM12 toolbox) is provided
Adaptive learning rates and parallelization for stochastic, sparse, non-smooth gradients
Recent work has established an empirically successful framework for adapting
learning rates for stochastic gradient descent (SGD). This effectively removes
all needs for tuning, while automatically reducing learning rates over time on
stationary problems, and permitting learning rates to grow appropriately in
non-stationary tasks. Here, we extend the idea in three directions, addressing
proper minibatch parallelization, including reweighted updates for sparse or
orthogonal gradients, improving robustness on non-smooth loss functions, in the
process replacing the diagonal Hessian estimation procedure that may not always
be available by a robust finite-difference approximation. The final algorithm
integrates all these components, has linear complexity and is hyper-parameter
free.Comment: Published at the First International Conference on Learning
Representations (ICLR-2013). Public reviews are available at
http://openreview.net/document/c14f2204-fd66-4d91-bed4-153523694041#c14f2204-fd66-4d91-bed4-15352369404
Proximal Multitask Learning over Networks with Sparsity-inducing Coregularization
In this work, we consider multitask learning problems where clusters of nodes
are interested in estimating their own parameter vector. Cooperation among
clusters is beneficial when the optimal models of adjacent clusters have a good
number of similar entries. We propose a fully distributed algorithm for solving
this problem. The approach relies on minimizing a global mean-square error
criterion regularized by non-differentiable terms to promote cooperation among
neighboring clusters. A general diffusion forward-backward splitting strategy
is introduced. Then, it is specialized to the case of sparsity promoting
regularizers. A closed-form expression for the proximal operator of a weighted
sum of -norms is derived to achieve higher efficiency. We also provide
conditions on the step-sizes that ensure convergence of the algorithm in the
mean and mean-square error sense. Simulations are conducted to illustrate the
effectiveness of the strategy
Counterfactual Learning from Bandit Feedback under Deterministic Logging: A Case Study in Statistical Machine Translation
The goal of counterfactual learning for statistical machine translation (SMT)
is to optimize a target SMT system from logged data that consist of user
feedback to translations that were predicted by another, historic SMT system. A
challenge arises by the fact that risk-averse commercial SMT systems
deterministically log the most probable translation. The lack of sufficient
exploration of the SMT output space seemingly contradicts the theoretical
requirements for counterfactual learning. We show that counterfactual learning
from deterministic bandit logs is possible nevertheless by smoothing out
deterministic components in learning. This can be achieved by additive and
multiplicative control variates that avoid degenerate behavior in empirical
risk minimization. Our simulation experiments show improvements of up to 2 BLEU
points by counterfactual learning from deterministic bandit feedback.Comment: Conference on Empirical Methods in Natural Language Processing
(EMNLP), 2017, Copenhagen, Denmar
- …