39,021 research outputs found
Probabilistic Routing Protocol for Intermittently Connected Networks
This document is a product of the Delay Tolerant Networking Research Group and has been reviewed by that group. No objections to its publication as an RFC were raised.
This document defines PRoPHET, a Probabilistic Routing Protocol using History of Encounters and Transitivity. PRoPHET is a variant of the epidemic routing protocol for intermittently connected networks that operates by pruning the epidemic distribution tree to minimize resource usage while still attempting to achieve the best-case routing capabilities of epidemic routing. It is intended for use in sparse mesh networks where there is no guarantee that a fully connected path between the source and destination exists at any time, rendering traditional routing protocols unable to deliver messages between hosts. These networks are examples of networks where there is a disparity between the latency requirements of applications and the capabilities of the underlying network (networks often referred to as delay and disruption tolerant). The document presents an architectural overview followed by the protocol specification
The H.E.S.S. central data acquisition system
The High Energy Stereoscopic System (H.E.S.S.) is a system of Imaging
Atmospheric Cherenkov Telescopes (IACTs) located in the Khomas Highland in
Namibia. It measures cosmic gamma rays of very high energies (VHE; >100 GeV)
using the Earth's atmosphere as a calorimeter. The H.E.S.S. Array entered Phase
II in September 2012 with the inauguration of a fifth telescope that is larger
and more complex than the other four. This paper will give an overview of the
current H.E.S.S. central data acquisition (DAQ) system with particular emphasis
on the upgrades made to integrate the fifth telescope into the array. At first,
the various requirements for the central DAQ are discussed then the general
design principles employed to fulfil these requirements are described. Finally,
the performance, stability and reliability of the H.E.S.S. central DAQ are
presented. One of the major accomplishments is that less than 0.8% of
observation time has been lost due to central DAQ problems since 2009.Comment: 17 pages, 8 figures, published in Astroparticle Physic
Computing hypergeometric functions rigorously
We present an efficient implementation of hypergeometric functions in
arbitrary-precision interval arithmetic. The functions , ,
and (or the Kummer -function) are supported for
unrestricted complex parameters and argument, and by extension, we cover
exponential and trigonometric integrals, error functions, Fresnel integrals,
incomplete gamma and beta functions, Bessel functions, Airy functions, Legendre
functions, Jacobi polynomials, complete elliptic integrals, and other special
functions. The output can be used directly for interval computations or to
generate provably correct floating-point approximations in any format.
Performance is competitive with earlier arbitrary-precision software, and
sometimes orders of magnitude faster. We also partially cover the generalized
hypergeometric function and computation of high-order parameter
derivatives.Comment: v2: corrected example in section 3.1; corrected timing data for case
E-G in section 8.5 (table 6, figure 2); adjusted paper siz
DIANet: Dense-and-Implicit Attention Network
Attention networks have successfully boosted the performance in various
vision problems. Previous works lay emphasis on designing a new attention
module and individually plug them into the networks. Our paper proposes a
novel-and-simple framework that shares an attention module throughout different
network layers to encourage the integration of layer-wise information and this
parameter-sharing module is referred as Dense-and-Implicit-Attention (DIA)
unit. Many choices of modules can be used in the DIA unit. Since Long Short
Term Memory (LSTM) has a capacity of capturing long-distance dependency, we
focus on the case when the DIA unit is the modified LSTM (refer as DIA-LSTM).
Experiments on benchmark datasets show that the DIA-LSTM unit is capable of
emphasizing layer-wise feature interrelation and leads to significant
improvement of image classification accuracy. We further empirically show that
the DIA-LSTM has a strong regularization ability on stabilizing the training of
deep networks by the experiments with the removal of skip connections or Batch
Normalization in the whole residual network. The code is released at
https://github.com/gbup-group/DIANet
Riemann-Theta Boltzmann Machine
A general Boltzmann machine with continuous visible and discrete integer
valued hidden states is introduced. Under mild assumptions about the connection
matrices, the probability density function of the visible units can be solved
for analytically, yielding a novel parametric density function involving a
ratio of Riemann-Theta functions. The conditional expectation of a hidden state
for given visible states can also be calculated analytically, yielding a
derivative of the logarithmic Riemann-Theta function. The conditional
expectation can be used as activation function in a feedforward neural network,
thereby increasing the modelling capacity of the network. Both the Boltzmann
machine and the derived feedforward neural network can be successfully trained
via standard gradient- and non-gradient-based optimization techniques.Comment: 29 pages, 11 figures, final version published in Neurocomputin
Neural Sampling by Irregular Gating Inhibition of Spiking Neurons and Attractor Networks
A long tradition in theoretical neuroscience casts sensory processing in the
brain as the process of inferring the maximally consistent interpretations of
imperfect sensory input. Recently it has been shown that Gamma-band inhibition
can enable neural attractor networks to approximately carry out such a sampling
mechanism. In this paper we propose a novel neural network model based on
irregular gating inhibition, show analytically how it implements a Monte-Carlo
Markov Chain (MCMC) sampler, and describe how it can be used to model networks
of both neural attractors as well as of single spiking neurons. Finally we show
how this model applied to spiking neurons gives rise to a new putative
mechanism that could be used to implement stochastic synaptic weights in
biological neural networks and in neuromorphic hardware
- …