1,888 research outputs found

    Confinement and diffusion time-scales of CR hadrons in AGN-inflated bubbles

    Full text link
    While rich clusters are powerful sources of X-rays, gamma-ray emission from these large cosmic structures has not been detected yet. X-ray radiative energy losses in the central regions of relaxed galaxy clusters are so strong that one needs to consider special sources of energy, likely AGN feedback, to suppress catastrophic cooling of the gas. We consider a model of AGN feedback that postulates that the AGN supplies the energy to the gas by inflating bubbles of relativistic plasma, whose energy content is dominated by cosmic-ray (CR) hadrons. If most of these hadrons can quickly escape the bubbles, then collisions of CRs with thermal protons in the intracluster medium (ICM) should lead to strong gamma-ray emission, unless fast diffusion of CRs removes them from the cluster. Therefore, the lack of detections with modern gamma-ray telescopes sets limits on the confinement time of CR hadrons in bubbles and CR diffusive propagation in the ICM.Comment: 8 pages, 2 figures, accepted for publication in MNRA

    Deep Neural Network for Structural Prediction and Lane Detection in Traffic Scene

    Full text link
    © 2016 IEEE. Hierarchical neural networks have been shown to be effective in learning representative image features and recognizing object classes. However, most existing networks combine the low/middle level cues for classification without accounting for any spatial structures. For applications such as understanding a scene, how the visual cues are spatially distributed in an image becomes essential for successful analysis. This paper extends the framework of deep neural networks by accounting for the structural cues in the visual signals. In particular, two kinds of neural networks have been proposed. First, we develop a multitask deep convolutional network, which simultaneously detects the presence of the target and the geometric attributes (location and orientation) of the target with respect to the region of interest. Second, a recurrent neuron layer is adopted for structured visual detection. The recurrent neurons can deal with the spatial distribution of visible cues belonging to an object whose shape or structure is difficult to explicitly define. Both the networks are demonstrated by the practical task of detecting lane boundaries in traffic scenes. The multitask convolutional neural network provides auxiliary geometric information to help the subsequent modeling of the given lane structures. The recurrent neural network automatically detects lane boundaries, including those areas containing no marks, without any explicit prior knowledge or secondary modeling

    An Equivalence Between Adaptive Dynamic Programming With a Critic and Backpropagation Through Time

    Get PDF
    We consider the adaptive dynamic programming technique called Dual Heuristic Programming (DHP), which is designed to learn a critic function, when using learned model functions of the environment. DHP is designed for optimizing control problems in large and continuous state spaces. We extend DHP into a new algorithm that we call Value-Gradient Learning, VGL(λ), and prove equivalence of an instance of the new algorithm to Backpropagation Through Time for Control with a greedy policy. Not only does this equivalence provide a link between these two different approaches, but it also enables our variant of DHP to have guaranteed convergence, under certain smoothness conditions and a greedy policy, when using a general smooth nonlinear function approximator for the critic. We consider several experimental scenarios including some that prove divergence of DHP under a greedy policy, which contrasts against our proven-convergent algorithm

    Clipping in Neurocontrol by Adaptive Dynamic Programming

    Get PDF
    In adaptive dynamic programming, neurocontrol, and reinforcement learning, the objective is for an agent to learn to choose actions so as to minimize a total cost function. In this paper, we show that when discretized time is used to model the motion of the agent, it can be very important to do clipping on the motion of the agent in the final time step of the trajectory. By clipping, we mean that the final time step of the trajectory is to be truncated such that the agent stops exactly at the first terminal state reached, and no distance further. We demonstrate that when clipping is omitted, learning performance can fail to reach the optimum, and when clipping is done properly, learning performance can improve significantly. The clipping problem we describe affects algorithms that use explicit derivatives of the model functions of the environment to calculate a learning gradient. These include backpropagation through time for control and methods based on dual heuristic programming. However, the clipping problem does not significantly affect methods based on heuristic dynamic programming, temporal differences learning, or policy-gradient learning algorithms
    • …
    corecore