3,525 research outputs found

    La sostenibilitat del sistema públic de salut: de qui o de què depèn?

    Get PDF

    Modeling contact formation between atomic-sized gold tips via molecular dynamics

    Get PDF
    The formation and rupture of atomic-sized contacts is modelled by means of molecular dynamics simulations. Such nano-contacts are realized in scanning tunnelling microscope and mechanically controlled break junction experiments. These instruments routinely measure the conductance across the nano-sized electrodes as they are brought into contact and separated, permitting conductance traces to be recorded that are plots of conductance versus the distance between the electrodes. One interesting feature of the conductance traces is that for some metals and geometric configurations a jump in the value of the conductance is observed right before contact between the electrodes, a phenomenon known as jump-to-contact. This paper considers, from a computational point of view, the dynamics of contact between two gold nano-electrodes. Repeated indentation of the two surfaces on each other is performed in two crystallographic orientations of face-centred cubic gold, namely (001) and (111). Ultimately, the intention is to identify the structures at the atomic level at the moment of first contact between the surfaces, since the value of the conductance is related to the minimum cross-section in the contact region. Conductance values obtained in this way are determined using first principles electronic transport calculations, with atomic configurations taken from the molecular dynamics simulations serving as input structures.Comment: 6 pages, 4 figures, conference submissio

    The cosmic evolution of radio-AGN feedback to z=1

    Full text link
    This paper presents the first measurement of the radio luminosity function of 'jet-mode' (radiatively-inefficient) radio-AGN out to z=1, in order to investigate the cosmic evolution of radio-AGN feedback. Eight radio source samples are combined to produce a catalogue of 211 radio-loud AGN with 0.5<z<1.0, which are spectroscopically classified into jet-mode and radiative-mode (radiatively-efficient) AGN classes. Comparing with large samples of local radio-AGN from the Sloan Digital Sky Survey, the cosmic evolution of the radio luminosity function of each radio-AGN class is independently derived. Radiative-mode radio-AGN show an order of magnitude increase in space density out to z~1 at all luminosities, consistent with these AGN being fuelled by cold gas. In contrast, the space density of jet-mode radio-AGN decreases with increasing redshift at low radio luminosities (L_1.4 < 1e24 W/Hz) but increases at higher radio luminosities. Simple models are developed to explain the observed evolution. In the best-fitting models, the characteristic space density of jet-mode AGN declines with redshift in accordance with the declining space density of massive quiescent galaxies, which fuel them via cooling of gas in their hot haloes. A time delay of 1.5-2 Gyr may be present between the quenching of star formation and the onset of jet-mode radio-AGN activity. The behaviour at higher radio luminosities can be explained either by an increasing characteristic luminosity of jet-mode radio-AGN activity with redshift (roughly as (1+z) cubed) or if the jet-mode radio-AGN population also includes some contribution of cold-gas-fuelled sources seen at a time when their accretion rate was low. Higher redshifts measurements would distinguish between these possibilities.Comment: Accepted for publication in MNRA

    Performance of object recognition in wearable videos

    Full text link
    Wearable technologies are enabling plenty of new applications of computer vision, from life logging to health assistance. Many of them are required to recognize the elements of interest in the scene captured by the camera. This work studies the problem of object detection and localization on videos captured by this type of camera. Wearable videos are a much more challenging scenario for object detection than standard images or even another type of videos, due to lower quality images (e.g. poor focus) or high clutter and occlusion common in wearable recordings. Existing work typically focuses on detecting the objects of focus or those being manipulated by the user wearing the camera. We perform a more general evaluation of the task of object detection in this type of video, because numerous applications, such as marketing studies, also need detecting objects which are not in focus by the user. This work presents a thorough study of the well known YOLO architecture, that offers an excellent trade-off between accuracy and speed, for the particular case of object detection in wearable video. We focus our study on the public ADL Dataset, but we also use additional public data for complementary evaluations. We run an exhaustive set of experiments with different variations of the original architecture and its training strategy. Our experiments drive to several conclusions about the most promising directions for our goal and point us to further research steps to improve detection in wearable videos.Comment: Emerging Technologies and Factory Automation, ETFA, 201

    Event Transformer+. A multi-purpose solution for efficient event data processing

    Get PDF
    Event cameras record sparse illumination changes with high temporal resolution and high dynamic range. Thanks to their sparse recording and low consumption, they are increasingly used in applications such as AR/VR and autonomous driving. Current top-performing methods often ignore specific event-data properties, leading to the development of generic but computationally expensive algorithms, while event-aware methods do not perform as well. We propose Event Transformer+, that improves our seminal work evtprev EvT with a refined patch-based event representation and a more robust backbone to achieve more accurate results, while still benefiting from event-data sparsity to increase its efficiency. Additionally, we show how our system can work with different data modalities and propose specific output heads, for event-stream predictions (i.e. action recognition) and per-pixel predictions (dense depth estimation). Evaluation results show better performance to the state-of-the-art while requiring minimal computation resources, both on GPU and CPU
    corecore