5,684 research outputs found

    High brightness fully coherent X-ray amplifier seeded by a free-electron laser oscillator

    Full text link
    X-ray free-electron laser oscillator (XFELO) is expected to be a cutting edge tool for fully coherent X-ray laser generation, and undulator taper technique is well-known for considerably increasing the efficiency of free-electron lasers (FELs). In order to combine the advantages of these two schemes, FEL amplifier seeded by XFELO is proposed by simply using a chirped electron beam. With the right choice of the beam parameters, the bunch tail is within the gain bandwidth of XFELO, and lase to saturation, which will be served as a seeding for further amplification. Meanwhile, the bunch head which is outside the gain bandwidth of XFELO, is preserved and used in the following FEL amplifier. It is found that the natural "double-horn" beam current as well as residual energy chirp from chicane compressor are quite suitable for the new scheme. Inheriting the advantages from XFELO seeding and undulator tapering, it is feasible to generate nearly terawatt level, fully coherent X-ray pulses with unprecedented shot-to-shot stability, which might open up new scientific opportunities in various research fields.Comment: 8 pages, 8 figure

    Object Detection in Videos with Tubelet Proposal Networks

    Full text link
    Object detection in videos has drawn increasing attention recently with the introduction of the large-scale ImageNet VID dataset. Different from object detection in static images, temporal information in videos is vital for object detection. To fully utilize temporal information, state-of-the-art methods are based on spatiotemporal tubelets, which are essentially sequences of associated bounding boxes across time. However, the existing methods have major limitations in generating tubelets in terms of quality and efficiency. Motion-based methods are able to obtain dense tubelets efficiently, but the lengths are generally only several frames, which is not optimal for incorporating long-term temporal information. Appearance-based methods, usually involving generic object tracking, could generate long tubelets, but are usually computationally expensive. In this work, we propose a framework for object detection in videos, which consists of a novel tubelet proposal network to efficiently generate spatiotemporal proposals, and a Long Short-term Memory (LSTM) network that incorporates temporal information from tubelet proposals for achieving high object detection accuracy in videos. Experiments on the large-scale ImageNet VID dataset demonstrate the effectiveness of the proposed framework for object detection in videos.Comment: CVPR 201

    Bilinear Graph Neural Network with Neighbor Interactions

    Full text link
    Graph Neural Network (GNN) is a powerful model to learn representations and make predictions on graph data. Existing efforts on GNN have largely defined the graph convolution as a weighted sum of the features of the connected nodes to form the representation of the target node. Nevertheless, the operation of weighted sum assumes the neighbor nodes are independent of each other, and ignores the possible interactions between them. When such interactions exist, such as the co-occurrence of two neighbor nodes is a strong signal of the target node's characteristics, existing GNN models may fail to capture the signal. In this work, we argue the importance of modeling the interactions between neighbor nodes in GNN. We propose a new graph convolution operator, which augments the weighted sum with pairwise interactions of the representations of neighbor nodes. We term this framework as Bilinear Graph Neural Network (BGNN), which improves GNN representation ability with bilinear interactions between neighbor nodes. In particular, we specify two BGNN models named BGCN and BGAT, based on the well-known GCN and GAT, respectively. Empirical results on three public benchmarks of semi-supervised node classification verify the effectiveness of BGNN -- BGCN (BGAT) outperforms GCN (GAT) by 1.6% (1.5%) in classification accuracy.Codes are available at: https://github.com/zhuhm1996/bgnn.Comment: Accepted by IJCAI 2020. SOLE copyright holder is IJCAI (International Joint Conferences on Artificial Intelligence), all rights reserve
    • …
    corecore