5,266 research outputs found

    Automatic Generation of Grounded Visual Questions

    Full text link
    In this paper, we propose the first model to be able to generate visually grounded questions with diverse types for a single image. Visual question generation is an emerging topic which aims to ask questions in natural language based on visual input. To the best of our knowledge, it lacks automatic methods to generate meaningful questions with various types for the same visual input. To circumvent the problem, we propose a model that automatically generates visually grounded questions with varying types. Our model takes as input both images and the captions generated by a dense caption model, samples the most probable question types, and generates the questions in sequel. The experimental results on two real world datasets show that our model outperforms the strongest baseline in terms of both correctness and diversity with a wide margin.Comment: VQ

    The evolution of magnetic structure driven by a synthetic spin-orbit coupling in two-component Bose-Hubbard model

    Full text link
    We study the evolution of magnetic structure driven by a synthetic spin-orbit coupling in a one-dimensional two-component Bose-Hubbard model. In addition to the Mott insulator-superfluid transition, we found in Mott insulator phases a transition from a gapped ferromagnetic phase to a gapless chiral phase by increasing the strength of spin-orbit coupling. Further increasing the spin-orbit coupling drives a transition from the gapless chiral phase to a gapped antiferromagnetic phase. These magnetic structures persist in superfluid phases. In particular, in the chiral Mott insulator and chiral superfluid phases, incommensurability is observed in characteristic correlation functions. These unconventional Mott insulator phase and superfluid phase demonstrate the novel effects arising from the competition between the kinetic energy and the spin-orbit coupling.Comment: 9 fig; English polished, note adde

    Balanced Sparsity for Efficient DNN Inference on GPU

    Full text link
    In trained deep neural networks, unstructured pruning can reduce redundant weights to lower storage cost. However, it requires the customization of hardwares to speed up practical inference. Another trend accelerates sparse model inference on general-purpose hardwares by adopting coarse-grained sparsity to prune or regularize consecutive weights for efficient computation. But this method often sacrifices model accuracy. In this paper, we propose a novel fine-grained sparsity approach, balanced sparsity, to achieve high model accuracy with commercial hardwares efficiently. Our approach adapts to high parallelism property of GPU, showing incredible potential for sparsity in the widely deployment of deep learning services. Experiment results show that balanced sparsity achieves up to 3.1x practical speedup for model inference on GPU, while retains the same high model accuracy as fine-grained sparsity
    corecore