2,664 research outputs found

    Physics Inspired Optimization on Semantic Transfer Features: An Alternative Method for Room Layout Estimation

    Full text link
    In this paper, we propose an alternative method to estimate room layouts of cluttered indoor scenes. This method enjoys the benefits of two novel techniques. The first one is semantic transfer (ST), which is: (1) a formulation to integrate the relationship between scene clutter and room layout into convolutional neural networks; (2) an architecture that can be end-to-end trained; (3) a practical strategy to initialize weights for very deep networks under unbalanced training data distribution. ST allows us to extract highly robust features under various circumstances, and in order to address the computation redundance hidden in these features we develop a principled and efficient inference scheme named physics inspired optimization (PIO). PIO's basic idea is to formulate some phenomena observed in ST features into mechanics concepts. Evaluations on public datasets LSUN and Hedau show that the proposed method is more accurate than state-of-the-art methods.Comment: To appear in CVPR 2017. Project Page: https://sites.google.com/view/st-pio

    Heavy quarkonium production through the top quark rare decays via the channels involving flavor changing neutral currents

    Full text link
    In the paper, we discuss the possibility of observation of heavy quarkoniums via the processes involving flavor changing neutral currents (FCNC). More explicitly, we systematically calculate the production of heavy charmonium and (cbˉ)(c\bar{b})-quarkonium through the top quark semi-exclusive rare FCNC decays in the framework of the non-relativistic QCD (NRQCD) factorization theory. Our results show that the total decay widths Γt→ηc=1.20−0.51−0.45+1.04+1.14×10−16\Gamma_{t\to \eta_c} =1.20^{+1.04+1.14}_{-0.51-0.45}\times 10^{-16} GeV, Γt→J/ψ=1.37−0.51−0.51+1.03+1.30×10−16\Gamma_{t\to J/\psi} =1.37^{+1.03+1.30}_{-0.51-0.51}\times 10^{-16} GeV, Γt→Bc=2.06−0.17−0.54+0.17+0.91×10−18\Gamma_{t\to B_c}=2.06^{+0.17+0.91}_{-0.17-0.54}\times 10^{-18} GeV, and Γt→Bc∗=6.27−0.62−1.64+0.63+2.78×10−18\Gamma_{t\to B^*_c}=6.27^{+0.63+2.78}_{-0.62-1.64}\times 10^{-18} GeV, where the uncertainties are from variation of quark masses and renormalization scales. Even though the decay widths are small, it is important to make a systematic study on the production of charmonium and (cbˉ)(c\bar{b})-quarkonium through the top-quark decays via FCNC in the Standard Model, which will provide useful guidance for future new physics research from the heavy quarkonium involved processes.Comment: 17 pages, 8 figures and 6 tables, to be published in European Physical Journal C. arXiv admin note: text overlap with arXiv:1304.1303 by other author

    SegNeXt: Rethinking Convolutional Attention Design for Semantic Segmentation

    Full text link
    We present SegNeXt, a simple convolutional network architecture for semantic segmentation. Recent transformer-based models have dominated the field of semantic segmentation due to the efficiency of self-attention in encoding spatial information. In this paper, we show that convolutional attention is a more efficient and effective way to encode contextual information than the self-attention mechanism in transformers. By re-examining the characteristics owned by successful segmentation models, we discover several key components leading to the performance improvement of segmentation models. This motivates us to design a novel convolutional attention network that uses cheap convolutional operations. Without bells and whistles, our SegNeXt significantly improves the performance of previous state-of-the-art methods on popular benchmarks, including ADE20K, Cityscapes, COCO-Stuff, Pascal VOC, Pascal Context, and iSAID. Notably, SegNeXt outperforms EfficientNet-L2 w/ NAS-FPN and achieves 90.6% mIoU on the Pascal VOC 2012 test leaderboard using only 1/10 parameters of it. On average, SegNeXt achieves about 2.0% mIoU improvements compared to the state-of-the-art methods on the ADE20K datasets with the same or fewer computations. Code is available at https://github.com/uyzhang/JSeg (Jittor) and https://github.com/Visual-Attention-Network/SegNeXt (Pytorch).Comment: SegNeXt, a simple CNN for semantic segmentation. Code is availabl

    AMT: All-Pairs Multi-Field Transforms for Efficient Frame Interpolation

    Full text link
    We present All-Pairs Multi-Field Transforms (AMT), a new network architecture for video frame interpolation. It is based on two essential designs. First, we build bidirectional correlation volumes for all pairs of pixels, and use the predicted bilateral flows to retrieve correlations for updating both flows and the interpolated content feature. Second, we derive multiple groups of fine-grained flow fields from one pair of updated coarse flows for performing backward warping on the input frames separately. Combining these two designs enables us to generate promising task-oriented flows and reduce the difficulties in modeling large motions and handling occluded areas during frame interpolation. These qualities promote our model to achieve state-of-the-art performance on various benchmarks with high efficiency. Moreover, our convolution-based model competes favorably compared to Transformer-based models in terms of accuracy and efficiency. Our code is available at https://github.com/MCG-NKU/AMT.Comment: Accepted to CVPR202
    • …
    corecore