8 research outputs found

    Asymptotically Compatible Schemes for Nonlocal Ohta Kawasaki Model

    Full text link
    We study the asymptotical compatibility of the Fourier spectral method in multidimensional space for the Nonlocal Ohta-Kawasaka (NOK) model, which is proposed in our previous work. By introducing the Fourier collocation discretization for the spatial variable, we show that the asymptotical compatibility holds in 2D and 3D over a periodic domain. For the temporal discretization, we adopt the second-order backward differentiation formula (BDF) method. We prove that for certain nonlocal kernels, the proposed time discretization schemes inherit the energy dissipation law. In the numerical experiments, we verify the asymptotical compatibility, the second-order temporal convergence rate, and the energy stability of the proposed schemes. More importantly, we discover a novel square lattice pattern when certain nonlocal kernel are applied in the model. In addition, our numerical experiments confirm the existence of an upper bound for the optimal number of bubbles in 2D for some specific nonlocal kernels. Finally, we numerically explore the promotion/demotion effect induced by the nonlocal horizon, which is consistent with the theoretical studies presented in our earlier work

    Multi-grained Temporal Prototype Learning for Few-shot Video Object Segmentation

    Full text link
    Few-Shot Video Object Segmentation (FSVOS) aims to segment objects in a query video with the same category defined by a few annotated support images. However, this task was seldom explored. In this work, based on IPMT, a state-of-the-art few-shot image segmentation method that combines external support guidance information with adaptive query guidance cues, we propose to leverage multi-grained temporal guidance information for handling the temporal correlation nature of video data. We decompose the query video information into a clip prototype and a memory prototype for capturing local and long-term internal temporal guidance, respectively. Frame prototypes are further used for each frame independently to handle fine-grained adaptive guidance and enable bidirectional clip-frame prototype communication. To reduce the influence of noisy memory, we propose to leverage the structural similarity relation among different predicted regions and the support for selecting reliable memory frames. Furthermore, a new segmentation loss is also proposed to enhance the category discriminability of the learned prototypes. Experimental results demonstrate that our proposed video IPMT model significantly outperforms previous models on two benchmark datasets. Code is available at https://github.com/nankepan/VIPMT.Comment: ICCV 202

    MMBench: Is Your Multi-modal Model an All-around Player?

    Full text link
    Large vision-language models have recently achieved remarkable progress, exhibiting great perception and reasoning abilities concerning visual information. However, how to effectively evaluate these large vision-language models remains a major obstacle, hindering future model development. Traditional benchmarks like VQAv2 or COCO Caption provide quantitative performance measurements but suffer from a lack of fine-grained ability assessment and non-robust evaluation metrics. Recent subjective benchmarks, such as OwlEval, offer comprehensive evaluations of a model's abilities by incorporating human labor, but they are not scalable and display significant bias. In response to these challenges, we propose MMBench, a novel multi-modality benchmark. MMBench methodically develops a comprehensive evaluation pipeline, primarily comprised of two elements. The first element is a meticulously curated dataset that surpasses existing similar benchmarks in terms of the number and variety of evaluation questions and abilities. The second element introduces a novel CircularEval strategy and incorporates the use of ChatGPT. This implementation is designed to convert free-form predictions into pre-defined choices, thereby facilitating a more robust evaluation of the model's predictions. MMBench is a systematically-designed objective benchmark for robustly evaluating the various abilities of vision-language models. We hope MMBench will assist the research community in better evaluating their models and encourage future advancements in this domain. Project page: https://opencompass.org.cn/mmbench

    Historical changes of functions of the Grand Canal (Hangzhou section) and its role in development of Hangzhou City(运河(杭州段)功能的历史变迁及其对杭州城市发展的作用)

    No full text
    在考察运河(杭州段)水道的历史变迁的基础上,明确了其含义和范围;并进而论述了其功能的演变及其对杭州城市发展所起的作用.认为:其主要功能有航运、水利、纳污和景观、生态、文化等几项;其中,航运功能在古代阶段最为突出,对杭州城市的作用也最为显著,尤其在早期,决定了城市的兴起和繁荣,也控制了城市的基本空间形态;当代景观、生态和文化功能凸显,运河必将成为推动和促进城市发展、城市整体形象完善的重要因素

    The Crystal Chemistry of High-Temperature Oxide Superconductors and Materials with Related Structures

    No full text
    corecore