8 research outputs found
Asymptotically Compatible Schemes for Nonlocal Ohta Kawasaki Model
We study the asymptotical compatibility of the Fourier spectral method in
multidimensional space for the Nonlocal Ohta-Kawasaka (NOK) model, which is
proposed in our previous work. By introducing the Fourier collocation
discretization for the spatial variable, we show that the asymptotical
compatibility holds in 2D and 3D over a periodic domain. For the temporal
discretization, we adopt the second-order backward differentiation formula
(BDF) method. We prove that for certain nonlocal kernels, the proposed time
discretization schemes inherit the energy dissipation law. In the numerical
experiments, we verify the asymptotical compatibility, the second-order
temporal convergence rate, and the energy stability of the proposed schemes.
More importantly, we discover a novel square lattice pattern when certain
nonlocal kernel are applied in the model. In addition, our numerical
experiments confirm the existence of an upper bound for the optimal number of
bubbles in 2D for some specific nonlocal kernels. Finally, we numerically
explore the promotion/demotion effect induced by the nonlocal horizon, which is
consistent with the theoretical studies presented in our earlier work
Multi-grained Temporal Prototype Learning for Few-shot Video Object Segmentation
Few-Shot Video Object Segmentation (FSVOS) aims to segment objects in a query
video with the same category defined by a few annotated support images.
However, this task was seldom explored. In this work, based on IPMT, a
state-of-the-art few-shot image segmentation method that combines external
support guidance information with adaptive query guidance cues, we propose to
leverage multi-grained temporal guidance information for handling the temporal
correlation nature of video data. We decompose the query video information into
a clip prototype and a memory prototype for capturing local and long-term
internal temporal guidance, respectively. Frame prototypes are further used for
each frame independently to handle fine-grained adaptive guidance and enable
bidirectional clip-frame prototype communication. To reduce the influence of
noisy memory, we propose to leverage the structural similarity relation among
different predicted regions and the support for selecting reliable memory
frames. Furthermore, a new segmentation loss is also proposed to enhance the
category discriminability of the learned prototypes. Experimental results
demonstrate that our proposed video IPMT model significantly outperforms
previous models on two benchmark datasets. Code is available at
https://github.com/nankepan/VIPMT.Comment: ICCV 202
MMBench: Is Your Multi-modal Model an All-around Player?
Large vision-language models have recently achieved remarkable progress,
exhibiting great perception and reasoning abilities concerning visual
information. However, how to effectively evaluate these large vision-language
models remains a major obstacle, hindering future model development.
Traditional benchmarks like VQAv2 or COCO Caption provide quantitative
performance measurements but suffer from a lack of fine-grained ability
assessment and non-robust evaluation metrics. Recent subjective benchmarks,
such as OwlEval, offer comprehensive evaluations of a model's abilities by
incorporating human labor, but they are not scalable and display significant
bias. In response to these challenges, we propose MMBench, a novel
multi-modality benchmark. MMBench methodically develops a comprehensive
evaluation pipeline, primarily comprised of two elements. The first element is
a meticulously curated dataset that surpasses existing similar benchmarks in
terms of the number and variety of evaluation questions and abilities. The
second element introduces a novel CircularEval strategy and incorporates the
use of ChatGPT. This implementation is designed to convert free-form
predictions into pre-defined choices, thereby facilitating a more robust
evaluation of the model's predictions. MMBench is a systematically-designed
objective benchmark for robustly evaluating the various abilities of
vision-language models. We hope MMBench will assist the research community in
better evaluating their models and encourage future advancements in this
domain. Project page: https://opencompass.org.cn/mmbench
Historical changes of functions of the Grand Canal (Hangzhou section) and its role in development of Hangzhou City(运河(杭州段)功能的历史变迁及其对杭州城市发展的作用)
在考察运河(杭州段)水道的历史变迁的基础上,明确了其含义和范围;并进而论述了其功能的演变及其对杭州城市发展所起的作用.认为:其主要功能有航运、水利、纳污和景观、生态、文化等几项;其中,航运功能在古代阶段最为突出,对杭州城市的作用也最为显著,尤其在早期,决定了城市的兴起和繁荣,也控制了城市的基本空间形态;当代景观、生态和文化功能凸显,运河必将成为推动和促进城市发展、城市整体形象完善的重要因素