170 research outputs found
SoccerDB: A Large-Scale Database for Comprehensive Video Understanding
Soccer videos can serve as a perfect research object for video understanding
because soccer games are played under well-defined rules while complex and
intriguing enough for researchers to study. In this paper, we propose a new
soccer video database named SoccerDB, comprising 171,191 video segments from
346 high-quality soccer games. The database contains 702,096 bounding boxes,
37,709 essential event labels with time boundary and 17,115 highlight
annotations for object detection, action recognition, temporal action
localization, and highlight detection tasks. To our knowledge, it is the
largest database for comprehensive sports video understanding on various
aspects. We further survey a collection of strong baselines on SoccerDB, which
have demonstrated state-of-the-art performances on independent tasks. Our
evaluation suggests that we can benefit significantly when jointly considering
the inner correlations among those tasks. We believe the release of SoccerDB
will tremendously advance researches around comprehensive video understanding.
{\itshape Our dataset and code published on
https://github.com/newsdata/SoccerDB.}Comment: accepted by MM2020 sports worksho
Beating the Clauser-Horne-Shimony-Holt and the Svetlichny games with Optimal States
We study the relation between the maximal violation of Svetlichny's
inequality and the mixedness of quantum states and obtain the optimal state
(i.e., maximally nonlocal mixed states, or MNMS, for each value of linear
entropy) to beat the Clauser-Horne-Shimony-Holt and the Svetlichny games. For
the two-qubit and three-qubit MNMS, we showed that these states are also the
most tolerant state against white noise, and thus serve as valuable quantum
resources for such games. In particular, the quantum prediction of the MNMS
decreases as the linear entropy increases, and then ceases to be nonlocal when
the linear entropy reaches the critical points and for the
two- and three-qubit cases, respectively. The MNMS are related to classical
errors in experimental preparation of maximally entangled states.Comment: 7 pages, 3 figures; minor changes; accepted in Physical Review
LightNER: A Lightweight Tuning Paradigm for Low-resource NER via Pluggable Prompting
Most NER methods rely on extensive labeled data for model training, which
struggles in the low-resource scenarios with limited training data. Existing
dominant approaches usually suffer from the challenge that the target domain
has different label sets compared with a resource-rich source domain, which can
be concluded as class transfer and domain transfer. In this paper, we propose a
lightweight tuning paradigm for low-resource NER via pluggable prompting
(LightNER). Specifically, we construct the unified learnable verbalizer of
entity categories to generate the entity span sequence and entity categories
without any label-specific classifiers, thus addressing the class transfer
issue. We further propose a pluggable guidance module by incorporating
learnable parameters into the self-attention layer as guidance, which can
re-modulate the attention and adapt pre-trained weights. Note that we only tune
those inserted module with the whole parameter of the pre-trained language
model fixed, thus, making our approach lightweight and flexible for
low-resource scenarios and can better transfer knowledge across domains.
Experimental results show that LightNER can obtain comparable performance in
the standard supervised setting and outperform strong baselines in low-resource
settings. Code is in
https://github.com/zjunlp/DeepKE/tree/main/example/ner/few-shot.Comment: Accepted by COLING 202
- …