122 research outputs found
Action Sensitivity Learning for Temporal Action Localization
Temporal action localization (TAL), which involves recognizing and locating
action instances, is a challenging task in video understanding. Most existing
approaches directly predict action classes and regress offsets to boundaries,
while overlooking the discrepant importance of each frame. In this paper, we
propose an Action Sensitivity Learning framework (ASL) to tackle this task,
which aims to assess the value of each frame and then leverage the generated
action sensitivity to recalibrate the training procedure. We first introduce a
lightweight Action Sensitivity Evaluator to learn the action sensitivity at the
class level and instance level, respectively. The outputs of the two branches
are combined to reweight the gradient of the two sub-tasks. Moreover, based on
the action sensitivity of each frame, we design an Action Sensitive Contrastive
Loss to enhance features, where the action-aware frames are sampled as positive
pairs to push away the action-irrelevant frames. The extensive studies on
various action localization benchmarks (i.e., MultiThumos, Charades,
Ego4D-Moment Queries v1.0, Epic-Kitchens 100, Thumos14 and ActivityNet1.3) show
that ASL surpasses the state-of-the-art in terms of average-mAP under multiple
types of scenarios, e.g., single-labeled, densely-labeled and egocentric.Comment: Accepted to ICCV 202
COPlanner: Plan to Roll Out Conservatively but to Explore Optimistically for Model-Based RL
Dyna-style model-based reinforcement learning contains two phases: model
rollouts to generate sample for policy learning and real environment
exploration using current policy for dynamics model learning. However, due to
the complex real-world environment, it is inevitable to learn an imperfect
dynamics model with model prediction error, which can further mislead policy
learning and result in sub-optimal solutions. In this paper, we propose
, a planning-driven framework for model-based methods to
address the inaccurately learned dynamics model problem with conservative model
rollouts and optimistic environment exploration. leverages
an uncertainty-aware policy-guided model predictive control (UP-MPC) component
to plan for multi-step uncertainty estimation. This estimated uncertainty then
serves as a penalty during model rollouts and as a bonus during real
environment exploration respectively, to choose actions. Consequently,
can avoid model uncertain regions through conservative
model rollouts, thereby alleviating the influence of model error.
Simultaneously, it explores high-reward model uncertain regions to reduce model
error actively through optimistic real environment exploration.
is a plug-and-play framework that can be applied to any
dyna-style model-based methods. Experimental results on a series of
proprioceptive and visual continuous control tasks demonstrate that both sample
efficiency and asymptotic performance of strong model-based methods are
significantly improved combined with .Comment: 22 pages, 17 figure
Game-Theoretic Robust Reinforcement Learning Handles Temporally-Coupled Perturbations
Robust reinforcement learning (RL) seeks to train policies that can perform
well under environment perturbations or adversarial attacks. Existing
approaches typically assume that the space of possible perturbations remains
the same across timesteps. However, in many settings, the space of possible
perturbations at a given timestep depends on past perturbations. We formally
introduce temporally-coupled perturbations, presenting a novel challenge for
existing robust RL methods. To tackle this challenge, we propose GRAD, a novel
game-theoretic approach that treats the temporally-coupled robust RL problem as
a partially-observable two-player zero-sum game. By finding an approximate
equilibrium in this game, GRAD ensures the agent's robustness against
temporally-coupled perturbations. Empirical experiments on a variety of
continuous control tasks demonstrate that our proposed approach exhibits
significant robustness advantages compared to baselines against both standard
and temporally-coupled attacks, in both state and action spaces
HOFA: Twitter Bot Detection with Homophily-Oriented Augmentation and Frequency Adaptive Attention
Twitter bot detection has become an increasingly important and challenging
task to combat online misinformation, facilitate social content moderation, and
safeguard the integrity of social platforms. Though existing graph-based
Twitter bot detection methods achieved state-of-the-art performance, they are
all based on the homophily assumption, which assumes users with the same label
are more likely to be connected, making it easy for Twitter bots to disguise
themselves by following a large number of genuine users. To address this issue,
we proposed HOFA, a novel graph-based Twitter bot detection framework that
combats the heterophilous disguise challenge with a homophily-oriented graph
augmentation module (Homo-Aug) and a frequency adaptive attention module
(FaAt). Specifically, the Homo-Aug extracts user representations and computes a
k-NN graph using an MLP and improves Twitter's homophily by injecting the k-NN
graph. For the FaAt, we propose an attention mechanism that adaptively serves
as a low-pass filter along a homophilic edge and a high-pass filter along a
heterophilic edge, preventing user features from being over-smoothed by their
neighborhood. We also introduce a weight guidance loss to guide the frequency
adaptive attention module. Our experiments demonstrate that HOFA achieves
state-of-the-art performance on three widely-acknowledged Twitter bot detection
benchmarks, which significantly outperforms vanilla graph-based bot detection
techniques and strong heterophilic baselines. Furthermore, extensive studies
confirm the effectiveness of our Homo-Aug and FaAt module, and HOFA's ability
to demystify the heterophilous disguise challenge.Comment: 11 pages, 7 figure
Mutually-paced Knowledge Distillation for Cross-lingual Temporal Knowledge Graph Reasoning
This paper investigates cross-lingual temporal knowledge graph reasoning
problem, which aims to facilitate reasoning on Temporal Knowledge Graphs (TKGs)
in low-resource languages by transfering knowledge from TKGs in high-resource
ones. The cross-lingual distillation ability across TKGs becomes increasingly
crucial, in light of the unsatisfying performance of existing reasoning methods
on those severely incomplete TKGs, especially in low-resource languages.
However, it poses tremendous challenges in two aspects. First, the
cross-lingual alignments, which serve as bridges for knowledge transfer, are
usually too scarce to transfer sufficient knowledge between two TKGs. Second,
temporal knowledge discrepancy of the aligned entities, especially when
alignments are unreliable, can mislead the knowledge distillation process. We
correspondingly propose a mutually-paced knowledge distillation model MP-KD,
where a teacher network trained on a source TKG can guide the training of a
student network on target TKGs with an alignment module. Concretely, to deal
with the scarcity issue, MP-KD generates pseudo alignments between TKGs based
on the temporal information extracted by our representation module. To maximize
the efficacy of knowledge transfer and control the noise caused by the temporal
knowledge discrepancy, we enhance MP-KD with a temporal cross-lingual attention
mechanism to dynamically estimate the alignment strength. The two procedures
are mutually paced along with model training. Extensive experiments on twelve
cross-lingual TKG transfer tasks in the EventKG benchmark demonstrate the
effectiveness of the proposed MP-KD method.Comment: This paper is accepted by The Web Conference 202
TACO: Temporal Latent Action-Driven Contrastive Loss for Visual Reinforcement Learning
Despite recent progress in reinforcement learning (RL) from raw pixel data,
sample inefficiency continues to present a substantial obstacle. Prior works
have attempted to address this challenge by creating self-supervised auxiliary
tasks, aiming to enrich the agent's learned representations with
control-relevant information for future state prediction. However, these
objectives are often insufficient to learn representations that can represent
the optimal policy or value function, and they often consider tasks with small,
abstract discrete action spaces and thus overlook the importance of action
representation learning in continuous control. In this paper, we introduce
TACO: Temporal Action-driven Contrastive Learning, a simple yet powerful
temporal contrastive learning approach that facilitates the concurrent
acquisition of latent state and action representations for agents. TACO
simultaneously learns a state and an action representation by optimizing the
mutual information between representations of current states paired with action
sequences and representations of the corresponding future states.
Theoretically, TACO can be shown to learn state and action representations that
encompass sufficient information for control, thereby improving sample
efficiency. For online RL, TACO achieves 40% performance boost after one
million environment interaction steps on average across nine challenging visual
continuous control tasks from Deepmind Control Suite. In addition, we show that
TACO can also serve as a plug-and-play module adding to existing offline visual
RL methods to establish the new state-of-the-art performance for offline visual
RL across offline datasets with varying quality
Antibiotics in the offshore waters of the Bohai Sea and the Yellow Sea in China: Occurrence, distribution and ecological risks
The ocean is an important sink of land-based pollutants. Previous studies showed that serious antibiotic pollution occurred in the coastal waters, but limited studies focused on their presence in offshore waters. In this study, eleven antibiotics in three different categories were investigated in offshore waters of the Bohai Sea and the Yellow Sea in China. The results indicated that three antibiotics dehydration erythromycin, sulfamethoxazole and trimethoprim occurred throughout the offshore waters at concentrations of 0.10-16.6 ng L-1 and they decreased exponentially from the rivers to the coastal and offshore waters. The other antibiotics all presented very low detection rates (<10%) and concentrations (<0.51 ng L-1). Although the concentrations were very low, risk assessment based on the calculated risk quotients (RQs) showed that sulfamethoxazole, dehydration erythromycin and clarithromycin at most of sampling sites posed medium or low ecological risks (0.01 < RQ < 1) to some sensitive aquatic organisms, including Synechococcus leopoliensis and Pseudokirchneriella subcapitata. (C) 2012 Elsevier Ltd. All rights reserved.The ocean is an important sink of land-based pollutants. Previous studies showed that serious antibiotic pollution occurred in the coastal waters, but limited studies focused on their presence in offshore waters. In this study, eleven antibiotics in three different categories were investigated in offshore waters of the Bohai Sea and the Yellow Sea in China. The results indicated that three antibiotics dehydration erythromycin, sulfamethoxazole and trimethoprim occurred throughout the offshore waters at concentrations of 0.10-16.6 ng L-1 and they decreased exponentially from the rivers to the coastal and offshore waters. The other antibiotics all presented very low detection rates (<10%) and concentrations (<0.51 ng L-1). Although the concentrations were very low, risk assessment based on the calculated risk quotients (RQs) showed that sulfamethoxazole, dehydration erythromycin and clarithromycin at most of sampling sites posed medium or low ecological risks (0.01 < RQ < 1) to some sensitive aquatic organisms, including Synechococcus leopoliensis and Pseudokirchneriella subcapitata. (C) 2012 Elsevier Ltd. All rights reserved
- …