122 research outputs found

    Action Sensitivity Learning for Temporal Action Localization

    Full text link
    Temporal action localization (TAL), which involves recognizing and locating action instances, is a challenging task in video understanding. Most existing approaches directly predict action classes and regress offsets to boundaries, while overlooking the discrepant importance of each frame. In this paper, we propose an Action Sensitivity Learning framework (ASL) to tackle this task, which aims to assess the value of each frame and then leverage the generated action sensitivity to recalibrate the training procedure. We first introduce a lightweight Action Sensitivity Evaluator to learn the action sensitivity at the class level and instance level, respectively. The outputs of the two branches are combined to reweight the gradient of the two sub-tasks. Moreover, based on the action sensitivity of each frame, we design an Action Sensitive Contrastive Loss to enhance features, where the action-aware frames are sampled as positive pairs to push away the action-irrelevant frames. The extensive studies on various action localization benchmarks (i.e., MultiThumos, Charades, Ego4D-Moment Queries v1.0, Epic-Kitchens 100, Thumos14 and ActivityNet1.3) show that ASL surpasses the state-of-the-art in terms of average-mAP under multiple types of scenarios, e.g., single-labeled, densely-labeled and egocentric.Comment: Accepted to ICCV 202

    SOUP:A fleet management system for passenger demand prediction and competitive taxi supply

    Get PDF

    COPlanner: Plan to Roll Out Conservatively but to Explore Optimistically for Model-Based RL

    Full text link
    Dyna-style model-based reinforcement learning contains two phases: model rollouts to generate sample for policy learning and real environment exploration using current policy for dynamics model learning. However, due to the complex real-world environment, it is inevitable to learn an imperfect dynamics model with model prediction error, which can further mislead policy learning and result in sub-optimal solutions. In this paper, we propose COPlanner\texttt{COPlanner}, a planning-driven framework for model-based methods to address the inaccurately learned dynamics model problem with conservative model rollouts and optimistic environment exploration. COPlanner\texttt{COPlanner} leverages an uncertainty-aware policy-guided model predictive control (UP-MPC) component to plan for multi-step uncertainty estimation. This estimated uncertainty then serves as a penalty during model rollouts and as a bonus during real environment exploration respectively, to choose actions. Consequently, COPlanner\texttt{COPlanner} can avoid model uncertain regions through conservative model rollouts, thereby alleviating the influence of model error. Simultaneously, it explores high-reward model uncertain regions to reduce model error actively through optimistic real environment exploration. COPlanner\texttt{COPlanner} is a plug-and-play framework that can be applied to any dyna-style model-based methods. Experimental results on a series of proprioceptive and visual continuous control tasks demonstrate that both sample efficiency and asymptotic performance of strong model-based methods are significantly improved combined with COPlanner\texttt{COPlanner}.Comment: 22 pages, 17 figure

    Game-Theoretic Robust Reinforcement Learning Handles Temporally-Coupled Perturbations

    Full text link
    Robust reinforcement learning (RL) seeks to train policies that can perform well under environment perturbations or adversarial attacks. Existing approaches typically assume that the space of possible perturbations remains the same across timesteps. However, in many settings, the space of possible perturbations at a given timestep depends on past perturbations. We formally introduce temporally-coupled perturbations, presenting a novel challenge for existing robust RL methods. To tackle this challenge, we propose GRAD, a novel game-theoretic approach that treats the temporally-coupled robust RL problem as a partially-observable two-player zero-sum game. By finding an approximate equilibrium in this game, GRAD ensures the agent's robustness against temporally-coupled perturbations. Empirical experiments on a variety of continuous control tasks demonstrate that our proposed approach exhibits significant robustness advantages compared to baselines against both standard and temporally-coupled attacks, in both state and action spaces

    HOFA: Twitter Bot Detection with Homophily-Oriented Augmentation and Frequency Adaptive Attention

    Full text link
    Twitter bot detection has become an increasingly important and challenging task to combat online misinformation, facilitate social content moderation, and safeguard the integrity of social platforms. Though existing graph-based Twitter bot detection methods achieved state-of-the-art performance, they are all based on the homophily assumption, which assumes users with the same label are more likely to be connected, making it easy for Twitter bots to disguise themselves by following a large number of genuine users. To address this issue, we proposed HOFA, a novel graph-based Twitter bot detection framework that combats the heterophilous disguise challenge with a homophily-oriented graph augmentation module (Homo-Aug) and a frequency adaptive attention module (FaAt). Specifically, the Homo-Aug extracts user representations and computes a k-NN graph using an MLP and improves Twitter's homophily by injecting the k-NN graph. For the FaAt, we propose an attention mechanism that adaptively serves as a low-pass filter along a homophilic edge and a high-pass filter along a heterophilic edge, preventing user features from being over-smoothed by their neighborhood. We also introduce a weight guidance loss to guide the frequency adaptive attention module. Our experiments demonstrate that HOFA achieves state-of-the-art performance on three widely-acknowledged Twitter bot detection benchmarks, which significantly outperforms vanilla graph-based bot detection techniques and strong heterophilic baselines. Furthermore, extensive studies confirm the effectiveness of our Homo-Aug and FaAt module, and HOFA's ability to demystify the heterophilous disguise challenge.Comment: 11 pages, 7 figure

    Mutually-paced Knowledge Distillation for Cross-lingual Temporal Knowledge Graph Reasoning

    Full text link
    This paper investigates cross-lingual temporal knowledge graph reasoning problem, which aims to facilitate reasoning on Temporal Knowledge Graphs (TKGs) in low-resource languages by transfering knowledge from TKGs in high-resource ones. The cross-lingual distillation ability across TKGs becomes increasingly crucial, in light of the unsatisfying performance of existing reasoning methods on those severely incomplete TKGs, especially in low-resource languages. However, it poses tremendous challenges in two aspects. First, the cross-lingual alignments, which serve as bridges for knowledge transfer, are usually too scarce to transfer sufficient knowledge between two TKGs. Second, temporal knowledge discrepancy of the aligned entities, especially when alignments are unreliable, can mislead the knowledge distillation process. We correspondingly propose a mutually-paced knowledge distillation model MP-KD, where a teacher network trained on a source TKG can guide the training of a student network on target TKGs with an alignment module. Concretely, to deal with the scarcity issue, MP-KD generates pseudo alignments between TKGs based on the temporal information extracted by our representation module. To maximize the efficacy of knowledge transfer and control the noise caused by the temporal knowledge discrepancy, we enhance MP-KD with a temporal cross-lingual attention mechanism to dynamically estimate the alignment strength. The two procedures are mutually paced along with model training. Extensive experiments on twelve cross-lingual TKG transfer tasks in the EventKG benchmark demonstrate the effectiveness of the proposed MP-KD method.Comment: This paper is accepted by The Web Conference 202

    TACO: Temporal Latent Action-Driven Contrastive Loss for Visual Reinforcement Learning

    Full text link
    Despite recent progress in reinforcement learning (RL) from raw pixel data, sample inefficiency continues to present a substantial obstacle. Prior works have attempted to address this challenge by creating self-supervised auxiliary tasks, aiming to enrich the agent's learned representations with control-relevant information for future state prediction. However, these objectives are often insufficient to learn representations that can represent the optimal policy or value function, and they often consider tasks with small, abstract discrete action spaces and thus overlook the importance of action representation learning in continuous control. In this paper, we introduce TACO: Temporal Action-driven Contrastive Learning, a simple yet powerful temporal contrastive learning approach that facilitates the concurrent acquisition of latent state and action representations for agents. TACO simultaneously learns a state and an action representation by optimizing the mutual information between representations of current states paired with action sequences and representations of the corresponding future states. Theoretically, TACO can be shown to learn state and action representations that encompass sufficient information for control, thereby improving sample efficiency. For online RL, TACO achieves 40% performance boost after one million environment interaction steps on average across nine challenging visual continuous control tasks from Deepmind Control Suite. In addition, we show that TACO can also serve as a plug-and-play module adding to existing offline visual RL methods to establish the new state-of-the-art performance for offline visual RL across offline datasets with varying quality

    Antibiotics in the offshore waters of the Bohai Sea and the Yellow Sea in China: Occurrence, distribution and ecological risks

    Get PDF
    The ocean is an important sink of land-based pollutants. Previous studies showed that serious antibiotic pollution occurred in the coastal waters, but limited studies focused on their presence in offshore waters. In this study, eleven antibiotics in three different categories were investigated in offshore waters of the Bohai Sea and the Yellow Sea in China. The results indicated that three antibiotics dehydration erythromycin, sulfamethoxazole and trimethoprim occurred throughout the offshore waters at concentrations of 0.10-16.6 ng L-1 and they decreased exponentially from the rivers to the coastal and offshore waters. The other antibiotics all presented very low detection rates (<10%) and concentrations (<0.51 ng L-1). Although the concentrations were very low, risk assessment based on the calculated risk quotients (RQs) showed that sulfamethoxazole, dehydration erythromycin and clarithromycin at most of sampling sites posed medium or low ecological risks (0.01 < RQ < 1) to some sensitive aquatic organisms, including Synechococcus leopoliensis and Pseudokirchneriella subcapitata. (C) 2012 Elsevier Ltd. All rights reserved.The ocean is an important sink of land-based pollutants. Previous studies showed that serious antibiotic pollution occurred in the coastal waters, but limited studies focused on their presence in offshore waters. In this study, eleven antibiotics in three different categories were investigated in offshore waters of the Bohai Sea and the Yellow Sea in China. The results indicated that three antibiotics dehydration erythromycin, sulfamethoxazole and trimethoprim occurred throughout the offshore waters at concentrations of 0.10-16.6 ng L-1 and they decreased exponentially from the rivers to the coastal and offshore waters. The other antibiotics all presented very low detection rates (<10%) and concentrations (<0.51 ng L-1). Although the concentrations were very low, risk assessment based on the calculated risk quotients (RQs) showed that sulfamethoxazole, dehydration erythromycin and clarithromycin at most of sampling sites posed medium or low ecological risks (0.01 < RQ < 1) to some sensitive aquatic organisms, including Synechococcus leopoliensis and Pseudokirchneriella subcapitata. (C) 2012 Elsevier Ltd. All rights reserved
    corecore