1,356 research outputs found

    Learning Multiple Temporal Matching for Time Series Classification

    No full text
    12International audienceIn real applications, time series are generally of complex structure, exhibiting different global behaviors within classes. To discriminate such challenging time series, we propose a multiple temporal matching approach that reveals the commonly shared features within classes, and the most differential ones across classes. For this, we rely on a new framework based on the variance/covariance criterion to strengthen or weaken matched observations according to the induced variability within and between classes. The experiments performed on real and synthetic datasets demonstrate the ability of the multiple temporal matching approach to capture fine-grained distinctions between time series

    Temporal Matching in Endoscopic Images for Remote-Controlled Robotic Surgery

    Get PDF
    Temporal matching is applied in the frame of the formation of high-level entities in remote-controlled robotic surgery. The objective is to track tumor boundaries over time to improve the segmentation stage in each image of the sequence to facilitate the tracking and localization of the tumor. It makes use of an attributed string matching technique to find the correspondence between tumor boundaries over time. Relationships are then exploited to reconstitute the tumor boundaries and remove the inconsistencies coming from the detection errors. Input data are free form shapes of different length representing the tumor boundary, extracted at a previous stage

    Rethinking matching-based few-shot action recognition

    Full text link
    Few-shot action recognition, i.e. recognizing new action classes given only a few examples, benefits from incorporating temporal information. Prior work either encodes such information in the representation itself and learns classifiers at test time, or obtains frame-level features and performs pairwise temporal matching. We first evaluate a number of matching-based approaches using features from spatio-temporal backbones, a comparison missing from the literature, and show that the gap in performance between simple baselines and more complicated methods is significantly reduced. Inspired by this, we propose Chamfer++, a non-temporal matching function that achieves state-of-the-art results in few-shot action recognition. We show that, when starting from temporal features, our parameter-free and interpretable approach can outperform all other matching-based and classifier methods for one-shot action recognition on three common datasets without using temporal information in the matching stage. Project page: https://jbertrand89.github.io/matching-based-fsarComment: Accepted at SCIA 202

    Computing maximum matchings in temporal graphs

    Get PDF
    Temporal graphs are graphs whose topology is subject to discrete changes over time. Given a static underlying graph G, a temporal graph is represented by assigning a set of integer time-labels to every edge e of G, indicating the discrete time steps at which e is active. We introduce and study the complexity of a natural temporal extension of the classical graph problem Maximum Matching, taking into account the dynamic nature of temporal graphs. In our problem, Maximum Temporal Matching, we are looking for the largest possible number of time-labeled edges (simply time-edges) (e,t) such that no vertex is matched more than once within any time window of Δ consecutive time slots, where Δ ∈ ℕ is given. The requirement that a vertex cannot be matched twice in any Δ-window models some necessary "recovery" period that needs to pass for an entity (vertex) after being paired up for some activity with another entity. We prove strong computational hardness results for Maximum Temporal Matching, even for elementary cases. To cope with this computational hardness, we mainly focus on fixed-parameter algorithms with respect to natural parameters, as well as on polynomial-time approximation algorithms

    Computing maximum matchings in temporal graphs.

    Get PDF
    Temporal graphs are graphs whose topology is subject to discrete changes over time. Given a static underlying graph G, a temporal graph is represented by assigning a set of integer time-labels to every edge e of G, indicating the discrete time steps at which e is active. We introduce and study the complexity of a natural temporal extension of the classical graph problem Maximum Matching, taking into account the dynamic nature of temporal graphs. In our problem, Maximum Temporal Matching, we are looking for the largest possible number of time-labeled edges (simply time-edges) (e,t) such that no vertex is matched more than once within any time window of Δ consecutive time slots, where Δ ∈ ℕ is given. The requirement that a vertex cannot be matched twice in any Δ-window models some necessary "recovery" period that needs to pass for an entity (vertex) after being paired up for some activity with another entity. We prove strong computational hardness results for Maximum Temporal Matching, even for elementary cases. To cope with this computational hardness, we mainly focus on fixed-parameter algorithms with respect to natural parameters, as well as on polynomial-time approximation algorithms

    Temporal matching between interoception and exteroception: electrophysiological responses in a heartbeat discrimination task

    Get PDF
    Recent studies on interoception emphasize the importance of multisensory integration between interoception and exteroception. One of the methods frequently applied for assessing interoceptive sensitivity is the heartbeat discrimination task, where individuals judge whether the timing of external stimuli (e.g., tones) are synchronized to their own heartbeat. Despite its extensive use in research, the neural dynamics underlying the temporal matching between interoceptive and exteroceptive stimuli in this task have remained unclear. The present study used electroencephalography (EEG) to examine the neural responses of healthy participants who performed a heartbeat discrimination task. We analyzed the differences between EEG responses to tones, which were likely to be perceived as “heartbeat-synchronous” (200 ms delayed from the R wave) or “heartbeat-asynchronous” (0 ms delayed). Possible associations of these neural differentiations with task performance were also investigated. Compared with the responses to heartbeat-asynchronous tones, heartbeat-synchronous tones caused a relative decrease in early gamma-band EEG response and an increase in later P2 event-related potential (ERP) amplitude. Condition differences in the EEG/ERP measures were not significantly correlated with the behavioral measures. The mechanisms underlying the observed neural responses and the possibility of electrophysiological measurement of interoceptive sensitivity are discussed in terms of two perspectives: the predictive coding framework and the cardiac-phase-dependent baroreceptor function.This version of the article may not completely replicate the final authoritative version published in Journal of Psychophysiology at https://doi.org/10.1027/0269-8803/a000224. It is not the version of record and is therefore not suitable for citation. Please do not copy or cite without the permission of the author(s)

    DropMAE: Masked Autoencoders with Spatial-Attention Dropout for Tracking Tasks

    Get PDF
    In this paper, we study masked autoencoder (MAE) pretraining on videos for matching-based downstream tasks, including visual object tracking (VOT) and video object segmentation (VOS). A simple extension of MAE is to randomly mask out frame patches in videos and reconstruct the frame pixels. However, we find that this simple baseline heavily relies on spatial cues while ignoring temporal relations for frame reconstruction, thus leading to sub-optimal temporal matching representations for VOT and VOS. To alleviate this problem, we propose DropMAE, which adaptively performs spatial-attention dropout in the frame reconstruction to facilitate temporal correspondence learning in videos. We show that our DropMAE is a strong and efficient temporal matching learner, which achieves better finetuning results on matching-based tasks than the ImageNet-based MAE with 2× faster pre-training speed. Moreover, we also find that motion diversity in pre-training videos is more important than scene diversity for improving the performance on VOT and VOS. Our pre-trained DropMAE model can be directly loaded in existing ViT-based trackers for fine-tuning without further modifications. Notably, DropMAE sets new state-of-the-art performance on 8 out of 9 highly competitive video tracking and segmentation datasets. Our code and pre-trained models are available at https://github.com/jimmy-dq/DropMAE.git
    corecore