413 research outputs found
Robust Component-based Network Localization with Noisy Range Measurements
Accurate and robust localization is crucial for wireless ad-hoc and sensor
networks. Among the localization techniques, component-based methods advance
themselves for conquering network sparseness and anchor sparseness. But
component-based methods are sensitive to ranging noises, which may cause a huge
accumulated error either in component realization or merging process. This
paper presents three results for robust component-based localization under
ranging noises. (1) For a rigid graph component, a novel method is proposed to
evaluate the graph's possible number of flip ambiguities under noises. In
particular, graph's \emph{MInimal sepaRators that are neaRly cOllineaR
(MIRROR)} is presented as the cause of flip ambiguity, and the number of
MIRRORs indicates the possible number of flip ambiguities under noise. (2) Then
the sensitivity of a graph's local deforming regarding ranging noises is
investigated by perturbation analysis. A novel Ranging Sensitivity Matrix (RSM)
is proposed to estimate the node location perturbations due to ranging noises.
(3) By evaluating component robustness via the flipping and the local deforming
risks, a Robust Component Generation and Realization (RCGR) algorithm is
developed, which generates components based on the robustness metrics. RCGR was
evaluated by simulations, which showed much better noise resistance and
locating accuracy improvements than state-of-the-art of component-based
localization algorithms.Comment: 9 pages, 15 figures, ICCCN 2018, Hangzhou, Chin
Sequential Experimental Design for X-Ray CT Using Deep Reinforcement Learning
In X-ray Computed Tomography (CT), projections from many angles are acquired
and used for 3D reconstruction. To make CT suitable for in-line quality
control, reducing the number of angles while maintaining reconstruction quality
is necessary. Sparse-angle tomography is a popular approach for obtaining 3D
reconstructions from limited data. To optimize its performance, one can adapt
scan angles sequentially to select the most informative angles for each scanned
object. Mathematically, this corresponds to solving and optimal experimental
design (OED) problem. OED problems are high-dimensional, non-convex, bi-level
optimization problems that cannot be solved online, i.e., during the scan. To
address these challenges, we pose the OED problem as a partially observable
Markov decision process in a Bayesian framework, and solve it through deep
reinforcement learning. The approach learns efficient non-greedy policies to
solve a given class of OED problems through extensive offline training rather
than solving a given OED problem directly via numerical optimization. As such,
the trained policy can successfully find the most informative scan angles
online. We use a policy training method based on the Actor-Critic approach and
evaluate its performance on 2D tomography with synthetic data
Benchmarking the Physical-world Adversarial Robustness of Vehicle Detection
Adversarial attacks in the physical world can harm the robustness of
detection models. Evaluating the robustness of detection models in the physical
world can be challenging due to the time-consuming and labor-intensive nature
of many experiments. Thus, virtual simulation experiments can provide a
solution to this challenge. However, there is no unified detection benchmark
based on virtual simulation environment. To address this challenge, we proposed
an instant-level data generation pipeline based on the CARLA simulator. Using
this pipeline, we generated the DCI dataset and conducted extensive experiments
on three detection models and three physical adversarial attacks. The dataset
covers 7 continuous and 1 discrete scenes, with over 40 angles, 20 distances,
and 20,000 positions. The results indicate that Yolo v6 had strongest
resistance, with only a 6.59% average AP drop, and ASA was the most effective
attack algorithm with a 14.51% average AP reduction, twice that of other
algorithms. Static scenes had higher recognition AP, and results under
different weather conditions were similar. Adversarial attack algorithm
improvement may be approaching its 'limitation'.Comment: CVPR 2023 worksho
Improved Motor Imagery Classification Using Adaptive Spatial Filters Based on Particle Swarm Optimization Algorithm
As a typical self-paced brain-computer interface (BCI) system, the motor
imagery (MI) BCI has been widely applied in fields such as robot control,
stroke rehabilitation, and assistance for patients with stroke or spinal cord
injury. Many studies have focused on the traditional spatial filters obtained
through the common spatial pattern (CSP) method. However, the CSP method can
only obtain fixed spatial filters for specific input signals. Besides, CSP
method only focuses on the variance difference of two types of
electroencephalogram (EEG) signals, so the decoding ability of EEG signals is
limited. To obtain more effective spatial filters for better extraction of
spatial features that can improve classification to MI-EEG, this paper proposes
an adaptive spatial filter solving method based on particle swarm optimization
algorithm (PSO). A training and testing framework based on filter bank and
spatial filters (FBCSP-ASP) is designed for MI EEG signal classification.
Comparative experiments are conducted on two public datasets (2a and 2b) from
BCI competition IV, which show the outstanding average recognition accuracy of
FBCSP-ASP. The proposed method has achieved significant performance improvement
on MI-BCI. The classification accuracy of the proposed method has reached
74.61% and 81.19% on datasets 2a and 2b, respectively. Compared with the
baseline algorithm (FBCSP), the proposed algorithm improves 11.44% and 7.11% on
two datasets respectively. Furthermore, the analysis based on mutual
information, t-SNE and Shapley values further proves that ASP features have
excellent decoding ability for MI-EEG signals, and explains the improvement of
classification performance by the introduction of ASP features.Comment: 25 pages, 8 figure
SkyScript: A Large and Semantically Diverse Vision-Language Dataset for Remote Sensing
Remote sensing imagery, despite its broad applications in helping achieve
Sustainable Development Goals and tackle climate change, has not yet benefited
from the recent advancements of versatile, task-agnostic vision language models
(VLMs). A key reason is that the large-scale, semantically diverse image-text
dataset required for developing VLMs is still absent for remote sensing images.
Unlike natural images, remote sensing images and their associated text
descriptions cannot be efficiently collected from the public Internet at scale.
In this work, we bridge this gap by using geo-coordinates to automatically
connect open, unlabeled remote sensing images with rich semantics covered in
OpenStreetMap, and thus construct SkyScript, a comprehensive vision-language
dataset for remote sensing images, comprising 2.6 million image-text pairs
covering 29K distinct semantic tags. With continual pre-training on this
dataset, we obtain a VLM that surpasses baseline models with a 6.2% average
accuracy gain in zero-shot scene classification across seven benchmark
datasets. It also demonstrates the ability of zero-shot transfer for
fine-grained object attribute classification and cross-modal retrieval. We hope
this dataset can support the advancement of VLMs for various multi-modal tasks
in remote sensing, such as open-vocabulary classification, retrieval,
captioning, and text-to-image synthesis.Comment: Accepted by AAAI 202
ViP3D: End-to-end Visual Trajectory Prediction via 3D Agent Queries
Existing autonomous driving pipelines separate the perception module from the
prediction module. The two modules communicate via hand-picked features such as
agent boxes and trajectories as interfaces. Due to this separation, the
prediction module only receives partial information from the perception module.
Even worse, errors from the perception modules can propagate and accumulate,
adversely affecting the prediction results. In this work, we propose ViP3D, a
visual trajectory prediction pipeline that leverages the rich information from
raw videos to predict future trajectories of agents in a scene. ViP3D employs
sparse agent queries throughout the pipeline, making it fully differentiable
and interpretable. Furthermore, we propose an evaluation metric for this novel
end-to-end visual trajectory prediction task. Extensive experimental results on
the nuScenes dataset show the strong performance of ViP3D over traditional
pipelines and previous end-to-end models.Comment: Project page is at https://tsinghua-mars-lab.github.io/ViP3
- …