234 research outputs found
OSNet & MNetO: Two Types of General Reconstruction Architectures for Linear Computed Tomography in Multi-Scenarios
Recently, linear computed tomography (LCT) systems have actively attracted
attention. To weaken projection truncation and image the region of interest
(ROI) for LCT, the backprojection filtration (BPF) algorithm is an effective
solution. However, in BPF for LCT, it is difficult to achieve stable interior
reconstruction, and for differentiated backprojection (DBP) images of LCT,
multiple rotation-finite inversion of Hilbert transform (Hilbert
filtering)-inverse rotation operations will blur the image. To satisfy multiple
reconstruction scenarios for LCT, including interior ROI, complete object, and
exterior region beyond field-of-view (FOV), and avoid the rotation operations
of Hilbert filtering, we propose two types of reconstruction architectures. The
first overlays multiple DBP images to obtain a complete DBP image, then uses a
network to learn the overlying Hilbert filtering function, referred to as the
Overlay-Single Network (OSNet). The second uses multiple networks to train
different directional Hilbert filtering models for DBP images of multiple
linear scannings, respectively, and then overlays the reconstructed results,
i.e., Multiple Networks Overlaying (MNetO). In two architectures, we introduce
a Swin Transformer (ST) block to the generator of pix2pixGAN to extract both
local and global features from DBP images at the same time. We investigate two
architectures from different networks, FOV sizes, pixel sizes, number of
projections, geometric magnification, and processing time. Experimental results
show that two architectures can both recover images. OSNet outperforms BPF in
various scenarios. For the different networks, ST-pix2pixGAN is superior to
pix2pixGAN and CycleGAN. MNetO exhibits a few artifacts due to the differences
among the multiple models, but any one of its models is suitable for imaging
the exterior edge in a certain direction.Comment: 13 pages, 13 figure
Learning the Network of Graphs for Graph Neural Networks
Graph neural networks (GNNs) have achieved great success in many scenarios
with graph-structured data. However, in many real applications, there are three
issues when applying GNNs: graphs are unknown, nodes have noisy features, and
graphs contain noisy connections. Aiming at solving these problems, we propose
a new graph neural network named as GL-GNN. Our model includes multiple
sub-modules, each sub-module selects important data features and learn the
corresponding key relation graph of data samples when graphs are unknown.
GL-GNN further obtains the network of graphs by learning the network of
sub-modules. The learned graphs are further fused using an aggregation method
over the network of graphs. Our model solves the first issue by simultaneously
learning multiple relation graphs of data samples as well as a relation network
of graphs, and solves the second and the third issue by selecting important
data features as well as important data sample relations. We compare our method
with 14 baseline methods on seven datasets when the graph is unknown and 11
baseline methods on two datasets when the graph is known. The results show that
our method achieves better accuracies than the baseline methods and is capable
of selecting important features and graph edges from the dataset. Our code will
be publicly available at \url{https://github.com/Looomo/GL-GNN}
Privacy-Aware UAV Flights through Self-Configuring Motion Planning
During flights, an unmanned aerial vehicle (UAV) may not be allowed to move across certain areas due to soft constraints such as privacy restrictions. Current methods on self-adaption focus mostly on motion planning such that the trajectory does not trespass predetermined restricted areas. When the environment is cluttered with uncertain obstacles, however, these motion planning algorithms are not flexible enough to find a trajectory that satisfies additional privacy-preserving requirements within a tight time budget during the flights. In this paper, we propose a privacy risk aware motion planning method through the reconfiguration of privacy-sensitive sensors. It minimises environmental impact by re-configuring the sensor during flight, while still guaranteeing the hard safety and energy constraints such as collision avoidance and timeliness. First, we formulate a model for assessing privacy risks of dynamically detected restricted areas. In case the UAV cannot find a feasible solution to satisfy both hard and soft constraints from the current configuration, our decision making method can then produce an optimal reconfiguration of the privacy-sensitive sensor with a more efficient trajectory. We evaluate the proposal through various simulations with different settings in a virtual environment and also validate the approach through real test flights on DJI Matrice 100 UAV
Two-View Topogram-Based Anatomy-Guided CT Reconstruction for Prospective Risk Minimization
To facilitate a prospective estimation of CT effective dose and risk
minimization process, a prospective spatial dose estimation and the known
anatomical structures are expected. To this end, a CT reconstruction method is
required to reconstruct CT volumes from as few projections as possible, i.e. by
using the topograms, with anatomical structures as correct as possible. In this
work, an optimized CT reconstruction model based on a generative adversarial
network (GAN) is proposed. The GAN is trained to reconstruct 3D volumes from an
anterior-posterior and a lateral CT projection. To enhance anatomical
structures, a pre-trained organ segmentation network and the 3D perceptual loss
are applied during the training phase, so that the model can then generate both
organ-enhanced CT volume and the organ segmentation mask. The proposed method
can reconstruct CT volumes with PSNR of 26.49, RMSE of 196.17, and SSIM of
0.64, compared to 26.21, 201.55 and 0.63 using the baseline method. In terms of
the anatomical structure, the proposed method effectively enhances the organ
shape and boundary and allows for a straight-forward identification of the
relevant anatomical structures. We note that conventional reconstruction
metrics fail to indicate the enhancement of anatomical structures. In addition
to such metrics, the evaluation is expanded with assessing the organ
segmentation performance. The average organ dice of the proposed method is 0.71
compared with 0.63 in baseline model, indicating the enhancement of anatomical
structures
AlignMiF: Geometry-Aligned Multimodal Implicit Field for LiDAR-Camera Joint Synthesis
Neural implicit fields have been a de facto standard in novel view synthesis.
Recently, there exist some methods exploring fusing multiple modalities within
a single field, aiming to share implicit features from different modalities to
enhance reconstruction performance. However, these modalities often exhibit
misaligned behaviors: optimizing for one modality, such as LiDAR, can adversely
affect another, like camera performance, and vice versa. In this work, we
conduct comprehensive analyses on the multimodal implicit field of LiDAR-camera
joint synthesis, revealing the underlying issue lies in the misalignment of
different sensors. Furthermore, we introduce AlignMiF, a geometrically aligned
multimodal implicit field with two proposed modules: Geometry-Aware Alignment
(GAA) and Shared Geometry Initialization (SGI). These modules effectively align
the coarse geometry across different modalities, significantly enhancing the
fusion process between LiDAR and camera data. Through extensive experiments
across various datasets and scenes, we demonstrate the effectiveness of our
approach in facilitating better interaction between LiDAR and camera modalities
within a unified neural field. Specifically, our proposed AlignMiF, achieves
remarkable improvement over recent implicit fusion methods (+2.01 and +3.11
image PSNR on the KITTI-360 and Waymo datasets) and consistently surpasses
single modality performance (13.8% and 14.2% reduction in LiDAR Chamfer
Distance on the respective datasets).Comment: CVPR202
- …