143 research outputs found
'Tax-free' 3DMM Conditional Face Generation
3DMM conditioned face generation has gained traction due to its well-defined
controllability; however, the trade-off is lower sample quality: Previous works
such as DiscoFaceGAN and 3D-FM GAN show a significant FID gap compared to the
unconditional StyleGAN, suggesting that there is a quality tax to pay for
controllability. In this paper, we challenge the assumption that quality and
controllability cannot coexist. To pinpoint the previous issues, we
mathematically formalize the problem of 3DMM conditioned face generation. Then,
we devise simple solutions to the problem under our proposed framework. This
results in a new model that effectively removes the quality tax between 3DMM
conditioned face GANs and the unconditional StyleGAN.Comment: Accepted to the AI for Content Creation Workshop at CVPR 202
Design of the Tsinghua Tabletop Kibble Balance
The Kibble balance is a precision instrument for realizing the mass unit, the
kilogram, in the new international system of units (SI). In recent years, an
important trend for Kibble balance experiments is to go tabletop, in which the
instrument's size is notably reduced while retaining a measurement accuracy of
. In this paper, we report a new design of a tabletop Kibble balance
to be built at Tsinghua University. The Tsinghua Kibble balance aims to deliver
a compact instrument for robust mass calibrations from 10 g to 1 kg with a
targeted measurement accuracy of 50 g or less. Some major features of the
Tsinghua Kibble balance system, including the design of a new magnet, one-mode
measurement scheme, the spring-compensated magnet moving mechanism, and
magnetic shielding considerations, are discussed.Comment: 8 pages, 9 figure
FE-Fusion-VPR: Attention-based Multi-Scale Network Architecture for Visual Place Recognition by Fusing Frames and Events
Traditional visual place recognition (VPR), usually using standard cameras,
is easy to fail due to glare or high-speed motion. By contrast, event cameras
have the advantages of low latency, high temporal resolution, and high dynamic
range, which can deal with the above issues. Nevertheless, event cameras are
prone to failure in weakly textured or motionless scenes, while standard
cameras can still provide appearance information in this case. Thus, exploiting
the complementarity of standard cameras and event cameras can effectively
improve the performance of VPR algorithms. In the paper, we propose
FE-Fusion-VPR, an attention-based multi-scale network architecture for VPR by
fusing frames and events. First, the intensity frame and event volume are fed
into the two-stream feature extraction network for shallow feature fusion.
Next, the three-scale features are obtained through the multi-scale fusion
network and aggregated into three sub-descriptors using the VLAD layer.
Finally, the weight of each sub-descriptor is learned through the descriptor
re-weighting network to obtain the final refined descriptor. Experimental
results show that on the Brisbane-Event-VPR and DDD20 datasets, the Recall@1 of
our FE-Fusion-VPR is 29.26% and 33.59% higher than Event-VPR and
Ensemble-EventVPR, and is 7.00% and 14.15% higher than MultiRes-NetVLAD and
NetVLAD. To our knowledge, this is the first end-to-end network that goes
beyond the existing event-based and frame-based SOTA methods to fuse frame and
events directly for VPR
Microstructure-Based Interfacial Tuning Mechanism of Capacitive Pressure Sensors for Electronic Skin
- …