321 research outputs found
Transparent Contribution Evaluation for Secure Federated Learning on Blockchain
Federated Learning is a promising machine learning paradigm when multiple
parties collaborate to build a high-quality machine learning model.
Nonetheless, these parties are only willing to participate when given enough
incentives, such as a fair reward based on their contributions. Many studies
explored Shapley value based methods to evaluate each party's contribution to
the learned model. However, they commonly assume a semi-trusted server to train
the model and evaluate the data owners' model contributions, which lacks
transparency and may hinder the success of federated learning in practice. In
this work, we propose a blockchain-based federated learning framework and a
protocol to transparently evaluate each participant's contribution. Our
framework protects all parties' privacy in the model building phase and
transparently evaluates contributions based on the model updates. The
experiment with the handwritten digits dataset demonstrates that the proposed
method can effectively evaluate the contributions
GyroFlow: Gyroscope-Guided Unsupervised Optical Flow Learning
Existing optical flow methods are erroneous in challenging scenes, such as
fog, rain, and night because the basic optical flow assumptions such as
brightness and gradient constancy are broken. To address this problem, we
present an unsupervised learning approach that fuses gyroscope into optical
flow learning. Specifically, we first convert gyroscope readings into motion
fields named gyro field. Then, we design a self-guided fusion module to fuse
the background motion extracted from the gyro field with the optical flow and
guide the network to focus on motion details. To the best of our knowledge,
this is the first deep learning-based framework that fuses gyroscope data and
image content for optical flow learning. To validate our method, we propose a
new dataset that covers regular and challenging scenes. Experiments show that
our method outperforms the state-of-art methods in both regular and challenging
scenes
Exposure Fusion for Hand-held Camera Inputs with Optical Flow and PatchMatch
This paper proposes a hybrid synthesis method for multi-exposure image fusion
taken by hand-held cameras. Motions either due to the shaky camera or caused by
dynamic scenes should be compensated before any content fusion. Any
misalignment can easily cause blurring/ghosting artifacts in the fused result.
Our hybrid method can deal with such motions and maintain the exposure
information of each input effectively. In particular, the proposed method first
applies optical flow for a coarse registration, which performs well with
complex non-rigid motion but produces deformations at regions with missing
correspondences. The absence of correspondences is due to the occlusions of
scene parallax or the moving contents. To correct such error registration, we
segment images into superpixels and identify problematic alignments based on
each superpixel, which is further aligned by PatchMatch. The method combines
the efficiency of optical flow and the accuracy of PatchMatch. After PatchMatch
correction, we obtain a fully aligned image stack that facilitates a
high-quality fusion that is free from blurring/ghosting artifacts. We compare
our method with existing fusion algorithms on various challenging examples,
including the static/dynamic, the indoor/outdoor and the daytime/nighttime
scenes. Experiment results demonstrate the effectiveness and robustness of our
method
Nonlinear fatigue damage of cracked cement paste after grouting enhancement
Grouting reinforcement is an important part of modern engineering and has grown in popularity due to the benefits of grouting enhancement on cyclic loading. Understanding the fatigue mechanism of grouting-enhanced structures is vital to the design and the long-term stability analysis of such structures. In this study, the fatigue mechanical properties of cracked cement paste after epoxy resin grouting enhancement under different cyclic conditions were investigated in the laboratory and an inverted S-shaped curve was proposed to describe the damage accumulation. The test results indicated that the fatigue axial deformation can be divided into three stages: the initial stage, constant velocity stage and accelerating stage. The irreversible deformation can be used to describe the damage accumulation. The fatigue process is significantly affected by the upper limit stress level and the stress amplitude. In addition, the exponential relationship between stress amplitude and fatigue life was obtained. The proposed S-shaped curve was validated by an experimental fatigue strain test. The tests result upon various load conditions and crack types represented a good agreement with the predicted data
GyroFlow+: Gyroscope-Guided Unsupervised Deep Homography and Optical Flow Learning
Existing homography and optical flow methods are erroneous in challenging
scenes, such as fog, rain, night, and snow because the basic assumptions such
as brightness and gradient constancy are broken. To address this issue, we
present an unsupervised learning approach that fuses gyroscope into homography
and optical flow learning. Specifically, we first convert gyroscope readings
into motion fields named gyro field. Second, we design a self-guided fusion
module (SGF) to fuse the background motion extracted from the gyro field with
the optical flow and guide the network to focus on motion details. Meanwhile,
we propose a homography decoder module (HD) to combine gyro field and
intermediate results of SGF to produce the homography. To the best of our
knowledge, this is the first deep learning framework that fuses gyroscope data
and image content for both deep homography and optical flow learning. To
validate our method, we propose a new dataset that covers regular and
challenging scenes. Experiments show that our method outperforms the
state-of-the-art methods in both regular and challenging scenes.Comment: 12 pages. arXiv admin note: substantial text overlap with
arXiv:2103.1372
Fully Test-Time Adaptation for Monocular 3D Object Detection
Monocular 3D object detection (Mono 3Det) aims to identify 3D objects from a
single RGB image. However, existing methods often assume training and test data
follow the same distribution, which may not hold in real-world test scenarios.
To address the out-of-distribution (OOD) problems, we explore a new adaptation
paradigm for Mono 3Det, termed Fully Test-time Adaptation. It aims to adapt a
well-trained model to unlabeled test data by handling potential data
distribution shifts at test time without access to training data and test
labels. However, applying this paradigm in Mono 3Det poses significant
challenges due to OOD test data causing a remarkable decline in object
detection scores. This decline conflicts with the pre-defined score thresholds
of existing detection methods, leading to severe object omissions (i.e., rare
positive detections and many false negatives). Consequently, the limited
positive detection and plenty of noisy predictions cause test-time adaptation
to fail in Mono 3Det. To handle this problem, we propose a novel Monocular
Test-Time Adaptation (MonoTTA) method, based on two new strategies. 1)
Reliability-driven adaptation: we empirically find that high-score objects are
still reliable and the optimization of high-score objects can enhance
confidence across all detections. Thus, we devise a self-adaptive strategy to
identify reliable objects for model adaptation, which discovers potential
objects and alleviates omissions. 2) Noise-guard adaptation: since high-score
objects may be scarce, we develop a negative regularization term to exploit the
numerous low-score objects via negative learning, preventing overfitting to
noise and trivial solutions. Experimental results show that MonoTTA brings
significant performance gains for Mono 3Det models in OOD test scenarios,
approximately 190% gains by average on KITTI and 198% gains on nuScenes
HandBooster: Boosting 3D Hand-Mesh Reconstruction by Conditional Synthesis and Sampling of Hand-Object Interactions
Reconstructing 3D hand mesh robustly from a single image is very challenging,
due to the lack of diversity in existing real-world datasets. While data
synthesis helps relieve the issue, the syn-to-real gap still hinders its usage.
In this work, we present HandBooster, a new approach to uplift the data
diversity and boost the 3D hand-mesh reconstruction performance by training a
conditional generative space on hand-object interactions and purposely sampling
the space to synthesize effective data samples. First, we construct versatile
content-aware conditions to guide a diffusion model to produce realistic images
with diverse hand appearances, poses, views, and backgrounds; favorably,
accurate 3D annotations are obtained for free. Then, we design a novel
condition creator based on our similarity-aware distribution sampling
strategies to deliberately find novel and realistic interaction poses that are
distinctive from the training set. Equipped with our method, several baselines
can be significantly improved beyond the SOTA on the HO3D and DexYCB
benchmarks. Our code will be released on
https://github.com/hxwork/HandBooster_Pytorch
Supervised Homography Learning with Realistic Dataset Generation
In this paper, we propose an iterative framework, which consists of two
phases: a generation phase and a training phase, to generate realistic training
data and yield a supervised homography network. In the generation phase, given
an unlabeled image pair, we utilize the pre-estimated dominant plane masks and
homography of the pair, along with another sampled homography that serves as
ground truth to generate a new labeled training pair with realistic motion. In
the training phase, the generated data is used to train the supervised
homography network, in which the training data is refined via a content
consistency module and a quality assessment module. Once an iteration is
finished, the trained network is used in the next data generation phase to
update the pre-estimated homography. Through such an iterative strategy, the
quality of the dataset and the performance of the network can be gradually and
simultaneously improved. Experimental results show that our method achieves
state-of-the-art performance and existing supervised methods can be also
improved based on the generated dataset. Code and dataset are available at
https://github.com/megvii-research/RealSH.Comment: Accepted by ICCV 202
- …
