2 research outputs found
BoostTree and BoostForest for Ensemble Learning
Bootstrap aggregating (Bagging) and boosting are two popular ensemble
learning approaches, which combine multiple base learners to generate a
composite model for more accurate and more reliable performance. They have been
widely used in biology, engineering, healthcare, etc. This article proposes
BoostForest, which is an ensemble learning approach using BoostTree as base
learners and can be used for both classification and regression. BoostTree
constructs a tree model by gradient boosting. It achieves high randomness
(diversity) by sampling its parameters randomly from a parameter pool, and
selecting a subset of features randomly at node splitting. BoostForest further
increases the randomness by bootstrapping the training data in constructing
different BoostTrees. BoostForest outperformed four classical ensemble learning
approaches (Random Forest, Extra-Trees, XGBoost and LightGBM) on 34
classification and regression datasets. Remarkably, BoostForest has only one
hyper-parameter (the number of BoostTrees), which can be easily specified. Our
code is publicly available, and the proposed ensemble learning framework can
also be used to combine many other base learners
Towards Generalizable Deepfake Detection by Primary Region Regularization
The existing deepfake detection methods have reached a bottleneck in
generalizing to unseen forgeries and manipulation approaches. Based on the
observation that the deepfake detectors exhibit a preference for overfitting
the specific primary regions in input, this paper enhances the generalization
capability from a novel regularization perspective. This can be simply achieved
by augmenting the images through primary region removal, thereby preventing the
detector from over-relying on data bias. Our method consists of two stages,
namely the static localization for primary region maps, as well as the dynamic
exploitation of primary region masks. The proposed method can be seamlessly
integrated into different backbones without affecting their inference
efficiency. We conduct extensive experiments over three widely used deepfake
datasets - DFDC, DF-1.0, and Celeb-DF with five backbones. Our method
demonstrates an average performance improvement of 6% across different
backbones and performs competitively with several state-of-the-art baselines.Comment: 12 pages. Code and Dataset: https://github.com/xaCheng1996/PRL