304 research outputs found
Bilingual sentence alignment of pre-Qin history literature for digital humanities study
Sentence aligned bilingual text of history literature provides support of digital resources for related digital humanities studies, but existing studies have done little work on sentence alignment of ancient Chinese and English. In this study, we made a preliminary attempt to align the sentence of ancient Chinese and English. We used the bilingual text of the Analects of Confucius and Zuo's Commentaries of the Spring and Autumn Annals, extracted features and adopted the classification method to divide the bilingual candidate sentence pairs based on probability scores. The bilingual sentence alignment model based on SVM had the best performance on a larger amount of data when using three features and confirmed the impact of candidate dataset
DiffLLE: Diffusion-guided Domain Calibration for Unsupervised Low-light Image Enhancement
Existing unsupervised low-light image enhancement methods lack enough
effectiveness and generalization in practical applications. We suppose this is
because of the absence of explicit supervision and the inherent gap between
real-world scenarios and the training data domain. In this paper, we develop
Diffusion-based domain calibration to realize more robust and effective
unsupervised Low-Light Enhancement, called DiffLLE. Since the diffusion model
performs impressive denoising capability and has been trained on massive clean
images, we adopt it to bridge the gap between the real low-light domain and
training degradation domain, while providing efficient priors of real-world
content for unsupervised models. Specifically, we adopt a naive unsupervised
enhancement algorithm to realize preliminary restoration and design two
zero-shot plug-and-play modules based on diffusion model to improve
generalization and effectiveness. The Diffusion-guided Degradation Calibration
(DDC) module narrows the gap between real-world and training low-light
degradation through diffusion-based domain calibration and a lightness
enhancement curve, which makes the enhancement model perform robustly even in
sophisticated wild degradation. Due to the limited enhancement effect of the
unsupervised model, we further develop the Fine-grained Target domain
Distillation (FTD) module to find a more visual-friendly solution space. It
exploits the priors of the pre-trained diffusion model to generate
pseudo-references, which shrinks the preliminary restored results from a coarse
normal-light domain to a finer high-quality clean field, addressing the lack of
strong explicit supervision for unsupervised methods. Benefiting from these,
our approach even outperforms some supervised methods by using only a simple
unsupervised baseline. Extensive experiments demonstrate the superior
effectiveness of the proposed DiffLLE
Neural Video Fields Editing
Diffusion models have revolutionized text-driven video editing. However,
applying these methods to real-world editing encounters two significant
challenges: (1) the rapid increase in GPU memory demand as the number of frames
grows, and (2) the inter-frame inconsistency in edited videos. To this end, we
propose NVEdit, a novel text-driven video editing framework designed to
mitigate memory overhead and improve consistent editing for real-world long
videos. Specifically, we construct a neural video field, powered by tri-plane
and sparse grid, to enable encoding long videos with hundreds of frames in a
memory-efficient manner. Next, we update the video field through off-the-shelf
Text-to-Image (T2I) models to impart text-driven editing effects. A progressive
optimization strategy is developed to preserve original temporal priors.
Importantly, both the neural video field and T2I model are adaptable and
replaceable, thus inspiring future research. Experiments demonstrate the
ability of our approach to edit hundreds of frames with impressive inter-frame
consistency. Our project is available at: https://nvedit.github.io/
The lithospheric lager and block velocity structure and motion of Substance for Tibetan plateau
Abstract HKT-ISTP 2013
A
Higher visceral adiposity index is associated with increased likelihood of abdominal aortic calcification
Background: The negative effects of visceral adiposity accumulation on cardiovascular health have drawn much attention. However, the association between the Visceral Adiposity Index (VAI) and Abdominal Aortic Calcification (AAC) has never been reported before. The authors aimed to investigate the association between the VAI and AAC in US adults.
Methods: Cross-sectional data were derived from the 2013 to 2014 National Health and Nutrition Examination Survey (NHANES) of participants with complete data of VAI and AAC scores. Weighted multivariable regression and logistic regression analysis were conducted to explore the independent relationship between VAI and AAC. Subgroup analysis and interaction tests were also performed.
Results: A total of 2958 participants were enrolled and participants in the higher VAI tertile tended to have a higher mean AAC score and prevalence of severe AAC. In the fully adjusted model, a positive association between VAI and AAC score and severe AAC was observed (β = 0.04, 95% CI 0.01‒0.08; OR = 1.04, 95% CI 1.01‒1.07). Participants in the highest VAI tertile had a 0.41-unit higher AAC score (β = 0.41, 95% CI 0.08‒0.73) and a significantly 68% higher risk of severe AAC than those in the lowest VAI tertile (OR = 1.68, 95% CI 1.04‒2.71). Subgroup analysis and interaction tests indicated that there was no dependence for the association of VAI and AAC.
Conclusion: Visceral adiposity accumulation evaluated by the VAI was associated with a higher AAC score and an increased likelihood of severe AAC
FLAG3D: A 3D Fitness Activity Dataset with Language Instruction
With the continuously thriving popularity around the world, fitness activity
analytic has become an emerging research topic in computer vision. While a
variety of new tasks and algorithms have been proposed recently, there are
growing hunger for data resources involved in high-quality data, fine-grained
labels, and diverse environments. In this paper, we present FLAG3D, a
large-scale 3D fitness activity dataset with language instruction containing
180K sequences of 60 categories. FLAG3D features the following three aspects:
1) accurate and dense 3D human pose captured from advanced MoCap system to
handle the complex activity and large movement, 2) detailed and professional
language instruction to describe how to perform a specific activity, 3)
versatile video resources from a high-tech MoCap system, rendering software,
and cost-effective smartphones in natural environments. Extensive experiments
and in-depth analysis show that FLAG3D contributes great research value for
various challenges, such as cross-domain human action recognition, dynamic
human mesh recovery, and language-guided human action generation. Our dataset
and source code will be publicly available at
https://andytang15.github.io/FLAG3D
Federated Learning of Large Language Models with Parameter-Efficient Prompt Tuning and Adaptive Optimization
Federated learning (FL) is a promising paradigm to enable collaborative model
training with decentralized data. However, the training process of Large
Language Models (LLMs) generally incurs the update of significant parameters,
which limits the applicability of FL techniques to tackle the LLMs in real
scenarios. Prompt tuning can significantly reduce the number of parameters to
update, but it either incurs performance degradation or low training
efficiency. The straightforward utilization of prompt tuning in the FL often
raises non-trivial communication costs and dramatically degrades performance.
In addition, the decentralized data is generally non-Independent and
Identically Distributed (non-IID), which brings client drift problems and thus
poor performance. This paper proposes a Parameter-efficient prompt Tuning
approach with Adaptive Optimization, i.e., FedPepTAO, to enable efficient and
effective FL of LLMs. First, an efficient partial prompt tuning approach is
proposed to improve performance and efficiency simultaneously. Second, a novel
adaptive optimization method is developed to address the client drift problems
on both the device and server sides to enhance performance further. Extensive
experiments based on 10 datasets demonstrate the superb performance (up to
60.8\% in terms of accuracy) and efficiency (up to 97.59\% in terms of training
time) of FedPepTAO compared with 9 baseline approaches. Our code is available
at https://github.com/llm-eff/FedPepTAO.Comment: 18 pages, accepted by EMNLP 202
- …