475 research outputs found
Semi-supervised Text Regression with Conditional Generative Adversarial Networks
Enormous online textual information provides intriguing opportunities for
understandings of social and economic semantics. In this paper, we propose a
novel text regression model based on a conditional generative adversarial
network (GAN), with an attempt to associate textual data and social outcomes in
a semi-supervised manner. Besides promising potential of predicting
capabilities, our superiorities are twofold: (i) the model works with
unbalanced datasets of limited labelled data, which align with real-world
scenarios; and (ii) predictions are obtained by an end-to-end framework,
without explicitly selecting high-level representations. Finally we point out
related datasets for experiments and future research directions
Festivals, Festival Foods, and Dietary Acculturation: A Journey of Hybridization and Identity Formation for Chinese International Students in Ottawa
Through participant observation at the 2018 Ottawa Night Market Chinatown and interviews with fifteen post-secondary Chinese national students in Ottawa about their dietary acculturation, this research aims to answer the following questions: How does hybridity play out in Chinese students' dietary acculturation? What are the impacts of festivals and festival foods on hybridization and identity formation? The findings suggest that Chinese students do become more "hybrid" in their food practices, but this is less so from incorporating Canadian food habits, and more a result of increased consumption of various Chinese regional cuisines and Asian cuisines. However, becoming more hybrid does not weaken the participants' Chinese identity; rather they retain it through attending the Night Market, celebrating traditional Chinese festivals, and maintaining cultural beliefs related to food choices, health and nutrition. This study suggests that hybridization involves multi-cultural and multi-dimensional influences, and confirms that hybridization is distinct from identity formation
Research on Insulator Detection Algorithm for High-Speed Rail Contact Network
To satisfy the requirements of intelligent inspection at a greater level, a rotated insulator target detection algorithm based on the improved YOLOv5 is proposed to solve the inadequacy of conventional detection algorithms used for contact networks in high-speed rails, e.g., low accuracy and non-consideration of insulation direction. First, Coordinated Attention (CA) and criss-cross attention mechanisms are introduced to efficiently extract the effective features and position information of insulators. The Reparameterization Visual Geometry Group (RepVGG) backbone network architecture is used to effectively improve the model representation and detection speed. In the backbone network of the detection head, the Alignment Convolution (AC) module is used to solve the tilt and feature misalignment of the insulator target, as well as to adjust the alignment degree of the prediction frame to the actual target. Finally, the Rotation Complete Intersection over Union (R-CIoU) is used to calculate the rotation loss function, which can be used to accurately position the prediction frame. Experimental results show that the proposed algorithm can detect different directions of insulators and that the mean Average Precision (mAP) can reach 97.5% while the detection speed is improved, thus satisfying the requirements of insulator target detection at a greater level
EdgeYOLO: An Edge-Real-Time Object Detector
This paper proposes an efficient, low-complexity and anchor-free object
detector based on the state-of-the-art YOLO framework, which can be implemented
in real time on edge computing platforms. We develop an enhanced data
augmentation method to effectively suppress overfitting during training, and
design a hybrid random loss function to improve the detection accuracy of small
objects. Inspired by FCOS, a lighter and more efficient decoupled head is
proposed, and its inference speed can be improved with little loss of
precision. Our baseline model can reach the accuracy of 50.6% AP50:95 and 69.8%
AP50 in MS COCO2017 dataset, 26.4% AP50:95 and 44.8% AP50 in VisDrone2019-DET
dataset, and it meets real-time requirements (FPS>=30) on edge-computing device
Nvidia Jetson AGX Xavier. We also designed lighter models with less parameters
for edge computing devices with lower computing power, which also show better
performances. Our source code, hyper-parameters and model weights are all
available at https://github.com/LSH9832/edgeyolo
Gitor: Scalable Code Clone Detection by Building Global Sample Graph
Code clone detection is about finding out similar code fragments, which has
drawn much attention in software engineering since it is important for software
maintenance and evolution. Researchers have proposed many techniques and tools
for source code clone detection, but current detection methods concentrate on
analyzing or processing code samples individually without exploring the
underlying connections among code samples. In this paper, we propose Gitor to
capture the underlying connections among different code samples. Specifically,
given a source code database, we first tokenize all code samples to extract the
pre-defined individual information. After obtaining all samples individual
information, we leverage them to build a large global sample graph where each
node is a code sample or a type of individual information. Then we apply a node
embedding technique on the global sample graph to extract all the samples
vector representations. After collecting all code samples vectors, we can
simply compare the similarity between any two samples to detect possible clone
pairs. More importantly, since the obtained vector of a sample is from a global
sample graph, we can combine it with its own code features to improve the code
clone detection performance. To demonstrate the effectiveness of Gitor, we
evaluate it on a widely used dataset namely BigCloneBench. Our experimental
results show that Gitor has higher accuracy in terms of code clone detection
and excellent execution time for inputs of various sizes compared to existing
state-of-the-art tools. Moreover, we also evaluate the combination of Gitor
with other traditional vector-based clone detection methods, the results show
that the use of Gitor enables them detect more code clones with higher F1.Comment: 12 pages, 5 figure
CC2Vec: Combining Typed Tokens with Contrastive Learning for Effective Code Clone Detection
With the development of the open source community, the code is often copied,
spread, and evolved in multiple software systems, which brings uncertainty and
risk to the software system (e.g., bug propagation and copyright infringement).
Therefore, it is important to conduct code clone detection to discover similar
code pairs. Many approaches have been proposed to detect code clones where
token-based tools can scale to big code. However, due to the lack of program
details, they cannot handle more complicated code clones, i.e., semantic code
clones. In this paper, we introduce CC2Vec, a novel code encoding method
designed to swiftly identify simple code clones while also enhancing the
capability for semantic code clone detection. To retain the program details
between tokens, CC2Vec divides them into different categories (i.e., typed
tokens) according to the syntactic types and then applies two self-attention
mechanism layers to encode them. To resist changes in the code structure of
semantic code clones, CC2Vec performs contrastive learning to reduce the
differences introduced by different code implementations. We evaluate CC2Vec on
two widely used datasets (i.e., BigCloneBench and Google Code Jam) and the
results report that our method can effectively detect simple code clones. In
addition, CC2Vec not only attains comparable performance to widely used
semantic code clone detection systems such as ASTNN, SCDetector, and FCCA by
simply fine-tuning, but also significantly surpasses these methods in both
detection efficiency.Comment: 21 pages, 7 figure
Fast and accurate extraction of ultra-high quality factor from cavity ring-down measurement
Cavity ring-down is an essential test to measure ultra-high quality factor
(UHQ) optical cavities, which is, however, frequently misinterpreted due to
lacking of a specified analysis guideline. Here we clarify the basic property
of cavity ring down and present a step-by-step method that enables extraction
of the overall quality factor, as well as the intrinsic loss and coupling state
of UHQ cavities with better fidelity and simplicity than prior schemes. Our
work can facilitate acurrate design and characterization of UHQ cavities for
ultra-low noise lasers, high finesse reference cavities, and ultra-narrow
optical filters
Semi-supervised Text Regression with Conditional Generative Adversarial Networks
Enormous online textual information provides intriguing opportunities for understandings of social and economic semantics. In this paper, we propose a novel text regression model based on a conditional generative adversarial network (GAN), with an attempt to associate textual data and social outcomes in a semi-supervised manner. Besides promising potential of predicting capabilities, our superiorities are twofold: (i) the model works with unbalanced datasets of limited labelled data, which align with real-world scenarios; and (ii) predictions are obtained by an end-to-end framework, without explicitly selecting high-level representations. Finally we point out related datasets for experiments and future research directions
A Community Detection and Graph Neural Network Based Link Prediction Approach for Scientific Literature
This study presents a novel approach that synergizes community detection
algorithms with various Graph Neural Network (GNN) models to bolster link
prediction in scientific literature networks. By integrating the Louvain
community detection algorithm into our GNN frameworks, we consistently enhance
performance across all models tested. For example, integrating Louvain with the
GAT model resulted in an AUC score increase from 0.777 to 0.823, exemplifying
the typical improvements observed. Similar gains are noted when Louvain is
paired with other GNN architectures, confirming the robustness and
effectiveness of incorporating community-level insights. This consistent uplift
in performance reflected in our extensive experimentation on bipartite graphs
of scientific collaborations and citations highlights the synergistic potential
of combining community detection with GNNs to overcome common link prediction
challenges such as scalability and resolution limits. Our findings advocate for
the integration of community structures as a significant step forward in the
predictive accuracy of network science models, offering a comprehensive
understanding of scientific collaboration patterns through the lens of advanced
machine learning techniques
Development of risk prediction model for cognitive impairment in patients with coronary heart disease: A study protocol for a prospective, cross-sectional analysis
BackgroundIschemic heart disease and degenerative encephalopathy are two main sources of disease burden for the global elderly population. Coronary heart disease (CHD) and cognitive impairment, as representative diseases, are prevalent and serious illnesses in the elderly. According to recent research, patients with CHD are more likely to experience cognitive impairment and their cognitive ability declines more quickly. Vascular risk factors have been associated with differences in cognitive performance in epidemiological studies, but evidence in patients with CHD is more limited. Inextricably linked between the heart and the brain. Considering the unique characteristics of recurrent cognitive impairment in patients with CHD, we will further study the related risk factors. We tried to investigate the potential predictors of cognitive impairment in patients with CHD through a prospective, cross-sectional study.MethodsThe cross-sectional study design will recruit 378 patients with CHD (≥65 years) from Xiyuan Hospital of China Academy of Chinese Medical Sciences. The subjects' cognitive function is evaluated with MoCA scale, and they are divided into cognitive impairment group and normal cognitive function group according to the score results. Demographic data, disease characteristics (results of coronary CT/ angiography, number of stents implanted, status of diseased vessels), laboratory tests (biochemistry, coagulation, serum iron levels, pulse wave velocity), metabolites (blood samples and intestinal metabolites), and lifestyle (smoking, alcohol consumption, sleep, physical activity) will be assessed as outcome indicators. Compare the two groups and the correlation analysis will be performed on the development of mild cognitive impairment. Mann-Whitney U or X2 test was selected to describe and evaluate the variation, and logistics regression analysis was employed to fit the prediction model. After that, do the calibration curve and decision curve to evaluate the model. The prediction model will be validated by a validation set.DiscussionTo explore the risk factors related to mild cognitive impairment (MCI) in patients with CHD, a new predictive model is established, which can achieve advanced intervention in the occurrence of MCI after CHD. Owing to its cross-sectional study design, the study has some limitations, but it will be further studied by increasing the observation period, adding follow-up data collection or prospective cohort study. The study has been registered with the China Clinical Trials Registry (ChiCTR2200063255) to conduct clinical trials
- …
