87 research outputs found
An informatics system for exploring eye movements in reading
Eye tracking techniques have been widely used in many research areas including
cognitive science, psychology, human-computer interaction, marketing research,
medical research etc. Many computer programs have emerged to help these
researchers to design experiments, present visual stimuli and process the large
quantity of numerical data produced by the eye tracker. However, most applications,
especially commercial products, are designed for a particular tracking device
and tend to be general purpose. Few of them are designed specifically for
reading research. This can be inconvenient when dealing with complex experimental
design, multi-source data collection, and text based data analysis, including
almost every aspect of a reading study lifecycle.
A flexible and powerful system that manages the lifecycle of different reading
studies is required to fulfill these demands. Therefore, we created an informatics
system with two major software suites: Experiment Executor and EyeMap. It
is a system designed specifically for reading research. Experiment Executor
helps reading researchers build complex experimental environments, which can
rapidly present display changes and support the co-registration of eye tracking
information with other data collection devices such as EEG (electroencephalography)
amplifiers. The EyeMap component helps researchers visualize and analysis
a wide range of writing systems including spaced and unspaced scripts, which
can be presented in proportional or non-proportional font types. The aim of the
system is to accelerate the life cycle of a reading experiment from design through
analysis.
Several experiments were conducted on this system. These experiments
confirmed the effectiveness and the capability of the system with several new
reading research findings from the visual information processing stages of reading
SGL-PT: A Strong Graph Learner with Graph Prompt Tuning
Recently, much exertion has been paid to design graph self-supervised methods
to obtain generalized pre-trained models, and adapt pre-trained models onto
downstream tasks through fine-tuning. However, there exists an inherent gap
between pretext and downstream graph tasks, which insufficiently exerts the
ability of pre-trained models and even leads to negative transfer. Meanwhile,
prompt tuning has seen emerging success in natural language processing by
aligning pre-training and fine-tuning with consistent training objectives. In
this paper, we identify the challenges for graph prompt tuning: The first is
the lack of a strong and universal pre-training task across sundry pre-training
methods in graph domain. The second challenge lies in the difficulty of
designing a consistent training objective for both pre-training and downstream
tasks. To overcome above obstacles, we propose a novel framework named SGL-PT
which follows the learning strategy ``Pre-train, Prompt, and Predict''.
Specifically, we raise a strong and universal pre-training task coined as SGL
that acquires the complementary merits of generative and contrastive
self-supervised graph learning. And aiming for graph classification task, we
unify pre-training and fine-tuning by designing a novel verbalizer-free
prompting function, which reformulates the downstream task in a similar format
as pretext task. Empirical results show that our method surpasses other
baselines under unsupervised setting, and our prompt tuning method can greatly
facilitate models on biological datasets over fine-tuning methods
MARIO: Model Agnostic Recipe for Improving OOD Generalization of Graph Contrastive Learning
In this work, we investigate the problem of out-of-distribution (OOD)
generalization for unsupervised learning methods on graph data. This scenario
is particularly challenging because graph neural networks (GNNs) have been
shown to be sensitive to distributional shifts, even when labels are available.
To address this challenge, we propose a \underline{M}odel-\underline{A}gnostic
\underline{R}ecipe for \underline{I}mproving \underline{O}OD generalizability
of unsupervised graph contrastive learning methods, which we refer to as MARIO.
MARIO introduces two principles aimed at developing distributional-shift-robust
graph contrastive methods to overcome the limitations of existing frameworks:
(i) Information Bottleneck (IB) principle for achieving generalizable
representations and (ii) Invariant principle that incorporates adversarial data
augmentation to obtain invariant representations. To the best of our knowledge,
this is the first work that investigates the OOD generalization problem of
graph contrastive learning, with a specific focus on node-level tasks. Through
extensive experiments, we demonstrate that our method achieves state-of-the-art
performance on the OOD test set, while maintaining comparable performance on
the in-distribution test set when compared to existing approaches. The source
code for our method can be found at: https://github.com/ZhuYun97/MARIOComment: 20 pages, 15 figure
Study on Helicopter Antitorque Device Based on Cross-Flow Fan Technology
In order to improve low-altitude flight security of single-rotor helicopter, an experimental model of a helicopter antitorque device is developed for wind tunnel test. The model is based on the flow control technology of the cross-flow fan (CFF). Wind tunnel tests show that the model can produce side force. It is concluded that the influence of the CFF rotating speed, the rotor collective pitch, and the forward flight speed on the side force of the model is great. At the same time, the numerical simulation calculation method of the model has been established. Good agreement between experimental and numerical side force and power shows that results of numerical solution are reliable. Therefore, the results in actual helicopter obtained from Computational Fluid Dynamics (CFD) solution are acceptable. This proves that this antitorque device can be used for a helicopter
Cross-relation Cross-bag Attention for Distantly-supervised Relation Extraction
Distant supervision leverages knowledge bases to automatically label
instances, thus allowing us to train relation extractor without human
annotations. However, the generated training data typically contain massive
noise, and may result in poor performances with the vanilla supervised
learning. In this paper, we propose to conduct multi-instance learning with a
novel Cross-relation Cross-bag Selective Attention (CSA), which leads to
noise-robust training for distant supervised relation extractor. Specifically,
we employ the sentence-level selective attention to reduce the effect of noisy
or mismatched sentences, while the correlation among relations were captured to
improve the quality of attention weights. Moreover, instead of treating all
entity-pairs equally, we try to pay more attention to entity-pairs with a
higher quality. Similarly, we adopt the selective attention mechanism to
achieve this goal. Experiments with two types of relation extractor demonstrate
the superiority of the proposed approach over the state-of-the-art, while
further ablation studies verify our intuitions and demonstrate the
effectiveness of our proposed two techniques.Comment: AAAI 201
Dancing Avatar: Pose and Text-Guided Human Motion Videos Synthesis with Image Diffusion Model
The rising demand for creating lifelike avatars in the digital realm has led
to an increased need for generating high-quality human videos guided by textual
descriptions and poses. We propose Dancing Avatar, designed to fabricate human
motion videos driven by poses and textual cues. Our approach employs a
pretrained T2I diffusion model to generate each video frame in an
autoregressive fashion. The crux of innovation lies in our adept utilization of
the T2I diffusion model for producing video frames successively while
preserving contextual relevance. We surmount the hurdles posed by maintaining
human character and clothing consistency across varying poses, along with
upholding the background's continuity amidst diverse human movements. To ensure
consistent human appearances across the entire video, we devise an intra-frame
alignment module. This module assimilates text-guided synthesized human
character knowledge into the pretrained T2I diffusion model, synergizing
insights from ChatGPT. For preserving background continuity, we put forth a
background alignment pipeline, amalgamating insights from segment anything and
image inpainting techniques. Furthermore, we propose an inter-frame alignment
module that draws inspiration from an auto-regressive pipeline to augment
temporal consistency between adjacent frames, where the preceding frame guides
the synthesis process of the current frame. Comparisons with state-of-the-art
methods demonstrate that Dancing Avatar exhibits the capacity to generate human
videos with markedly superior quality, both in terms of human and background
fidelity, as well as temporal coherence compared to existing state-of-the-art
approaches.Comment: 11 pages, 3 figure
Degeneration-Tuning: Using Scrambled Grid shield Unwanted Concepts from Stable Diffusion
Owing to the unrestricted nature of the content in the training data, large
text-to-image diffusion models, such as Stable Diffusion (SD), are capable of
generating images with potentially copyrighted or dangerous content based on
corresponding textual concepts information. This includes specific intellectual
property (IP), human faces, and various artistic styles. However, Negative
Prompt, a widely used method for content removal, frequently fails to conceal
this content due to inherent limitations in its inference logic. In this work,
we propose a novel strategy named \textbf{Degeneration-Tuning (DT)} to shield
contents of unwanted concepts from SD weights. By utilizing Scrambled Grid to
reconstruct the correlation between undesired concepts and their corresponding
image domain, we guide SD to generate meaningless content when such textual
concepts are provided as input. As this adaptation occurs at the level of the
model's weights, the SD, after DT, can be grafted onto other conditional
diffusion frameworks like ControlNet to shield unwanted concepts. In addition
to qualitatively showcasing the effectiveness of our DT method in protecting
various types of concepts, a quantitative comparison of the SD before and after
DT indicates that the DT method does not significantly impact the generative
quality of other contents. The FID and IS scores of the model on COCO-30K
exhibit only minor changes after DT, shifting from 12.61 and 39.20 to 13.04 and
38.25, respectively, which clearly outperforms the previous methods
- …