157 research outputs found
Factors that affect Chinese teachers’ use of the L1 and TL at tertiary level: an investigation from sociolinguistic perspective
Communicative Language Teaching (CLT) has been dominantly advocated in various educational contexts for many years. A number of countries promoted CLT in their English curricula, and the development of students’ communicative competence has become increasingly important in the era of globalisation. It is because people from different areas in the world are increasingly interconnected and communicative competence entails international sensitivities to communication needs of global citizens. Emergence of English as a global language and the changing situation of English learning have been acknowledged globally. However, there appears to be increasing resistances against the implementation of CLT in countries like China, Japan and Vietnam. This study is a timely research that revisits English Language Teaching (ELT) in China with the focus on Chinese teachers’ use of the Target Language (TL) and First Language (L1). The purpose is to have an in-depth look at the factors that affect Chinese teachers’ use of the TL and L1 in College English classes.
The current study took place in a regional university in China. The research methods employed included 53 Classroom Observations, 4 teachers’ interviews and 4 students’ focus-group interviews, and document analysis. The findings suggest multiple resistances existing in the current research site and call for changes that should be made from different dimensions within the context
Non-Autoregressive Neural Machine Translation with Enhanced Decoder Input
Non-autoregressive translation (NAT) models, which remove the dependence on
previous target tokens from the inputs of the decoder, achieve significantly
inference speedup but at the cost of inferior accuracy compared to
autoregressive translation (AT) models. Previous work shows that the quality of
the inputs of the decoder is important and largely impacts the model accuracy.
In this paper, we propose two methods to enhance the decoder inputs so as to
improve NAT models. The first one directly leverages a phrase table generated
by conventional SMT approaches to translate source tokens to target tokens,
which are then fed into the decoder as inputs. The second one transforms
source-side word embeddings to target-side word embeddings through
sentence-level alignment and word-level adversary learning, and then feeds the
transformed word embeddings into the decoder as inputs. Experimental results
show our method largely outperforms the NAT baseline~\citep{gu2017non} by
BLEU scores on WMT14 English-German task and BLEU scores on WMT16
English-Romanian task.Comment: AAAI 201
Efficient Bi-Level Optimization for Recommendation Denoising
The acquisition of explicit user feedback (e.g., ratings) in real-world
recommender systems is often hindered by the need for active user involvement.
To mitigate this issue, implicit feedback (e.g., clicks) generated during user
browsing is exploited as a viable substitute. However, implicit feedback
possesses a high degree of noise, which significantly undermines recommendation
quality. While many methods have been proposed to address this issue by
assigning varying weights to implicit feedback, two shortcomings persist: (1)
the weight calculation in these methods is iteration-independent, without
considering the influence of weights in previous iterations, and (2) the weight
calculation often relies on prior knowledge, which may not always be readily
available or universally applicable.
To overcome these two limitations, we model recommendation denoising as a
bi-level optimization problem. The inner optimization aims to derive an
effective model for the recommendation, as well as guiding the weight
determination, thereby eliminating the need for prior knowledge. The outer
optimization leverages gradients of the inner optimization and adjusts the
weights in a manner considering the impact of previous weights. To efficiently
solve this bi-level optimization problem, we employ a weight generator to avoid
the storage of weights and a one-step gradient-matching-based loss to
significantly reduce computational time. The experimental results on three
benchmark datasets demonstrate that our proposed approach outperforms both
state-of-the-art general and denoising recommendation models. The code is
available at https://github.com/CoderWZW/BOD.Comment: 11pages, 5 figures, 6 table
Fine-Tuning by Curriculum Learning for Non-Autoregressive Neural Machine Translation
Non-autoregressive translation (NAT) models remove the dependence on previous
target tokens and generate all target tokens in parallel, resulting in
significant inference speedup but at the cost of inferior translation accuracy
compared to autoregressive translation (AT) models. Considering that AT models
have higher accuracy and are easier to train than NAT models, and both of them
share the same model configurations, a natural idea to improve the accuracy of
NAT models is to transfer a well-trained AT model to an NAT model through
fine-tuning. However, since AT and NAT models differ greatly in training
strategy, straightforward fine-tuning does not work well. In this work, we
introduce curriculum learning into fine-tuning for NAT. Specifically, we design
a curriculum in the fine-tuning process to progressively switch the training
from autoregressive generation to non-autoregressive generation. Experiments on
four benchmark translation datasets show that the proposed method achieves good
improvement (more than BLEU score) over previous NAT baselines in terms of
translation accuracy, and greatly speed up (more than times) the inference
process over AT baselines.Comment: AAAI 202
Complex Pathways to Cooperation Emergent from Asymmetry in Heterogeneous Populations
Cooperation within asymmetric populations has garnered significant attention
in evolutionary games. This paper explores cooperation evolution in populations
with weak and strong players, using a game model where players choose between
cooperation and defection. Asymmetry stems from different benefits for strong
and weak cooperators, with their benefit ratio indicating the degree of
asymmetry. Varied rankings of parameters including the asymmetry degree,
cooperation costs, and benefits brought by weak players give rise to scenarios
including the prisoner's dilemma (PDG) for both player types, the snowdrift
game (SDG), and mixed PDG-SDG interactions. Our results indicate that in an
infinite well-mixed population, defection remains the dominant strategy when
strong players engage in the prisoner's dilemma game. However, if strong
players play snowdrift games, global cooperation increases with the proportion
of strong players. In this scenario, strong cooperators can prevail over strong
defectors when the proportion of strong players is low, but the prevalence of
cooperation among strong players decreases as their proportion increases. In
contrast, within a square lattice, the optimum global cooperation emerges at
intermediate proportions of strong players with moderate degrees of asymmetry.
Additionally, weak players protect cooperative clusters from exploitation by
strong defectors. This study highlights the complex dynamics of cooperation in
asymmetric interactions, contributing to the theory of cooperation in
asymmetric games.Comment: 10 pages, 8 figure
UniEdit: A Unified Tuning-Free Framework for Video Motion and Appearance Editing
Recent advances in text-guided video editing have showcased promising results
in appearance editing (e.g., stylization). However, video motion editing in the
temporal dimension (e.g., from eating to waving), which distinguishes video
editing from image editing, is underexplored. In this work, we present UniEdit,
a tuning-free framework that supports both video motion and appearance editing
by harnessing the power of a pre-trained text-to-video generator within an
inversion-then-generation framework. To realize motion editing while preserving
source video content, based on the insights that temporal and spatial
self-attention layers encode inter-frame and intra-frame dependency
respectively, we introduce auxiliary motion-reference and reconstruction
branches to produce text-guided motion and source features respectively. The
obtained features are then injected into the main editing path via temporal and
spatial self-attention layers. Extensive experiments demonstrate that UniEdit
covers video motion editing and various appearance editing scenarios, and
surpasses the state-of-the-art methods. Our code will be publicly available.Comment: Project page: https://jianhongbai.github.io/UniEdit
A Survey on Non-Autoregressive Generation for Neural Machine Translation and Beyond
Non-autoregressive (NAR) generation, which is first proposed in neural
machine translation (NMT) to speed up inference, has attracted much attention
in both machine learning and natural language processing communities. While NAR
generation can significantly accelerate inference speed for machine
translation, the speedup comes at the cost of sacrificed translation accuracy
compared to its counterpart, auto-regressive (AR) generation. In recent years,
many new models and algorithms have been designed/proposed to bridge the
accuracy gap between NAR generation and AR generation. In this paper, we
conduct a systematic survey with comparisons and discussions of various
non-autoregressive translation (NAT) models from different aspects.
Specifically, we categorize the efforts of NAT into several groups, including
data manipulation, modeling methods, training criterion, decoding algorithms,
and the benefit from pre-trained models. Furthermore, we briefly review other
applications of NAR models beyond machine translation, such as dialogue
generation, text summarization, grammar error correction, semantic parsing,
speech synthesis, and automatic speech recognition. In addition, we also
discuss potential directions for future exploration, including releasing the
dependency of KD, dynamic length prediction, pre-training for NAR, and wider
applications, etc. We hope this survey can help researchers capture the latest
progress in NAR generation, inspire the design of advanced NAR models and
algorithms, and enable industry practitioners to choose appropriate solutions
for their applications. The web page of this survey is at
\url{https://github.com/LitterBrother-Xiao/Overview-of-Non-autoregressive-Applications}.Comment: 25 pages, 11 figures, 4 table
- …