157 research outputs found

    Factors that affect Chinese teachers’ use of the L1 and TL at tertiary level: an investigation from sociolinguistic perspective

    Get PDF
    Communicative Language Teaching (CLT) has been dominantly advocated in various educational contexts for many years. A number of countries promoted CLT in their English curricula, and the development of students’ communicative competence has become increasingly important in the era of globalisation. It is because people from different areas in the world are increasingly interconnected and communicative competence entails international sensitivities to communication needs of global citizens. Emergence of English as a global language and the changing situation of English learning have been acknowledged globally. However, there appears to be increasing resistances against the implementation of CLT in countries like China, Japan and Vietnam. This study is a timely research that revisits English Language Teaching (ELT) in China with the focus on Chinese teachers’ use of the Target Language (TL) and First Language (L1). The purpose is to have an in-depth look at the factors that affect Chinese teachers’ use of the TL and L1 in College English classes. The current study took place in a regional university in China. The research methods employed included 53 Classroom Observations, 4 teachers’ interviews and 4 students’ focus-group interviews, and document analysis. The findings suggest multiple resistances existing in the current research site and call for changes that should be made from different dimensions within the context

    Non-Autoregressive Neural Machine Translation with Enhanced Decoder Input

    Full text link
    Non-autoregressive translation (NAT) models, which remove the dependence on previous target tokens from the inputs of the decoder, achieve significantly inference speedup but at the cost of inferior accuracy compared to autoregressive translation (AT) models. Previous work shows that the quality of the inputs of the decoder is important and largely impacts the model accuracy. In this paper, we propose two methods to enhance the decoder inputs so as to improve NAT models. The first one directly leverages a phrase table generated by conventional SMT approaches to translate source tokens to target tokens, which are then fed into the decoder as inputs. The second one transforms source-side word embeddings to target-side word embeddings through sentence-level alignment and word-level adversary learning, and then feeds the transformed word embeddings into the decoder as inputs. Experimental results show our method largely outperforms the NAT baseline~\citep{gu2017non} by 5.115.11 BLEU scores on WMT14 English-German task and 4.724.72 BLEU scores on WMT16 English-Romanian task.Comment: AAAI 201

    Efficient Bi-Level Optimization for Recommendation Denoising

    Full text link
    The acquisition of explicit user feedback (e.g., ratings) in real-world recommender systems is often hindered by the need for active user involvement. To mitigate this issue, implicit feedback (e.g., clicks) generated during user browsing is exploited as a viable substitute. However, implicit feedback possesses a high degree of noise, which significantly undermines recommendation quality. While many methods have been proposed to address this issue by assigning varying weights to implicit feedback, two shortcomings persist: (1) the weight calculation in these methods is iteration-independent, without considering the influence of weights in previous iterations, and (2) the weight calculation often relies on prior knowledge, which may not always be readily available or universally applicable. To overcome these two limitations, we model recommendation denoising as a bi-level optimization problem. The inner optimization aims to derive an effective model for the recommendation, as well as guiding the weight determination, thereby eliminating the need for prior knowledge. The outer optimization leverages gradients of the inner optimization and adjusts the weights in a manner considering the impact of previous weights. To efficiently solve this bi-level optimization problem, we employ a weight generator to avoid the storage of weights and a one-step gradient-matching-based loss to significantly reduce computational time. The experimental results on three benchmark datasets demonstrate that our proposed approach outperforms both state-of-the-art general and denoising recommendation models. The code is available at https://github.com/CoderWZW/BOD.Comment: 11pages, 5 figures, 6 table

    Fine-Tuning by Curriculum Learning for Non-Autoregressive Neural Machine Translation

    Full text link
    Non-autoregressive translation (NAT) models remove the dependence on previous target tokens and generate all target tokens in parallel, resulting in significant inference speedup but at the cost of inferior translation accuracy compared to autoregressive translation (AT) models. Considering that AT models have higher accuracy and are easier to train than NAT models, and both of them share the same model configurations, a natural idea to improve the accuracy of NAT models is to transfer a well-trained AT model to an NAT model through fine-tuning. However, since AT and NAT models differ greatly in training strategy, straightforward fine-tuning does not work well. In this work, we introduce curriculum learning into fine-tuning for NAT. Specifically, we design a curriculum in the fine-tuning process to progressively switch the training from autoregressive generation to non-autoregressive generation. Experiments on four benchmark translation datasets show that the proposed method achieves good improvement (more than 11 BLEU score) over previous NAT baselines in terms of translation accuracy, and greatly speed up (more than 1010 times) the inference process over AT baselines.Comment: AAAI 202

    Complex Pathways to Cooperation Emergent from Asymmetry in Heterogeneous Populations

    Full text link
    Cooperation within asymmetric populations has garnered significant attention in evolutionary games. This paper explores cooperation evolution in populations with weak and strong players, using a game model where players choose between cooperation and defection. Asymmetry stems from different benefits for strong and weak cooperators, with their benefit ratio indicating the degree of asymmetry. Varied rankings of parameters including the asymmetry degree, cooperation costs, and benefits brought by weak players give rise to scenarios including the prisoner's dilemma (PDG) for both player types, the snowdrift game (SDG), and mixed PDG-SDG interactions. Our results indicate that in an infinite well-mixed population, defection remains the dominant strategy when strong players engage in the prisoner's dilemma game. However, if strong players play snowdrift games, global cooperation increases with the proportion of strong players. In this scenario, strong cooperators can prevail over strong defectors when the proportion of strong players is low, but the prevalence of cooperation among strong players decreases as their proportion increases. In contrast, within a square lattice, the optimum global cooperation emerges at intermediate proportions of strong players with moderate degrees of asymmetry. Additionally, weak players protect cooperative clusters from exploitation by strong defectors. This study highlights the complex dynamics of cooperation in asymmetric interactions, contributing to the theory of cooperation in asymmetric games.Comment: 10 pages, 8 figure

    UniEdit: A Unified Tuning-Free Framework for Video Motion and Appearance Editing

    Full text link
    Recent advances in text-guided video editing have showcased promising results in appearance editing (e.g., stylization). However, video motion editing in the temporal dimension (e.g., from eating to waving), which distinguishes video editing from image editing, is underexplored. In this work, we present UniEdit, a tuning-free framework that supports both video motion and appearance editing by harnessing the power of a pre-trained text-to-video generator within an inversion-then-generation framework. To realize motion editing while preserving source video content, based on the insights that temporal and spatial self-attention layers encode inter-frame and intra-frame dependency respectively, we introduce auxiliary motion-reference and reconstruction branches to produce text-guided motion and source features respectively. The obtained features are then injected into the main editing path via temporal and spatial self-attention layers. Extensive experiments demonstrate that UniEdit covers video motion editing and various appearance editing scenarios, and surpasses the state-of-the-art methods. Our code will be publicly available.Comment: Project page: https://jianhongbai.github.io/UniEdit

    A Survey on Non-Autoregressive Generation for Neural Machine Translation and Beyond

    Full text link
    Non-autoregressive (NAR) generation, which is first proposed in neural machine translation (NMT) to speed up inference, has attracted much attention in both machine learning and natural language processing communities. While NAR generation can significantly accelerate inference speed for machine translation, the speedup comes at the cost of sacrificed translation accuracy compared to its counterpart, auto-regressive (AR) generation. In recent years, many new models and algorithms have been designed/proposed to bridge the accuracy gap between NAR generation and AR generation. In this paper, we conduct a systematic survey with comparisons and discussions of various non-autoregressive translation (NAT) models from different aspects. Specifically, we categorize the efforts of NAT into several groups, including data manipulation, modeling methods, training criterion, decoding algorithms, and the benefit from pre-trained models. Furthermore, we briefly review other applications of NAR models beyond machine translation, such as dialogue generation, text summarization, grammar error correction, semantic parsing, speech synthesis, and automatic speech recognition. In addition, we also discuss potential directions for future exploration, including releasing the dependency of KD, dynamic length prediction, pre-training for NAR, and wider applications, etc. We hope this survey can help researchers capture the latest progress in NAR generation, inspire the design of advanced NAR models and algorithms, and enable industry practitioners to choose appropriate solutions for their applications. The web page of this survey is at \url{https://github.com/LitterBrother-Xiao/Overview-of-Non-autoregressive-Applications}.Comment: 25 pages, 11 figures, 4 table
    • …
    corecore