1,655 research outputs found

    ReLoop2: Building Self-Adaptive Recommendation Models via Responsive Error Compensation Loop

    Full text link
    Industrial recommender systems face the challenge of operating in non-stationary environments, where data distribution shifts arise from evolving user behaviors over time. To tackle this challenge, a common approach is to periodically re-train or incrementally update deployed deep models with newly observed data, resulting in a continual training process. However, the conventional learning paradigm of neural networks relies on iterative gradient-based updates with a small learning rate, making it slow for large recommendation models to adapt. In this paper, we introduce ReLoop2, a self-correcting learning loop that facilitates fast model adaptation in online recommender systems through responsive error compensation. Inspired by the slow-fast complementary learning system observed in human brains, we propose an error memory module that directly stores error samples from incoming data streams. These stored samples are subsequently leveraged to compensate for model prediction errors during testing, particularly under distribution shifts. The error memory module is designed with fast access capabilities and undergoes continual refreshing with newly observed data samples during the model serving phase to support fast model adaptation. We evaluate the effectiveness of ReLoop2 on three open benchmark datasets as well as a real-world production dataset. The results demonstrate the potential of ReLoop2 in enhancing the responsiveness and adaptiveness of recommender systems operating in non-stationary environments.Comment: Accepted by KDD 2023. See the project page at https://xpai.github.io/ReLoo

    Diffusion Model is an Effective Planner and Data Synthesizer for Multi-Task Reinforcement Learning

    Full text link
    Diffusion models have demonstrated highly-expressive generative capabilities in vision and NLP. Recent studies in reinforcement learning (RL) have shown that diffusion models are also powerful in modeling complex policies or trajectories in offline datasets. However, these works have been limited to single-task settings where a generalist agent capable of addressing multi-task predicaments is absent. In this paper, we aim to investigate the effectiveness of a single diffusion model in modeling large-scale multi-task offline data, which can be challenging due to diverse and multimodal data distribution. Specifically, we propose Multi-Task Diffusion Model (\textsc{MTDiff}), a diffusion-based method that incorporates Transformer backbones and prompt learning for generative planning and data synthesis in multi-task offline settings. \textsc{MTDiff} leverages vast amounts of knowledge available in multi-task data and performs implicit knowledge sharing among tasks. For generative planning, we find \textsc{MTDiff} outperforms state-of-the-art algorithms across 50 tasks on Meta-World and 8 maps on Maze2D. For data synthesis, \textsc{MTDiff} generates high-quality data for testing tasks given a single demonstration as a prompt, which enhances the low-quality datasets for even unseen tasks.Comment: 21 page

    Universal Trading for Order Execution with Oracle Policy Distillation

    Full text link
    As a fundamental problem in algorithmic trading, order execution aims at fulfilling a specific trading order, either liquidation or acquirement, for a given instrument. Towards effective execution strategy, recent years have witnessed the shift from the analytical view with model-based market assumptions to model-free perspective, i.e., reinforcement learning, due to its nature of sequential decision optimization. However, the noisy and yet imperfect market information that can be leveraged by the policy has made it quite challenging to build up sample efficient reinforcement learning methods to achieve effective order execution. In this paper, we propose a novel universal trading policy optimization framework to bridge the gap between the noisy yet imperfect market states and the optimal action sequences for order execution. Particularly, this framework leverages a policy distillation method that can better guide the learning of the common policy towards practically optimal execution by an oracle teacher with perfect information to approximate the optimal trading strategy. The extensive experiments have shown significant improvements of our method over various strong baselines, with reasonable trading actions.Comment: Accepted in AAAI 2021, the code and the supplementary materials are in https://seqml.github.io/opd

    Text-Only Domain Adaptation for End-to-End Speech Recognition through Down-Sampling Acoustic Representation

    Full text link
    Mapping two modalities, speech and text, into a shared representation space, is a research topic of using text-only data to improve end-to-end automatic speech recognition (ASR) performance in new domains. However, the length of speech representation and text representation is inconsistent. Although the previous method up-samples the text representation to align with acoustic modality, it may not match the expected actual duration. In this paper, we proposed novel representations match strategy through down-sampling acoustic representation to align with text modality. By introducing a continuous integrate-and-fire (CIF) module generating acoustic representations consistent with token length, our ASR model can learn unified representations from both modalities better, allowing for domain adaptation using text-only data of the target domain. Experiment results of new domain data demonstrate the effectiveness of the proposed method.Comment: Accepted by INTERSPEECH 2023. arXiv admin note: text overlap with arXiv:2309.0143

    Early Identification of High-Risk TIA or Minor Stroke Using Artificial Neural Network

    Get PDF
    Background and Purpose: The risk of recurrent stroke following a transient ischemic attack (TIA) or minor stroke is high, despite of a significant reduction in the past decade. In this study, we investigated the feasibility of using artificial neural network (ANN) for risk stratification of TIA or minor stroke patients.Methods: Consecutive patients with acute TIA or minor ischemic stroke presenting at a tertiary hospital during a 2-year period were recruited. We collected demographics, clinical and imaging data at baseline. The primary outcome was recurrent ischemic stroke within 1 year. We developed ANN models to predict the primary outcome. We randomly down-sampled patients without a primary outcome to 1:1 match with those with a primary outcome to mitigate data imbalance. We used a 5-fold cross-validation approach to train and test the ANN models to avoid overfitting. We employed 19 independent variables at baseline as the input neurons in the ANN models, using a learning algorithm based on backpropagation to minimize the loss function. We obtained the sensitivity, specificity, accuracy and the c statistic of each ANN model from the 5 rounds of cross-validation and compared that of support vector machine (SVM) and Naïve Bayes classifier in risk stratification of the patients.Results: A total of 451 acute TIA or minor stroke patients were enrolled. Forty (8.9%) patients had a recurrent ischemic stroke within 1 year. Another 40 patients were randomly selected from those with no recurrent stroke, so that data from 80 patients in total were used for 5 rounds of training and testing of ANN models. The median sensitivity, specificity, accuracy and c statistic of the ANN models to predict recurrent stroke at 1 year was 75%, 75%, 75%, and 0.77, respectively. ANN model outperformed SVM and Naïve Bayes classifier in our dataset for predicting relapse after TIA or minor stroke.Conclusion: This pilot study indicated that ANN may yield a novel and effective method in risk stratification of TIA and minor stroke. Further studies are warranted for verification and improvement of the current ANN model

    Who reports insufficient and disturbed sleep? Results from a representative population-based health survey in Hong Kong.

    Get PDF
    OBJECTIVES: To highlight the prevalence of sleep problems and identify associated risk factors among a representative sample recruited from the general population of Hong Kong. DESIGN, SETTING AND PARTICIPANTS: Participants included 12 022 individuals (aged 15 or above) who took part in the Population Health Survey 2014/15, a territory-wide survey conducted by the Department of Health of the Government of the Hong Kong Special Administrative Region. PRIMARY AND SECONDARY OUTCOME MEASURES: Outcomes were the prevalence of (1) insufficient sleep (<6 hours sleep per day) and (2) any sleep disturbance (difficulty initiating sleep, intermittent awakenings, early awakening) ≥3 times per week in the past 30 days. Multivariable logistic regression identified associations between sleep problems and sociodemographic, clinical and lifestyle factors. RESULTS: 9.7% of respondents reported insufficient sleep and 10.5% reported sleep disturbances ≥3 times a week. Female gender, monthly household income <$12 250 (Hong Kong dollar), lower education level, mental health condition and physical health condition were significantly associated with both insufficient and disturbed sleep (all p<0.05). Unemployment, homemaker, insufficient physical activity, current/former smoking status and harmful alcohol consumption were associated with sleep disturbances only (all p<0.01). CONCLUSIONS: Sleep problems are highly prevalent in Hong Kong. As such problems are associated with a range of health conditions, it is important to facilitate improvements in sleep. Our results show that harmful alcohol consumption, insufficient physical activity and current smoking are modifiable risk factors for sleep disturbances. Public health campaigns should focus on these risk factors in order to promote a healthy lifestyle and ultimately reduce sleep disturbances. Targeted interventions for high-risk groups may also be warranted, particularly for those with doctor-diagnosed physical and mental health conditions

    A functional variant in the Stearoyl-CoA desaturase gene promoter enhances fatty acid desaturation in pork

    Get PDF
    There is growing public concern about reducing saturated fat intake. Stearoyl-CoA desaturase (SCD) is the lipogenic enzyme responsible for the biosynthesis of oleic acid (18:1) by desaturating stearic acid (18:0). Here we describe a total of 18 mutations in the promoter and 3′ non-coding region of the pig SCD gene and provide evidence that allele T at AY487830:g.2228T>C in the promoter region enhances fat desaturation (the ratio 18:1/18:0 in muscle increases from 3.78 to 4.43 in opposite homozygotes) without affecting fat content (18:0+18:1, intramuscular fat content, and backfat thickness). No mutations that could affect the functionality of the protein were found in the coding region. First, we proved in a purebred Duroc line that the C-T-A haplotype of the 3 single nucleotide polymorphisms (SNPs) (g.2108C>T; g.2228T>C; g.2281A>G) of the promoter region was additively associated to enhanced 18:1/18:0 both in muscle and subcutaneous fat, but not in liver. We show that this association was consistent over a 10-year period of overlapping generations and, in line with these results, that the C-T-A haplotype displayed greater SCD mRNA expression in muscle. The effect of this haplotype was validated both internally, by comparing opposite homozygote siblings, and externally, by using experimental Duroc-based crossbreds. Second, the g.2281A>G and the g.2108C>T SNPs were excluded as causative mutations using new and previously published data, restricting the causality to g.2228T>C SNP, the last source of genetic variation within the haplotype. This mutation is positioned in the core sequence of several putative transcription factor binding sites, so that there are several plausible mechanisms by which allele T enhances 18:1/18:0 and, consequently, the proportion of monounsaturated to saturated fat.This research was supported by grants from the Spanish Ministry of Science and Innovation (AGL2009-09779 and AGL2012-33529). RRF is recipient of a PhD scholarship from the Spanish Ministry of Science and Innovation (BES-2010-034607). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of manuscript

    Optimasi Portofolio Resiko Menggunakan Model Markowitz MVO Dikaitkan dengan Keterbatasan Manusia dalam Memprediksi Masa Depan dalam Perspektif Al-Qur`an

    Full text link
    Risk portfolio on modern finance has become increasingly technical, requiring the use of sophisticated mathematical tools in both research and practice. Since companies cannot insure themselves completely against risk, as human incompetence in predicting the future precisely that written in Al-Quran surah Luqman verse 34, they have to manage it to yield an optimal portfolio. The objective here is to minimize the variance among all portfolios, or alternatively, to maximize expected return among all portfolios that has at least a certain expected return. Furthermore, this study focuses on optimizing risk portfolio so called Markowitz MVO (Mean-Variance Optimization). Some theoretical frameworks for analysis are arithmetic mean, geometric mean, variance, covariance, linear programming, and quadratic programming. Moreover, finding a minimum variance portfolio produces a convex quadratic programming, that is minimizing the objective function ðð¥with constraintsð ð 𥠥 ðandð´ð¥ = ð. The outcome of this research is the solution of optimal risk portofolio in some investments that could be finished smoothly using MATLAB R2007b software together with its graphic analysis

    Measurement of t(t)over-bar normalised multi-differential cross sections in pp collisions at root s=13 TeV, and simultaneous determination of the strong coupling strength, top quark pole mass, and parton distribution functions

    Get PDF
    Peer reviewe

    Measurement of the top quark forward-backward production asymmetry and the anomalous chromoelectric and chromomagnetic moments in pp collisions at √s = 13 TeV

    Get PDF
    Abstract The parton-level top quark (t) forward-backward asymmetry and the anomalous chromoelectric (d̂ t) and chromomagnetic (μ̂ t) moments have been measured using LHC pp collisions at a center-of-mass energy of 13 TeV, collected in the CMS detector in a data sample corresponding to an integrated luminosity of 35.9 fb−1. The linearized variable AFB(1) is used to approximate the asymmetry. Candidate t t ¯ events decaying to a muon or electron and jets in final states with low and high Lorentz boosts are selected and reconstructed using a fit of the kinematic distributions of the decay products to those expected for t t ¯ final states. The values found for the parameters are AFB(1)=0.048−0.087+0.095(stat)−0.029+0.020(syst),μ̂t=−0.024−0.009+0.013(stat)−0.011+0.016(syst), and a limit is placed on the magnitude of | d̂ t| &lt; 0.03 at 95% confidence level. [Figure not available: see fulltext.
    corecore