740 research outputs found

    NADiffuSE: Noise-aware Diffusion-based Model for Speech Enhancement

    Full text link
    The goal of speech enhancement (SE) is to eliminate the background interference from the noisy speech signal. Generative models such as diffusion models (DM) have been applied to the task of SE because of better generalization in unseen noisy scenes. Technical routes for the DM-based SE methods can be summarized into three types: task-adapted diffusion process formulation, generator-plus-conditioner (GPC) structures and the multi-stage frameworks. We focus on the first two approaches, which are constructed under the GPC architecture and use the task-adapted diffusion process to better deal with the real noise. However, the performance of these SE models is limited by the following issues: (a) Non-Gaussian noise estimation in the task-adapted diffusion process. (b) Conditional domain bias caused by the weak conditioner design in the GPC structure. (c) Large amount of residual noise caused by unreasonable interpolation operations during inference. To solve the above problems, we propose a noise-aware diffusion-based SE model (NADiffuSE) to boost the SE performance, where the noise representation is extracted from the noisy speech signal and introduced as a global conditional information for estimating the non-Gaussian components. Furthermore, the anchor-based inference algorithm is employed to achieve a compromise between the speech distortion and noise residual. In order to mitigate the performance degradation caused by the conditional domain bias in the GPC framework, we investigate three model variants, all of which can be viewed as multi-stage SE based on the preprocessing networks for Mel spectrograms. Experimental results show that NADiffuSE outperforms other DM-based SE models under the GPC infrastructure. Audio samples are available at: https://square-of-w.github.io/NADiffuSE-demo/

    Research hotspots and trends of fresh e-commerce in China: A knowledge mapping analysis based on bibliometrics

    Get PDF
    The fresh e-commerce industry has seen a sudden and substantial rise since the outbreak of COVID-19. The rapid development of this industry calls for a comprehensive and systematic review of its research status, hotspots and future trends, which will have significant implications for researchers in related fields. This paper first conducts a current situation analysis of the core literature on fresh e-commerce retrieved from four databases – CNKI, CSSCI, Wanfang and VIP – to categorize the research status of fresh e-commerce in three dimensions: the year of publication, article sources, and distribution of subjects. CiteSpace is then used to perform a bibliometric analysis of the data and to create visualized knowledge maps. The results show that the research on fresh e-commerce can be divided into three stages: rapid development (2012-2015), exploration and transformation (2016-2019), maturity and upgrade (2020-present). At each stage, the research evolves toward diversity and maturity with policy developments and changes in the external environment. Cold chain logistics, business models, freshness-keeping of products and e-commerce are ongoing research hotspots in fresh produce e-commerce, while later studies focus more on the transformation and upgrade of products, logistics, distribution and platforms to better serve consumers’ consumption habits and environmental requirements. This study provides valuable insights for researchers and enterprises who are engaged in the industry and for those who are interested in the development of fresh e-commerce in China

    Sales forecasting of stores in shopping malls: A study based on external data and transaction data

    Get PDF
    To improve the forecast accuracy of the sales of stores in shopping malls, this paper proposes a prediction method based on deep learning that comprehensively considers the external data, such as online review data of shopping mall stores, weather data, weekday/weekend data, and historical transaction data of the stores. To begin with, the online review data of the stores are pre-trained with BERT (Bidirectional Encoder Representations from Transformers) to complete the multi-label sentiment classification and obtain the intensity index of perceived sentiment of reviews. The index, together with other external data, such as online ratings, weather, weekday/weekend differences, and historical transactions of the stores, is pre-processed. At last, the Long Short-Term Memory (LSTM) and the Attention models are used to predict the sales volume of stores in a certain shopping mall. The results show that the addition of external data – weather, weekday/weekend, online ratings and intensity index of sentiment of reviews – to the historical sales data-based model can effectively improve the forecast accuracy of store sales

    Qilin-Med: Multi-stage Knowledge Injection Advanced Medical Large Language Model

    Full text link
    Integrating large language models (LLMs) into healthcare presents potential but faces challenges. Directly pre-training LLMs for domains like medicine is resource-heavy and sometimes unfeasible. Sole reliance on Supervised Fine-tuning (SFT) can result in overconfident predictions and may not tap into domain specific insights. Addressing these challenges, we present a multi-stage training method combining Domain-specific Continued Pre-training (DCPT), SFT, and Direct Preference Optimization (DPO). A notable contribution of our study is the introduction of a 3Gb Chinese Medicine (ChiMed) dataset, encompassing medical question answering, plain texts, knowledge graphs, and dialogues, segmented into three training stages. The medical LLM trained with our pipeline, Qilin-Med, exhibits significant performance boosts. In the CPT and SFT phases, it achieves 38.4% and 40.0% accuracy on the CMExam, surpassing Baichuan-7B's 33.5%. In the DPO phase, on the Huatuo-26M test set, it scores 16.66 in BLEU-1 and 27.44 in ROUGE1, outperforming the SFT's 12.69 and 24.21. This highlights the strength of our training approach in refining LLMs for medical applications

    ML-LMCL: Mutual Learning and Large-Margin Contrastive Learning for Improving ASR Robustness in Spoken Language Understanding

    Full text link
    Spoken language understanding (SLU) is a fundamental task in the task-oriented dialogue systems. However, the inevitable errors from automatic speech recognition (ASR) usually impair the understanding performance and lead to error propagation. Although there are some attempts to address this problem through contrastive learning, they (1) treat clean manual transcripts and ASR transcripts equally without discrimination in fine-tuning; (2) neglect the fact that the semantically similar pairs are still pushed away when applying contrastive learning; (3) suffer from the problem of Kullback-Leibler (KL) vanishing. In this paper, we propose Mutual Learning and Large-Margin Contrastive Learning (ML-LMCL), a novel framework for improving ASR robustness in SLU. Specifically, in fine-tuning, we apply mutual learning and train two SLU models on the manual transcripts and the ASR transcripts, respectively, aiming to iteratively share knowledge between these two models. We also introduce a distance polarization regularizer to avoid pushing away the intra-cluster pairs as much as possible. Moreover, we use a cyclical annealing schedule to mitigate KL vanishing issue. Experiments on three datasets show that ML-LMCL outperforms existing models and achieves new state-of-the-art performance

    Rethinking Multi-Interest Learning for Candidate Matching in Recommender Systems

    Full text link
    Existing research efforts for multi-interest candidate matching in recommender systems mainly focus on improving model architecture or incorporating additional information, neglecting the importance of training schemes. This work revisits the training framework and uncovers two major problems hindering the expressiveness of learned multi-interest representations. First, the current training objective (i.e., uniformly sampled softmax) fails to effectively train discriminative representations in a multi-interest learning scenario due to the severe increase in easy negative samples. Second, a routing collapse problem is observed where each learned interest may collapse to express information only from a single item, resulting in information loss. To address these issues, we propose the REMI framework, consisting of an Interest-aware Hard Negative mining strategy (IHN) and a Routing Regularization (RR) method. IHN emphasizes interest-aware hard negatives by proposing an ideal sampling distribution and developing a Monte-Carlo strategy for efficient approximation. RR prevents routing collapse by introducing a novel regularization term on the item-to-interest routing matrices. These two components enhance the learned multi-interest representations from both the optimization objective and the composition information. REMI is a general framework that can be readily applied to various existing multi-interest candidate matching methods. Experiments on three real-world datasets show our method can significantly improve state-of-the-art methods with easy implementation and negligible computational overhead. The source code will be released.Comment: RecSys 202

    Histone deacetylase 11 regulates oligodendrocyte-specific gene expression and cell development in OL-1 oligodendroglia cells

    Get PDF
    Both in vivo and in vitro studies indicate a correlation between reduced acetylation of histone core proteins and oligodendrocyte development. The nature of these histone modifications and the mechanisms mediating them remain undefined. To address these issues we utilized OL-1 cells, a rat non-transformed oligodendrocyte cell line, and primary oligodendrocyte cultures. We found that the acetylated histone H3 at lysine 9 and lysine 14 (H3K9/K14ac) is reduced in both the myelin basic protein (MBP) and proteolipid protein (PLP) genes of maturing oligodendroglial OL-1 cells, and furthermore, this temporally correlates with increases in MBP, PLP, and histone deacetylase (HDAC) 11 expression. Disruption of developmentally-regulated histone H3 deacetylation within the MBP and PLP genes by the HDAC inhibitor trichostatin A blunts MBP and PLP expression. With its increased expression, interaction of HDAC 11 with acetylated histone H3 and recruitment of HDAC 11 to the MBP and PLP genes markedly increases in maturing OL-1 cells. Moreover, suppressing HDAC 11 expression with small interfering RNA significantly: 1) increases H3K9/K14ac globally and within the MBP and PLP genes, 2) decreases MBP and PLP mRNA expression, and 3) blunts the morphological changes associated with oligodendrocyte development. Our data strongly support a specific role for HDAC 11 in histone deacetylation and in turn the regulation of oligodendrocyte-specific protein gene expression and oligodendrocyte development

    Equivariant Contrastive Learning for Sequential Recommendation

    Full text link
    Contrastive learning (CL) benefits the training of sequential recommendation models with informative self-supervision signals. Existing solutions apply general sequential data augmentation strategies to generate positive pairs and encourage their representations to be invariant. However, due to the inherent properties of user behavior sequences, some augmentation strategies, such as item substitution, can lead to changes in user intent. Learning indiscriminately invariant representations for all augmentation strategies might be suboptimal. Therefore, we propose Equivariant Contrastive Learning for Sequential Recommendation (ECL-SR), which endows SR models with great discriminative power, making the learned user behavior representations sensitive to invasive augmentations (e.g., item substitution) and insensitive to mild augmentations (e.g., featurelevel dropout masking). In detail, we use the conditional discriminator to capture differences in behavior due to item substitution, which encourages the user behavior encoder to be equivariant to invasive augmentations. Comprehensive experiments on four benchmark datasets show that the proposed ECL-SR framework achieves competitive performance compared to state-of-the-art SR models. The source code is available at https://github.com/Tokkiu/ECL.Comment: Accepted by RecSys 202

    β-catenin mediates insulin-like growth factor-I actions to promote cyclin D1 mRNA expression, cell proliferation and survival in oligodendroglial cultures

    Get PDF
    By promoting cell proliferation, survival and maturation insulin-like growth factor (IGF)-I is essential to the normal growth and development of the central nervous system. It is clear that IGF-I actions are primarily mediated by the type I IGF receptor (IGF1R), and that phosphoinositide 3 (PI3)-Akt kinases and MAP kinases signal many of IGF-I-IGF1R actions in neural cells, including oligodendrocyte lineage cells. The precise downstream targets of these signaling pathways, however, remain to be defined. We studied oligodendroglial cells to determine whether β-catenin, a molecule that is a downstream target of glycogen synthase kinase-3β (GSK3β) and plays a key role in the Wnt canonical signaling pathway, mediates IGF-I actions. We found that IGF-I increases β-catenin protein abundance within an hour after IGF-I-induced phosphorylation of Akt and GSK3β. Inhibiting the PI3-Akt pathway suppressed IGF-I-induced increases in β-catenin and cyclin D1 mRNA, while suppression of GSK3β activity simulated IGF-I actions. Knocking-down β-catenin mRNA by RNA interference suppressed IGF-I-stimulated increases in the abundance of cyclin D1 mRNA, cell proliferation, and cell survival. Our data suggest that β-catenin is an important downstream molecule in the PI3-Akt-GSK3β pathway, and as such it mediates IGF-I upregulation of cyclin D1 mRNA and promotion of cell proliferation and survival in oligodendroglial cells

    Developmental expression of histone deacetylase 11 in the murine brain

    Get PDF
    Recent studies indicate that neural cell development in the central nervous system (CNS) correlates with a reduction in acetylation of histone core proteins. Moreover, histone hypoacetylation is thought to be important to oligodendrocyte lineage development. The mechanisms mediating the reduction in acetylation during postnatal neural development remain to be defined. To begin to understand these mechanisms, we investigated the expression of histone deacetylase 11 (HDAC11), a newly identified HDAC, in mouse brain during postnatal development. We show that HDAC11 was widely expressed in the brain and that this expression gradually increased in a region-specific pattern between birth and 4 weeks of age. At the cellular level HDAC11 protein was predominately localized in the nuclei of mature oligodendrocytes but only minimally in astrocytes. Although dentate gyrus granule neurons abundantly expressed HDAC11, granule neuron precursors in the subgranule layer exhibited little HDAC11 immunoreactivity. Double-immunostaining of the corpus callosum and dentate gyrus demonstrated that HDAC11 and Ki67, a cell-proliferating marker, are rarely colocalized in same cells. Our data show that HDAC11 was expressed in the developing brain in a temporal and spatial pattern that correlates with the maturation of neural cells, including cells of the oligodendrocyte lineage. These findings support a role for HDAC11 in CNS histone deacetylation and the development of oligodendrocytes and neurons during postnatal development
    • …
    corecore