56 research outputs found

    CONVERSER: Few-Shot Conversational Dense Retrieval with Synthetic Data Generation

    Full text link
    Conversational search provides a natural interface for information retrieval (IR). Recent approaches have demonstrated promising results in applying dense retrieval to conversational IR. However, training dense retrievers requires large amounts of in-domain paired data. This hinders the development of conversational dense retrievers, as abundant in-domain conversations are expensive to collect. In this paper, we propose CONVERSER, a framework for training conversational dense retrievers with at most 6 examples of in-domain dialogues. Specifically, we utilize the in-context learning capability of large language models to generate conversational queries given a passage in the retrieval corpus. Experimental results on conversational retrieval benchmarks OR-QuAC and TREC CAsT 19 show that the proposed CONVERSER achieves comparable performance to fully-supervised models, demonstrating the effectiveness of our proposed framework in few-shot conversational dense retrieval. All source code and generated datasets are available at https://github.com/MiuLab/CONVERSERComment: Accepted to SIGDIAL 202

    Unsupervised Multilingual Dense Retrieval via Generative Pseudo Labeling

    Full text link
    Dense retrieval methods have demonstrated promising performance in multilingual information retrieval, where queries and documents can be in different languages. However, dense retrievers typically require a substantial amount of paired data, which poses even greater challenges in multilingual scenarios. This paper introduces UMR, an Unsupervised Multilingual dense Retriever trained without any paired data. Our approach leverages the sequence likelihood estimation capabilities of multilingual language models to acquire pseudo labels for training dense retrievers. We propose a two-stage framework which iteratively improves the performance of multilingual dense retrievers. Experimental results on two benchmark datasets show that UMR outperforms supervised baselines, showcasing the potential of training multilingual retrievers without paired data, thereby enhancing their practicality. Our source code, data, and models are publicly available at https://github.com/MiuLab/UMRComment: Accepted to Findings of EACL 202

    Model Extraction Attack against Self-supervised Speech Models

    Full text link
    Self-supervised learning (SSL) speech models generate meaningful representations of given clips and achieve incredible performance across various downstream tasks. Model extraction attack (MEA) often refers to an adversary stealing the functionality of the victim model with only query access. In this work, we study the MEA problem against SSL speech model with a small number of queries. We propose a two-stage framework to extract the model. In the first stage, SSL is conducted on the large-scale unlabeled corpus to pre-train a small speech model. Secondly, we actively sample a small portion of clips from the unlabeled corpus and query the target model with these clips to acquire their representations as labels for the small model's second-stage training. Experiment results show that our sampling methods can effectively extract the target model without knowing any information about its model architecture

    Risk factors and outcomes of carbapenem-nonsusceptible Escherichia coli bacteremia: A matched case–control study

    Get PDF
    BackgroundInfections due to carbapenem-resistant Enterobacteriaceae have been the emerging problem worldwide. This primary object of this study was to understand the risk factors and clinical outcomes of carbapenem-nonsusceptible Escherichia coli (CNSEc) bacteremia.MethodsWe conducted a matched case–control study in a 3,715-bed tertiary care medical center in northern Taiwan. The controls were selected among patients with carbapenem-susceptible E coli and were matched with CNSEc for bacteremia.ResultsFifty-one patients were included in this study (17 cases and 34 controls). Bivariate analysis showed that prior exposure to carbapenems (p<0.001), stay in intensive care units (p=0.016), placement of central venous catheters (p=0.001), chronic liver diseases (p<0.001), uremia with regular dialysis (p=0.004), and mechanical ventilation (p=0.004) were associated with CNSEc bacteremia. Multivariate analysis revealed that prior exposure to carbapenems [odds ratio (OR), 29.17; 95% confidence interval (CI), 1.76–484.70; p=0.019], uremia with regular dialysis (OR, 98.58; 95% CI, 4.02–999; p=0.005) and chronic liver diseases (OR, 27.86; 95% CI, 2.31–335.83; p=0.009) were independent risk factors for CNSEc bacteremia. Compared with carbapenem-susceptible E coli group, CNSEc group had a longer hospital stay (68.4 days vs. 35.8 days; p=0.04) and a higher disease severity, as indicated by a Pittsburgh bacteremia score greater than or equal to 4 (5.6% vs. 2.5%; p=0.015). Patients with CNSEc bacteremia had a higher overall in-hospital mortality rate (94.12% vs. 50.00%; p=0.002), but there was no difference in the 28-day mortality between these two groups.ConclusionsCNSEc bacteremia would lead to a poor outcome among patients with prior exposure to carbapenems, chronic liver disease, and uremia with regular dialysis

    Towards General-Purpose Text-Instruction-Guided Voice Conversion

    Full text link
    This paper introduces a novel voice conversion (VC) model, guided by text instructions such as "articulate slowly with a deep tone" or "speak in a cheerful boyish voice". Unlike traditional methods that rely on reference utterances to determine the attributes of the converted speech, our model adds versatility and specificity to voice conversion. The proposed VC model is a neural codec language model which processes a sequence of discrete codes, resulting in the code sequence of converted speech. It utilizes text instructions as style prompts to modify the prosody and emotional information of the given speech. In contrast to previous approaches, which often rely on employing separate encoders like prosody and content encoders to handle different aspects of the source speech, our model handles various information of speech in an end-to-end manner. Experiments have demonstrated the impressive capabilities of our model in comprehending instructions and delivering reasonable results.Comment: Accepted to ASRU 202

    BPR1K653, a Novel Aurora Kinase Inhibitor, Exhibits Potent Anti-Proliferative Activity in MDR1 (P-gp170)-Mediated Multidrug-Resistant Cancer Cells

    Get PDF
    Over-expression of Aurora kinases promotes the tumorigenesis of cells. The aim of this study was to determine the preclinical profile of a novel pan-Aurora kinase inhibitor, BPR1K653, as a candidate for anti-cancer therapy. Since expression of the drug efflux pump, MDR1, reduces the effectiveness of various chemotherapeutic compounds in human cancers, this study also aimed to determine whether the potency of BPR1K653 could be affected by the expression of MDR1 in cancer cells.BPR1K653 specifically inhibited the activity of Aurora-A and Aurora-B kinase at low nano-molar concentrations in vitro. Anti-proliferative activity of BPR1K653 was evaluated in various human cancer cell lines. Results of the clonogenic assay showed that BPR1K653 was potent in targeting a variety of cancer cell lines regardless of the tissue origin, p53 status, or expression of MDR1. At the cellular level, BPR1K653 induced endo-replication and subsequent apoptosis in both MDR1-negative and MDR1-positive cancer cells. Importantly, it showed potent activity against the growth of xenograft tumors of the human cervical carcinoma KB and KB-derived MDR1-positive KB-VIN10 cells in nude mice. Finally, BPR1K653 also exhibited favorable pharmacokinetic properties in rats.BPR1K653 is a novel potent anti-cancer compound, and its potency is not affected by the expression of the multiple drug resistant protein, MDR1, in cancer cells. Therefore, BPR1K653 is a promising anti-cancer compound that has potential for the management of various malignancies, particularly for patients with MDR1-related drug resistance after prolonged chemotherapeutic treatments

    The Effects of Aβ1-42 Binding to the SARS-CoV-2 Spike Protein S1 Subunit and Angiotensin-Converting Enzyme 2

    No full text
    Increasing evidence suggests that elderly people with dementia are vulnerable to the development of severe coronavirus disease 2019 (COVID-19). In Alzheimer’s disease (AD), the major form of dementia, β-amyloid (Aβ) levels in the blood are increased; however, the impact of elevated Aβ levels on the progression of COVID-19 remains largely unknown. Here, our findings demonstrate that Aβ1-42, but not Aβ1-40, bound to various viral proteins with a preferentially high affinity for the spike protein S1 subunit (S1) of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) and the viral receptor, angiotensin-converting enzyme 2 (ACE2). These bindings were mainly through the C-terminal residues of Aβ1-42. Furthermore, Aβ1-42 strengthened the binding of the S1 of SARS-CoV-2 to ACE2 and increased the viral entry and production of IL-6 in a SARS-CoV-2 pseudovirus infection model. Intriguingly, data from a surrogate mouse model with intravenous inoculation of Aβ1-42 show that the clearance of Aβ1-42 in the blood was dampened in the presence of the extracellular domain of the spike protein trimers of SARS-CoV-2, whose effects can be prevented by a novel anti-Aβ antibody. In conclusion, these findings suggest that the binding of Aβ1-42 to the S1 of SARS-CoV-2 and ACE2 may have a negative impact on the course and severity of SARS-CoV-2 infection. Further investigations are warranted to elucidate the underlying mechanisms and examine whether reducing the level of Aβ1-42 in the blood is beneficial to the fight against COVID-19 and AD
    corecore