4,826 research outputs found
Cross-domain paraphrasing for improving language modelling using out-of-domain data
In natural languages the variability in the underlying linguistic
generation rules significantly alters the observed surface word
sequence they create, and thus introduces a mismatch against
other data generated via alternative realizations associated with,
for example, a different domain. Hence, direct modelling of
out-of-domain data can result in poor generalization to the indomain
data of interest. To handle this problem, this paper
investigated using cross-domain paraphrastic language models
to improve in-domain language modelling (LM) using out-ofdomain
data. Phrase level paraphrase models learnt from each
domain were used to generate paraphrase variants for the data
of other domains. These were used to both improve the context
coverage of in-domain data, and reduce the domain mismatch of
the out-of-domain data. Significant error rate reduction of 0.6%
absolute was obtained on a state-of-the-art conversational telephone
speech recognition task using a cross-domain paraphrastic
multi-level LM trained on a billion words of mixed conversational
and broadcast news data. Consistent improvements on
the in-domain data context coverage were also obtained.The research leading to these results was supported by EPSRC Programme
Grant EP/I031022/1 (Natural Speech Technology) and DARPA
under the Broad Operational Language Translation (BOLT) program.This is the accepted manuscript. The final version is available at http://www.isca-speech.org/archive/interspeech_2013/i13_3424.htm
Conversational Process Modelling: State of the Art, Applications, and Implications in Practice
Chatbots such as ChatGPT have caused a tremendous hype lately. For BPM
applications, it is often not clear how to apply chatbots to generate business
value. Hence, this work aims at the systematic analysis of existing chatbots
for their support of conversational process modelling as process-oriented
capability. Application scenarios are identified along the process life cycle.
Then a systematic literature review on conversational process modelling is
performed. The resulting taxonomy serves as input for the identification of
application scenarios for conversational process modelling, including
paraphrasing and improvement of process descriptions. The application scenarios
are evaluated for existing chatbots based on a real-world test set from the
higher education domain. It contains process descriptions as well as
corresponding process models, together with an assessment of the model quality.
Based on the literature and application scenario analyses, recommendations for
the usage (practical implications) and further development (research
directions) of conversational process modelling are derived
Clinical dialogue transcription error correction using Seq2Seq models.
Good communication is critical to good healthcare. Clinical dialogue is a conversation between health practitioners and their patients, with the explicit goal of obtaining and sharing medical information. This information contributes to medical decision-making regarding the patient and plays a crucial role in their healthcare journey. The reliance on note taking and manual scribing processes are extremely inefficient and leads to manual transcription errors when digitizing notes. Automatic Speech Recognition (ASR) plays a significant role in speech-to-text applications, and can be directly used as a text generator in conversational applications. However, recording clinical dialogue presents a number of general and domain-specific challenges. In this paper, we present a seq2seq learning approach for ASR transcription error correction of clinical dialogues. We introduce a new Gastrointestinal Clinical Dialogue (GCD) Dataset which was gathered by health-care professionals from a NHS Inflammatory Bowel Disease clinic and use this in a comparative study with four commercial ASR systems. Using self-supervision strategies, we fine-tune a seq2seq model on a mask-filling task using a domain-specific PubMed dataset which we have shared publicly for future research. The BART model fine-tuned for mask-filling was able to correct transcription errors and achieve lower word error rates for three out of four commercial ASR outputs
Collaborative Business Process Management - A Literature-based Analysis of Methods for Supporting Model Understandability
Due to the growing amount of cooperative business scenarios, collaborative Business Process Management (cBPM) has emerged. The increased number of stakeholders with minor expertise in process modeling leads to a high relevance of model understandability in cBPM contexts. Despite extensive works in the research fields of cBPM and model understandability in BPM, there is no analysis and comprehensive overview of methods supporting process model understandability in cBPM scenarios. To address this research gap, this paper presents the results of a literature review. The paper identifies concepts for supporting model understandability in BPM, provides an overview of methods implementing these concepts, and discusses the methodsâ applicability in cBPM. The four concepts process model transformation, process model visualization, process model description, and modeling support are introduced. Subsequently, 69 methods are classified and discussed in the context of cBPM. Results contribute to revealing existing academic voids and can guide practitioners in cBPM scenarios
Neural Natural Language Generation: A Survey on Multilinguality, Multimodality, Controllability and Learning
Developing artificial learning systems that can understand and generate natural language has been one of the long-standing goals of artificial intelligence. Recent decades have witnessed an impressive progress on both of these problems, giving rise to a new family of approaches. Especially, the advances in deep learning over the past couple of years have led to neural approaches to natural language generation (NLG). These methods combine generative language learning techniques with neural-networks based frameworks. With a wide range of applications in natural language processing, neural NLG (NNLG) is a new and fast growing field of research. In this state-of-the-art report, we investigate the recent developments and applications of NNLG in its full extent from a multidimensional view, covering critical perspectives such as multimodality, multilinguality, controllability and learning strategies. We summarize the fundamental building blocks of NNLG approaches from these aspects and provide detailed reviews of commonly used preprocessing steps and basic neural architectures. This report also focuses on the seminal applications of these NNLG models such as machine translation, description generation, automatic speech recognition, abstractive summarization, text simplification, question answering and generation, and dialogue generation. Finally, we conclude with a thorough discussion of the described frameworks by pointing out some open research directions.This work has been partially supported by the European Commission ICT COST Action âMulti-task, Multilingual, Multi-modal Language Generationâ (CA18231). AE was supported by BAGEP 2021 Award of the Science Academy. EE was supported in part by TUBA GEBIP 2018 Award. BP is in in part funded by Independent Research Fund Denmark (DFF) grant 9063-00077B. IC has received funding from the European Unionâs Horizon 2020 research and innovation programme under the Marie Sklodowska-Curie grant agreement No 838188. EL is partly funded by Generalitat Valenciana and the Spanish Government throught projects PROMETEU/2018/089 and RTI2018-094649-B-I00, respectively. SMI is partly funded by UNIRI project uniri-drustv-18-20. GB is partly supported by the Ministry of Innovation and the National Research, Development and Innovation Office within the framework of the Hungarian Artificial Intelligence National Laboratory Programme. COT is partially funded by the Romanian Ministry of European Investments and Projects through the Competitiveness Operational Program (POC) project âHOLOTRAINâ (grant no. 29/221 ap2/07.04.2020, SMIS code: 129077) and by the German Academic Exchange Service (DAAD) through the project âAWAKEN: content-Aware and netWork-Aware faKE News mitigationâ (grant no. 91809005). ESA is partially funded by the German Academic Exchange Service (DAAD) through the project âDeep-Learning Anomaly Detection for Human and Automated Users Behaviorâ (grant no. 91809358)
Automated Question-Answering for Interactive Decision Support in Operations & Maintenance of Wind Turbines
Intelligent question-answering (QA) systems have witnessed increased interest in recent years, particularly in their ability to facilitate information access, data interpretation or decision support. The wind energy sector is one of the most promising sources of renewable energy, yet turbines regularly suffer from failures and operational inconsistencies, leading to downtimes and significant maintenance costs. Addressing these issues requires rapid interpretation of complex and dynamic data patterns under time-critical conditions. In this article, we present a novel approach that leverages interactive, natural language-based decision support for operations & maintenance (O&M) of wind turbines. The proposed interactive QA system allows engineers to pose domain-specific questions in natural language, and provides answers (in natural language) based on the automated retrieval of information on turbine sub-components, their properties and interactions, from a bespoke domain-specific knowledge graph. As data for specific faults is often sparse, we propose the use of paraphrase generation as a way to augment the existing dataset. Our QA system leverages encoder-decoder models to generate Cypher queries to obtain domain-specific facts from the KG database in response to user-posed natural language questions. Experiments with an attention-based sequence-to-sequence (Seq2Seq) model and a transformer show that the transformer accurately predicts up to 89.75% of responses to input questions, outperforming the Seq2Seq model marginally by 0.76%, though being 9.46 times more computationally efficient. The proposed QA system can help support engineers and technicians during O&M to reduce turbine downtime and operational costs, thus improving the reliability of wind energy as a source of renewable energy
Multi-Task Instruction Tuning of LLaMa for Specific Scenarios: A Preliminary Study on Writing Assistance
Proprietary Large Language Models (LLMs), such as ChatGPT, have garnered
significant attention due to their exceptional capabilities in handling a
diverse range of tasks. Recent studies demonstrate that open-sourced smaller
foundational models, such as 7B-size LLaMA, can also display remarkable
proficiency in tackling diverse tasks when fine-tuned using instruction-driven
data. In this work, we investigate a practical problem setting where the
primary focus is on one or a few particular tasks rather than general-purpose
instruction following, and explore whether LLMs can be beneficial and further
improved for such targeted scenarios. We choose the writing-assistant scenario
as the testbed, which includes seven writing tasks. We collect training data
for these tasks, reframe them in an instruction-following format, and
subsequently refine the LLM, specifically LLaMA, via instruction tuning.
Experimental results show that fine-tuning LLaMA on writing instruction data
significantly improves its ability on writing tasks. We also conduct more
experiments and analyses to offer insights for future work on effectively
fine-tuning LLaMA for specific scenarios. Finally, we initiate a discussion
regarding the necessity of employing LLMs for only one targeted task, taking
into account the efforts required for tuning and the resources consumed during
deployment
- âŠ