273 research outputs found
Survey of the State of the Art in Natural Language Generation: Core tasks, applications and evaluation
This paper surveys the current state of the art in Natural Language
Generation (NLG), defined as the task of generating text or speech from
non-linguistic input. A survey of NLG is timely in view of the changes that the
field has undergone over the past decade or so, especially in relation to new
(usually data-driven) methods, as well as new applications of NLG technology.
This survey therefore aims to (a) give an up-to-date synthesis of research on
the core tasks in NLG and the architectures adopted in which such tasks are
organised; (b) highlight a number of relatively recent research topics that
have arisen partly as a result of growing synergies between NLG and other areas
of artificial intelligence; (c) draw attention to the challenges in NLG
evaluation, relating them to similar challenges faced in other areas of Natural
Language Processing, with an emphasis on different evaluation methods and the
relationships between them.Comment: Published in Journal of AI Research (JAIR), volume 61, pp 75-170. 118
pages, 8 figures, 1 tabl
Explainable pattern modelling and summarization in sensor equipped smart homes of elderly
In the next several decades, the proportion of the elderly population is expected to increase significantly. This has led to various efforts to help live them independently for longer periods of time. Smart homes equipped with sensors provide a potential solution by capturing various behavioral and physiological patterns of the residents. In this work, we develop techniques to model and detect changes in these patterns. The focus is on methods that are explainable in nature and allow for generating natural language descriptions. We propose a comprehensive change description framework that can detect unusual changes in the sensor parameters and describe the data leading to those changes in natural language. An approach that models and detects variations in physiological and behavioral routines of the elderly forms one part of the change description framework. The second part comes from a natural language generation system in which we identify important health-relevant features from the sensor parameters. Throughout this dissertation, we validate the developed techniques using both synthetic and real data obtained from the homes of the elderly living in sensor-equipped facilities. Using multiple real data retrospective case studies, we show that our methods are able to detect variations in the sensor data that are correlated with important health events in the elderly as recorded in their Electronic Health Records.Includes bibliographical reference
A Review on Human-Computer Interaction and Intelligent Robots
In the field of artificial intelligence, human–computer interaction (HCI) technology and its related intelligent robot technologies are essential and interesting contents of research. From the perspective of software algorithm and hardware system, these above-mentioned technologies study and try to build a natural HCI environment. The purpose of this research is to provide an overview of HCI and intelligent robots. This research highlights the existing technologies of listening, speaking, reading, writing, and other senses, which are widely used in human interaction. Based on these same technologies, this research introduces some intelligent robot systems and platforms. This paper also forecasts some vital challenges of researching HCI and intelligent robots. The authors hope that this work will help researchers in the field to acquire the necessary information and technologies to further conduct more advanced research
Data-driven approaches to content selection for data-to-text generation
Data-to-text systems are powerful in generating reports from data automatically and
thus they simplify the presentation of complex data. Rather than presenting data using
visualisation techniques, data-to-text systems use human language, which is the most
common way for human-human communication. In addition, data-to-text systems can
adapt their output content to users’ preferences, background or interests and therefore
they can be pleasant for users to interact with. Content selection is an important part
of every data-to-text system, because it is the module that decides which from the
available information should be conveyed to the user.
This thesis makes three important contributions. Firstly, it investigates data-driven
approaches to content selection with respect to users’ preferences. It develops, compares
and evaluates two novel content selection methods. The first method treats content
selection as a Markov Decision Process (MDP), where the content selection decisions
are made sequentially, i.e. given the already chosen content, decide what to talk about
next. The MDP is solved using Reinforcement Learning (RL) and is optimised with
respect to a cumulative reward function. The second approach considers all content
selection decisions simultaneously by taking into account data relationships and treats
content selection as a multi-label classification task. The evaluation shows that the users
significantly prefer the output produced by the RL framework, whereas the multi-label
classification approach scores significantly higher than the RL method in automatic
metrics. The results also show that the end users’ preferences should be taken into
account when developing Natural Language Generation (NLG) systems.
NLG systems are developed with the assistance of domain experts, however the end
users are normally non-experts. Consider for instance a student feedback generation
system, where the system imitates the teachers. The system will produce feedback based
on the lecturers’ rather than the students’ preferences although students are the end
users. Therefore, the second contribution of this thesis is an approach that adapts the
content to “speakers” and “hearers” simultaneously. It considers initially two types of
known stakeholders; lecturers and students. It develops a novel approach that analyses
the preferences of the two groups using Principal Component Regression and uses the derived knowledge to hand-craft a reward function that is then optimised using RL.
The results show that the end users prefer the output generated by this system, rather
than the output that is generated by a system that mimics the experts. Therefore, it is
possible to model the middle ground of the preferences of different known stakeholders.
In most real world applications however, first-time users are generally unknown,
which is a common problem for NLG and interactive systems: the system cannot adapt
to user preferences without prior knowledge. This thesis contributes a novel framework
for addressing unknown stakeholders such as first time users, using Multi-objective Optimisation
to minimise regret for multiple possible user types. In this framework, the
content preferences of potential users are modelled as objective functions, which are
simultaneously optimised using Multi-objective Optimisation. This approach outperforms
two meaningful baselines and minimises regret for unknown users
Language modelling for clinical natural language understanding and generation
One of the long-standing objectives of Artificial Intelligence (AI) is to design and develop algorithms for social good including tackling public health challenges. In the era of digitisation, with an unprecedented amount of healthcare data being captured in digital form, the analysis of the healthcare data at scale can lead to better research of diseases, better monitoring patient conditions and more importantly improving patient outcomes. However, many AI-based analytic algorithms rely solely on structured healthcare data such as bedside measurements and test results which only account for 20% of all healthcare data, whereas the remaining 80% of healthcare data is unstructured including textual data such as clinical notes and discharge summaries which is still underexplored.
Conventional Natural Language Processing (NLP) algorithms that are designed for clinical applications rely on the shallow matching, templates and non-contextualised word embeddings which lead to limited understanding of contextual semantics. Though recent advances in NLP algorithms have demonstrated promising performance on a variety of NLP tasks in the general domain with contextualised language models, most of these generic NLP algorithms struggle at specific clinical NLP tasks which require biomedical knowledge and reasoning. Besides, there is limited research to study generative NLP algorithms to generate clinical reports and summaries automatically by considering salient clinical information.
This thesis aims to design and develop novel NLP algorithms especially clinical-driven contextualised language models to understand textual healthcare data and generate clinical narratives which can potentially support clinicians, medical scientists and patients. The first contribution of this thesis focuses on capturing phenotypic information of patients from clinical notes which is important to profile patient situation and improve patient outcomes. The thesis proposes a novel self-supervised language model, named Phenotypic Intelligence Extraction (PIE), to annotate phenotypes from clinical notes with the detection of contextual synonyms and the enhancement to reason with numerical values. The second contribution is to demonstrate the utility and benefits of using phenotypic features of patients in clinical use cases by predicting patient outcomes in Intensive Care Units (ICU) and identifying patients at risk of specific diseases with better accuracy and model interpretability. The third contribution is to propose generative models to generate clinical narratives to automate and accelerate the process of report writing and summarisation by clinicians. This thesis first proposes a novel summarisation language model named PEGASUS which surpasses or is on par with the state-of-the-art performance on 12 downstream datasets including biomedical literature from PubMed. PEGASUS is further extended to generate medical scientific documents from input tabular data.Open Acces
Making effective use of healthcare data using data-to-text technology
Healthcare organizations are in a continuous effort to improve health
outcomes, reduce costs and enhance patient experience of care. Data is
essential to measure and help achieving these improvements in healthcare
delivery. Consequently, a data influx from various clinical, financial and
operational sources is now overtaking healthcare organizations and their
patients. The effective use of this data, however, is a major challenge.
Clearly, text is an important medium to make data accessible. Financial reports
are produced to assess healthcare organizations on some key performance
indicators to steer their healthcare delivery. Similarly, at a clinical level,
data on patient status is conveyed by means of textual descriptions to
facilitate patient review, shift handover and care transitions. Likewise,
patients are informed about data on their health status and treatments via
text, in the form of reports or via ehealth platforms by their doctors.
Unfortunately, such text is the outcome of a highly labour-intensive process if
it is done by healthcare professionals. It is also prone to incompleteness,
subjectivity and hard to scale up to different domains, wider audiences and
varying communication purposes. Data-to-text is a recent breakthrough
technology in artificial intelligence which automatically generates natural
language in the form of text or speech from data. This chapter provides a
survey of data-to-text technology, with a focus on how it can be deployed in a
healthcare setting. It will (1) give an up-to-date synthesis of data-to-text
approaches, (2) give a categorized overview of use cases in healthcare, (3)
seek to make a strong case for evaluating and implementing data-to-text in a
healthcare setting, and (4) highlight recent research challenges.Comment: 27 pages, 2 figures, book chapte
Natural Language Interfaces to Data
Recent advances in NLU and NLP have resulted in renewed interest in natural
language interfaces to data, which provide an easy mechanism for non-technical
users to access and query the data. While early systems evolved from keyword
search and focused on simple factual queries, the complexity of both the input
sentences as well as the generated SQL queries has evolved over time. More
recently, there has also been a lot of focus on using conversational interfaces
for data analytics, empowering a line of non-technical users with quick
insights into the data. There are three main challenges in natural language
querying (NLQ): (1) identifying the entities involved in the user utterance,
(2) connecting the different entities in a meaningful way over the underlying
data source to interpret user intents, and (3) generating a structured query in
the form of SQL or SPARQL.
There are two main approaches for interpreting a user's NLQ. Rule-based
systems make use of semantic indices, ontologies, and KGs to identify the
entities in the query, understand the intended relationships between those
entities, and utilize grammars to generate the target queries. With the
advances in deep learning (DL)-based language models, there have been many
text-to-SQL approaches that try to interpret the query holistically using DL
models. Hybrid approaches that utilize both rule-based techniques as well as DL
models are also emerging by combining the strengths of both approaches.
Conversational interfaces are the next natural step to one-shot NLQ by
exploiting query context between multiple turns of conversation for
disambiguation. In this article, we review the background technologies that are
used in natural language interfaces, and survey the different approaches to
NLQ. We also describe conversational interfaces for data analytics and discuss
several benchmarks used for NLQ research and evaluation.Comment: The full version of this manuscript, as published by Foundations and
Trends in Databases, is available at http://dx.doi.org/10.1561/190000007
Controlling Hallucinations at Word Level in Data-to-Text Generation
Data-to-Text Generation (DTG) is a subfield of Natural Language Generation
aiming at transcribing structured data in natural language descriptions. The
field has been recently boosted by the use of neural-based generators which
exhibit on one side great syntactic skills without the need of hand-crafted
pipelines; on the other side, the quality of the generated text reflects the
quality of the training data, which in realistic settings only offer
imperfectly aligned structure-text pairs. Consequently, state-of-art neural
models include misleading statements - usually called hallucinations - in their
outputs. The control of this phenomenon is today a major challenge for DTG, and
is the problem addressed in the paper.
Previous work deal with this issue at the instance level: using an alignment
score for each table-reference pair. In contrast, we propose a finer-grained
approach, arguing that hallucinations should rather be treated at the word
level. Specifically, we propose a Multi-Branch Decoder which is able to
leverage word-level labels to learn the relevant parts of each training
instance. These labels are obtained following a simple and efficient scoring
procedure based on co-occurrence analysis and dependency parsing. Extensive
evaluations, via automated metrics and human judgment on the standard WikiBio
benchmark, show the accuracy of our alignment labels and the effectiveness of
the proposed Multi-Branch Decoder. Our model is able to reduce and control
hallucinations, while keeping fluency and coherence in generated texts. Further
experiments on a degraded version of ToTTo show that our model could be
successfully used on very noisy settings.Comment: 20 pages, 6 figures, 5 tables (excluding Appendix). Source code:
https://github.com/KaijuML/dtt-multi-branc
- …