5,890 research outputs found
A Survey of Natural Language Generation
This paper offers a comprehensive review of the research on Natural Language
Generation (NLG) over the past two decades, especially in relation to
data-to-text generation and text-to-text generation deep learning methods, as
well as new applications of NLG technology. This survey aims to (a) give the
latest synthesis of deep learning research on the NLG core tasks, as well as
the architectures adopted in the field; (b) detail meticulously and
comprehensively various NLG tasks and datasets, and draw attention to the
challenges in NLG evaluation, focusing on different evaluation methods and
their relationships; (c) highlight some future emphasis and relatively recent
research issues that arise due to the increasing synergy between NLG and other
artificial intelligence areas, such as computer vision, text and computational
creativity.Comment: Accepted by ACM Computing Survey (CSUR) 202
Automatic Personality Prediction; an Enhanced Method Using Ensemble Modeling
Human personality is significantly represented by those words which he/she
uses in his/her speech or writing. As a consequence of spreading the
information infrastructures (specifically the Internet and social media), human
communications have reformed notably from face to face communication.
Generally, Automatic Personality Prediction (or Perception) (APP) is the
automated forecasting of the personality on different types of human
generated/exchanged contents (like text, speech, image, video, etc.). The major
objective of this study is to enhance the accuracy of APP from the text. To
this end, we suggest five new APP methods including term frequency
vector-based, ontology-based, enriched ontology-based, latent semantic analysis
(LSA)-based, and deep learning-based (BiLSTM) methods. These methods as the
base ones, contribute to each other to enhance the APP accuracy through
ensemble modeling (stacking) based on a hierarchical attention network (HAN) as
the meta-model. The results show that ensemble modeling enhances the accuracy
of APP
On the primacy and irreducible nature of first-person versus third-person information
In\ua0this\ua0essay,\ua0we\ua0will\ua0support\ua0the\ua0claim\ua0that\ua0at\ua0the\ua0current\ua0level\ua0of\ua0scientific advancement\ua0a)\ua0some\ua0first-person\ua0accounts\ua0cannot\ua0be\ua0reduced\ua0to\ua0their third-person\ua0neural\ua0and\ua0psychophysiological\ua0correlates\ua0and\ua0b)\ua0that\ua0these first-person\ua0accounts\ua0are\ua0the\ua0only\ua0information\ua0to\ua0reckon\ua0when\ua0it\ua0is\ua0necessary\ua0to
analyse\ua0qualia\ua0contents.
Consequently,\ua0for\ua0many\ua0phenomena,\ua0first-person\ua0accounts\ua0are\ua0the\ua0only\ua0reliable source\ua0of\ua0information\ua0available\ua0and\ua0the\ua0knowledge\ua0of\ua0their\ua0neural\ua0and
psychophysical\ua0correlates\ua0don\u2019t\ua0offer\ua0any\ua0additional\ua0information\ua0about\ua0them
When Automated Assessment Meets Automated Content Generation: Examining Text Quality in the Era of GPTs
The use of machine learning (ML) models to assess and score textual data has
become increasingly pervasive in an array of contexts including natural
language processing, information retrieval, search and recommendation, and
credibility assessment of online content. A significant disruption at the
intersection of ML and text are text-generating large-language models such as
generative pre-trained transformers (GPTs). We empirically assess the
differences in how ML-based scoring models trained on human content assess the
quality of content generated by humans versus GPTs. To do so, we propose an
analysis framework that encompasses essay scoring ML-models, human and
ML-generated essays, and a statistical model that parsimoniously considers the
impact of type of respondent, prompt genre, and the ML model used for
assessment model. A rich testbed is utilized that encompasses 18,460
human-generated and GPT-based essays. Results of our benchmark analysis reveal
that transformer pretrained language models (PLMs) more accurately score human
essay quality as compared to CNN/RNN and feature-based ML methods.
Interestingly, we find that the transformer PLMs tend to score GPT-generated
text 10-15\% higher on average, relative to human-authored documents.
Conversely, traditional deep learning and feature-based ML models score human
text considerably higher. Further analysis reveals that although the
transformer PLMs are exclusively fine-tuned on human text, they more
prominently attend to certain tokens appearing only in GPT-generated text,
possibly due to familiarity/overlap in pre-training. Our framework and results
have implications for text classification settings where automated scoring of
text is likely to be disrupted by generative AI.Comment: Data available at:
https://github.com/nd-hal/automated-ML-scoring-versus-generatio
Conversational AI for Natural Language Processing: An Review of ChatGPT
ChatGPT is a conversational artificial intelligence model developed by OpenAI, which was introduced in 2019. It employs a transformer-based neural mesh to produce human being responses in real-time, allowing for natural language conversations with a machine. ChatGPT is instructed on huge quantities of data captured using the internet, making it knowledgeable in an extensive span of topics, from news & entertainment to politics and sports. This allows it to generate contextually relevant responses to questions and statements, making the conversation seem more lifelike. The model can be used in various applications, including customer service, personal assistants, and virtual assistants. ChatGPT has also shown promising results in generating creative content, such as jokes and poetry, showcasing its versatility and potential for future applications.This paper provides a comprehensive review of the existing literature on ChatGPT, highlighting its key advantages, such as improved accuracy and flexibility compared to traditional NLP tools, as well as its limitations and the need for further research to address potential ethical concerns. The review also highlights the potential for ChatGPT to be used in NLP applications, including question-answering and dialogue generation, and highlights the need for further research and development in these areas
- …