3,303 research outputs found

    Cross-lingual Data Augmentation for Document-grounded Dialog Systems in Low Resource Languages

    Full text link
    This paper proposes a framework to address the issue of data scarcity in Document-Grounded Dialogue Systems(DGDS). Our model leverages high-resource languages to enhance the capability of dialogue generation in low-resource languages. Specifically, We present a novel pipeline CLEM (Cross-Lingual Enhanced Model) including adversarial training retrieval (Retriever and Re-ranker), and Fid (fusion-in-decoder) generator. To further leverage high-resource language, we also propose an innovative architecture to conduct alignment across different languages with translated training. Extensive experiment results demonstrate the effectiveness of our model and we achieved 4th place in the DialDoc 2023 Competition. Therefore, CLEM can serve as a solution to resource scarcity in DGDS and provide useful guidance for multi-lingual alignment tasks

    A Multilingual Chat System with Image Presentation for Detecting Mistranslation

    Get PDF
    We have designed and developed a multilingual chat system, MCHI (Multilingual Chat with Hint Images), which is based on machine translation and equipped with a presentation function of images related to the contents of the messages by utterers so that listeners are able to notice mistranslation. MCHI accepts English, French, Chinese, Japanese, Korean and Vietnamese languages. It uses the Google API to retrieve related images from the image posting site Flickr. As a result of evaluation experiment, we have observed that participants detected the mismatch of a translated message with its related image. According to the answers of participants for a questionnaire, it turned out that the usability of the MCHI system is good enough though the related images are not satisfactory

    The Impact of Artificial Intelligence on the Evolution of Digital Education: A Comparative Study of OpenAI Text Generation Tools including ChatGPT, Bing Chat, Bard, and Ernie

    Full text link
    In the digital era, the integration of artificial intelligence (AI) in education has ushered in transformative changes, redefining teaching methodologies, curriculum planning, and student engagement. This review paper delves deep into the rapidly evolving landscape of digital education by contrasting the capabilities and impact of OpenAI's pioneering text generation tools like Bing Chat, Bard, Ernie with a keen focus on the novel ChatGPT. Grounded in a typology that views education through the lenses of system, process, and result, the paper navigates the multifaceted applications of AI. From decentralizing global education and personalizing curriculums to digitally documenting competence-based outcomes, AI stands at the forefront of educational modernization. Highlighting ChatGPT's meteoric rise to one million users in just five days, the study underscores its role in democratizing education, fostering autodidacticism, and magnifying student engagement. However, with such transformative power comes the potential for misuse, as text-generation tools can inadvertently challenge academic integrity. By juxtaposing the promise and pitfalls of AI in education, this paper advocates for a harmonized synergy between AI tools and the educational community, emphasizing the urgent need for ethical guidelines, pedagogical adaptations, and strategic collaborations

    Can ChatGPT pass the Vietnamese National High School Graduation Examination?

    Full text link
    This research article highlights the potential of AI-powered chatbots in education and presents the results of using ChatGPT, a large language model, to complete the Vietnamese National High School Graduation Examination (VNHSGE). The study dataset included 30 essays in the literature test case and 1,700 multiple-choice questions designed for other subjects. The results showed that ChatGPT was able to pass the examination with an average score of 6-7, demonstrating the technology's potential to revolutionize the educational landscape. The analysis of ChatGPT performance revealed its proficiency in a range of subjects, including mathematics, English, physics, chemistry, biology, history, geography, civic education, and literature, which suggests its potential to provide effective support for learners. However, further research is needed to assess ChatGPT performance on more complex exam questions and its potential to support learners in different contexts. As technology continues to evolve and improve, we can expect to see the use of AI tools like ChatGPT become increasingly common in educational settings, ultimately enhancing the educational experience for both students and educators.Comment: 9 pages, 13 figures, 4 table

    Typological parameters of genericity

    Get PDF
    Different languages employ different morphosyntactic devices for expressing genericity. And, of course, they also make use of different morphosyntactic and semantic or pragmatic cues which may contribute to the interpretation of a sentence as generic rather than episodic. [...] We will advance the strong hypo thesis that it is a fundamental property of lexical elements in natural language that they are neutral with respect to different modes of reference or non-reference. That is, we reject the idea that a certain use of a lexical element, e.g. a use which allows reference to particular spatio-temporally bounded objects in the world, should be linguistically prior to all other possible uses, e.g. to generic and non-specific uses. From this it follows that we do not consider generic uses as derived from non-generic uses as it is occasionally assumed in the literature. Rather, we regard these two possibilities of use as equivalent alternative uses of lexical elements. The typological differences to be noted therefore concern the formal and semantic relationship of generic and non-generic uses to each other; they do not pertain to the question of whether lexical elements are predetermined for one of these two uses. Even supposing we found a language where generic uses are always zero-marked and identical to lexical sterns, we would still not assume that lexical elements in this language primarily have a generic use from which the non-generic uses are derived. (Incidentally, none of the languages examined, not even Vietnamese, meets this criterion.

    Modality markers and politeness strategies in British and American ambassadorial speeches: A corpus-based approach

    Get PDF
    This study investigates modality markers used as expressions of politeness in British and American ambassadorial speeches via a corpus-based method. Results of the research reflect the semantic and pragmatic perspectives of modality markers on the theories of modality and politeness. Although modality and politeness are the central topics in a wide range of studies, the two domains have been discussed separately and their relationship has not been empirically investigated. Moreover, there has been no study on modality markers in British and American discourse, nor has the use of modality markers in British and American ambassadorial speeches been examined. Therefore, this research examines the relation of modality to politeness via the use of British and American ambassadorial speeches. The research contributes to the practice of the discourse community with the analysis of modality markers as politeness strategies in ambassadorial speeches. The results of a comparative analysis of modality markers as speakers’ politeness strategies collected in ambassadorial speeches reveal that American and British ambassadors are strikingly different in their frequency of modality markers expressing particular politeness categories. American ambassadors use more modality markers expressing positive politeness strategies such as paying attention to hearers, expressing strong commitment, hedging on hearers’ positive face, expressing optimism, complimenting to mitigate the force of comments, making claims and minimising the imposition of face-threatening acts. British ambassadors, however, employ more modality markers expressing negative politeness strategies such as hedging on negative face-threatening acts, expressing hypotheses, expressing humbleness and mitigating the force of obligation. Therefore, this thesis claims that American ambassadors use more modality markers expressing positive politeness in terms of personal emotions and directness, while British ambassadors prefer modality markers expressing negative politeness such as tentativeness, indirectness and mitigation. It is noted that modality is represented in a range of syntactic structures and patterns other than single modal auxiliary verbs. In addition, since modality markers as expressions of politeness are culture-specific, the use of modality markers differs from one culture to another. Moreover, modality markers cannot totally be treated as hedges in intercultural communication since some modality markers which seem to be semantically similar between languages are actually different in their pragmatic functions among different cultures

    GAIA: a benchmark for General AI Assistants

    Full text link
    We introduce GAIA, a benchmark for General AI Assistants that, if solved, would represent a milestone in AI research. GAIA proposes real-world questions that require a set of fundamental abilities such as reasoning, multi-modality handling, web browsing, and generally tool-use proficiency. GAIA questions are conceptually simple for humans yet challenging for most advanced AIs: we show that human respondents obtain 92\% vs. 15\% for GPT-4 equipped with plugins. This notable performance disparity contrasts with the recent trend of LLMs outperforming humans on tasks requiring professional skills in e.g. law or chemistry. GAIA's philosophy departs from the current trend in AI benchmarks suggesting to target tasks that are ever more difficult for humans. We posit that the advent of Artificial General Intelligence (AGI) hinges on a system's capability to exhibit similar robustness as the average human does on such questions. Using GAIA's methodology, we devise 466 questions and their answer. We release our questions while retaining answers to 300 of them to power a leader-board available at https://huggingface.co/gaia-benchmark

    CloChat: Understanding How People Customize, Interact, and Experience Personas in Large Language Models

    Full text link
    Large language models (LLMs) have facilitated significant strides in generating conversational agents, enabling seamless, contextually relevant dialogues across diverse topics. However, the existing LLM-driven conversational agents have fixed personalities and functionalities, limiting their adaptability to individual user needs. Creating personalized agent personas with distinct expertise or traits can address this issue. Nonetheless, we lack knowledge of how people customize and interact with agent personas. In this research, we investigated how users customize agent personas and their impact on interaction quality, diversity, and dynamics. To this end, we developed CloChat, an interface supporting easy and accurate customization of agent personas in LLMs. We conducted a study comparing how participants interact with CloChat and ChatGPT. The results indicate that participants formed emotional bonds with the customized agents, engaged in more dynamic dialogues, and showed interest in sustaining interactions. These findings contribute to design implications for future systems with conversational agents using LLMs.Comment: In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '24
    corecore