17 research outputs found

    YORC: Yoruba Reading Comprehension dataset

    Full text link
    In this paper, we create YORC: a new multi-choice Yoruba Reading Comprehension dataset that is based on Yoruba high-school reading comprehension examination. We provide baseline results by performing cross-lingual transfer using existing English RACE dataset based on a pre-trained encoder-only model. Additionally, we provide results by prompting large language models (LLMs) like GPT-4

    Forward-backward splitting algorithm with self-adaptive method for finite family of split minimization and fixed point problems in Hilbert spaces

    Get PDF
    In this paper, we introduce an inertial forward-backward splitting method together with a Halpern iterative algorithm for approximating a common solution of a finite family of split minimization problem involving two proper, lower semicontinuous and convex functions and fixed point problem of a nonexpansive mapping in real Hilbert spaces. Under suitable conditions, we proved that the sequence generated by our algorithm converges strongly to a solution of the aforementioned problems. The stepsizes studied in this paper are designed in such a way that they do not require the Lipschitz continuity condition on the gradient and prior knowledge of operator norm. Finally, we illustrate a numerical experiment to show the performance of the proposed method. The result discussed in this paper extends and complements many related results in literature

    Consultative engagement of stakeholders toward a roadmap for African language technologies

    Get PDF
    There has been a rise in natural language processing (NLP) communities across the African continent (Masakhane, AfricaNLP workshops). With this momentum noted, and given the existing power asymmetries that plague the African continent, there is an urgent need to ensure that these technologies move toward shared goals between organizations and stakeholders, not only to improve the representation of African languages in cutting-edge NLP research but also to ensure that NLP research enables technological advances toward human dignity, well-being, and equity for those who speak African languages. This study investigates the motivations, focus, and challenges faced by various stakeholders who are at the core of the NLP process. We perform structured stakeholder identification to identify core stakeholders in the NLP process. Interviews with representatives of these stakeholder groups are performed and are collated into relevant themes. Finally, a set of recommendations are proposed for use by policy and artificial intelligence (AI) researchers

    MasakhaNEWS: News Topic Classification for African languages

    Full text link
    African languages are severely under-represented in NLP research due to lack of datasets covering several NLP tasks. While there are individual language specific datasets that are being expanded to different tasks, only a handful of NLP tasks (e.g. named entity recognition and machine translation) have standardized benchmark datasets covering several geographical and typologically-diverse African languages. In this paper, we develop MasakhaNEWS -- a new benchmark dataset for news topic classification covering 16 languages widely spoken in Africa. We provide an evaluation of baseline models by training classical machine learning models and fine-tuning several language models. Furthermore, we explore several alternatives to full fine-tuning of language models that are better suited for zero-shot and few-shot learning such as cross-lingual parameter-efficient fine-tuning (like MAD-X), pattern exploiting training (PET), prompting language models (like ChatGPT), and prompt-free sentence transformer fine-tuning (SetFit and Cohere Embedding API). Our evaluation in zero-shot setting shows the potential of prompting ChatGPT for news topic classification in low-resource African languages, achieving an average performance of 70 F1 points without leveraging additional supervision like MAD-X. In few-shot setting, we show that with as little as 10 examples per label, we achieved more than 90\% (i.e. 86.0 F1 points) of the performance of full supervised training (92.6 F1 points) leveraging the PET approach.Comment: Accepted to IJCNLP-AACL 2023 (main conference

    AfriQA:Cross-lingual Open-Retrieval Question Answering for African Languages

    Get PDF
    African languages have far less in-language content available digitally, making it challenging for question answering systems to satisfy the information needs of users. Cross-lingual open-retrieval question answering (XOR QA) systems -- those that retrieve answer content from other languages while serving people in their native language -- offer a means of filling this gap. To this end, we create AfriQA, the first cross-lingual QA dataset with a focus on African languages. AfriQA includes 12,000+ XOR QA examples across 10 African languages. While previous datasets have focused primarily on languages where cross-lingual QA augments coverage from the target language, AfriQA focuses on languages where cross-lingual answer content is the only high-coverage source of answer content. Because of this, we argue that African languages are one of the most important and realistic use cases for XOR QA. Our experiments demonstrate the poor performance of automatic translation and multilingual retrieval methods. Overall, AfriQA proves challenging for state-of-the-art QA models. We hope that the dataset enables the development of more equitable QA technology

    AfriMTE and AfriCOMET : Empowering COMET to Embrace Under-resourced African Languages

    Get PDF
    Despite the progress we have recorded in scaling multilingual machine translation (MT) models and evaluation data to several under-resourced African languages, it is difficult to measure accurately the progress we have made on these languages because evaluation is often performed on n-gram matching metrics like BLEU that often have worse correlation with human judgments. Embedding-based metrics such as COMET correlate better; however, lack of evaluation data with human ratings for under-resourced languages, complexity of annotation guidelines like Multidimensional Quality Metrics (MQM), and limited language coverage of multilingual encoders have hampered their applicability to African languages. In this paper, we address these challenges by creating high-quality human evaluation data with a simplified MQM guideline for error-span annotation and direct assessment (DA) scoring for 13 typologically diverse African languages. Furthermore, we develop AfriCOMET, a COMET evaluation metric for African languages by leveraging DA training data from high-resource languages and African-centric multilingual encoder (AfroXLM-Roberta) to create the state-of-the-art evaluation metric for African languages MT with respect to Spearman-rank correlation with human judgments (+0.406)

    MasakhaNEWS:News Topic Classification for African languages

    Get PDF
    African languages are severely under-represented in NLP research due to lack of datasets covering several NLP tasks. While there are individual language specific datasets that are being expanded to different tasks, only a handful of NLP tasks (e.g. named entity recognition and machine translation) have standardized benchmark datasets covering several geographical and typologically-diverse African languages. In this paper, we develop MasakhaNEWS -- a new benchmark dataset for news topic classification covering 16 languages widely spoken in Africa. We provide an evaluation of baseline models by training classical machine learning models and fine-tuning several language models. Furthermore, we explore several alternatives to full fine-tuning of language models that are better suited for zero-shot and few-shot learning such as cross-lingual parameter-efficient fine-tuning (like MAD-X), pattern exploiting training (PET), prompting language models (like ChatGPT), and prompt-free sentence transformer fine-tuning (SetFit and Cohere Embedding API). Our evaluation in zero-shot setting shows the potential of prompting ChatGPT for news topic classification in low-resource African languages, achieving an average performance of 70 F1 points without leveraging additional supervision like MAD-X. In few-shot setting, we show that with as little as 10 examples per label, we achieved more than 90\% (i.e. 86.0 F1 points) of the performance of full supervised training (92.6 F1 points) leveraging the PET approach

    Quality Assessment of Maritime AIS data

    No full text
    This thesis includes a quality assessment investigation of Automatic Identification System (AIS) data retrieved through the ARPA project data platform from Digitraffic. Automatic Identification System (AIS) data is essential in improving the global shipping industry's safety, efficiency, environmental performance, and operations. The dataset includes location, navigational, and static data from thousands of ships from the Baltic Sea geographical region. The research examines the literature on AIS data quality through which an assessment concept was constructed. This mixed-method approach combines qualitative and quantitative data analysis techniques to identify the elements that influence the data's quality and develop strategies for measuring it. The investigation focuses on four critical aspects of data quality: accuracy, completeness, consistency, and timeliness. The findings show that AIS technology, communication protocols, ambient conditions, and human variables all impact the quality of marine AIS data. Therefore, to address these issues, the dissertation provides a set of quality indicators and data validation procedures. The effectiveness of the quality assessment procedures in identifying AIS data quality concerns was demonstrated. According to the study, ongoing monitoring and improvement of AIS data quality are still required to improve marine safety and decision-making, ultimately making it ideal for autonomous shipping where the data is needed with a high degree of integrity
    corecore