Towards Understanding Large Language Models for Multilingual Semantic Encoding

Abstract

Natural Language Processing (NLP) has witnessed significant advancements with the emergence of large language models (LLM) capable of understanding and generating human-like text. However, there remains a critical need to explore and understand their efficiency and effectiveness, especially in processing languages beyond English. This study aims to evaluate the efficiency of various large language models in capturing semantic meaning across English, German, and Spanish sentences. Principal Component Analysis (PCA) is utilized to identify important weights for understanding semantics. Through further experimentation with various sentence structures, we aim to identify factors contributing to the effectiveness of certain models. By pinpointing the strengths and weaknesses of different models, we aim to advance NLP research for multilingual applications

Similar works

Full text

thumbnail-image

Illinois Mathematics and Science Academy: DigitalCommons@IMSA

redirect
Last time updated on 22/06/2025

Having an issue?

Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.