评估单语和多语大型语言模型在希腊语问答中的表现:DemosQA基准
📄 中文摘要
随着自然语言处理和深度学习的进步,大型语言模型(LLMs)在问答等多种任务中显著提升了技术水平。然而,现有研究主要集中在高资源语言(如英语)上,最近才开始关注多语言模型。这些模型往往存在训练数据偏向少数流行语言的问题,或依赖于从高资源语言向低资源语言的迁移学习,这可能导致社会、文化和历史方面的误解。为了解决这一挑战,针对低资源语言开发了单语大型语言模型,以提高其在特定语言环境下的表现,并推动问答系统的多样性和准确性。该研究通过DemosQA基准评估了这些模型在希腊语问答任务中的有效性。
📄 English Summary
Evaluating Monolingual and Multilingual Large Language Models for Greek Question Answering: The DemosQA Benchmark
Recent advancements in Natural Language Processing and Deep Learning have led to the development of Large Language Models (LLMs), which have significantly improved performance across various tasks, including Question Answering (QA). However, existing research has predominantly focused on high-resourced languages, such as English, with recent attention shifting towards multilingual models. These models often exhibit a training data bias towards a limited number of popular languages or rely on transfer learning from high-resourced to low-resourced languages, potentially misrepresenting social, cultural, and historical contexts. To address this issue, monolingual LLMs have been developed for under-resourced languages, aiming to enhance their performance in specific linguistic environments and promote diversity and accuracy in QA systems. This study evaluates the effectiveness of these models in Greek Question Answering tasks using the DemosQA benchmark.
Powered by Cloudflare Workers + Payload CMS + Claude 3.5
数据源: OpenAI, Google AI, DeepMind, AWS ML Blog, HuggingFace 等