Author
Listed:
- Alexandros Tassios
(School of Informatics, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece)
- Stergios Tegos
(School of Informatics, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece)
- Christos Bouas
(School of Informatics, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece)
- Konstantinos Manousaridis
(School of Informatics, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece)
- Maria Papoutsoglou
(School of Informatics, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece)
- Maria Kaltsa
(Department of Theoretical & Applied Linguistics, School of English, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece
Information Technologies Institute, Center for Research and Technology Hellas, 57001 Thessaloniki, Greece)
- Eleni Dimopoulou
(PRAKSIS, 10432 Athens, Greece)
- Thanassis Mavropoulos
(Information Technologies Institute, Center for Research and Technology Hellas, 57001 Thessaloniki, Greece)
- Stefanos Vrochidis
(Information Technologies Institute, Center for Research and Technology Hellas, 57001 Thessaloniki, Greece)
- Georgios Meditskos
(School of Informatics, Aristotle University of Thessaloniki, 54124 Thessaloniki, Greece)
Abstract
The integration of Large Language Models (LLMs) in chatbot applications gains momentum. However, to successfully deploy such systems, the underlying capabilities of LLMs must be carefully considered, especially when dealing with low-resource languages and specialized fields. This paper presents the results of a comprehensive evaluation of several LLMs conducted in the context of a chatbot agent designed to assist migrants in their integration process. Our aim is to identify the optimal LLM that can effectively process and generate text in Greek and provide accurate information, addressing the specific needs of migrant populations. The design of the evaluation methodology leverages input from experts on social assistance initiatives, social impact and technological solutions, as well as from automated LLM self-evaluations. Given the linguistic challenges specific to the Greek language and the application domain, research findings indicate that Claude 3.7 Sonnet and Gemini 2.0 Flash demonstrate superior performance across all criteria, with Claude 3.7 Sonnet emerging as the leading candidate for the chatbot. Moreover, the results suggest that automated custom evaluations of LLMs can align with human assessments, offering a viable option for preliminary low-cost analysis to assist stakeholders in selecting the optimal LLM based on user and application domain requirements.
Suggested Citation
Alexandros Tassios & Stergios Tegos & Christos Bouas & Konstantinos Manousaridis & Maria Papoutsoglou & Maria Kaltsa & Eleni Dimopoulou & Thanassis Mavropoulos & Stefanos Vrochidis & Georgios Meditsko, 2025.
"LLM Performance in Low-Resource Languages: Selecting an Optimal Model for Migrant Integration Support in Greek,"
Future Internet, MDPI, vol. 17(6), pages 1-24, May.
Handle:
RePEc:gam:jftint:v:17:y:2025:i:6:p:235-:d:1664873
Download full text from publisher
References listed on IDEAS
- Christopher J. Lynch & Erik J. Jensen & Virginia Zamponi & Kevin O’Brien & Erika Frydenlund & Ross Gore, 2023.
"A Structured Narrative Prompt for Prompting Narratives from Large Language Models: Sentiment Assessment of ChatGPT-Generated Narratives and Real Tweets,"
Future Internet, MDPI, vol. 15(12), pages 1-36, November.
- Goran Bubaš & Antonela Čižmešija & Andreja Kovačić, 2023.
"Development of an Assessment Scale for Measurement of Usability and User Experience Characteristics of Bing Chat Conversational AI,"
Future Internet, MDPI, vol. 16(1), pages 1-19, December.
Full references (including those not matched with items on IDEAS)
Most related items
These are the items that most often cite the same works as this one and are cited by the same works as this one.
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jftint:v:17:y:2025:i:6:p:235-:d:1664873. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.