The other DICE contribution at #COLING2025 comes from Nikit, who presented "LOLA - An Open-Source Massively Multilingual Large Language Mode" by Nikit Srivastava, Denis Kuchelev, Tatiana Moteu Ngoli, Kshitij Shetty, Michael Röder, @hamadazahera, Diego Moussallem & Axel Ngonga.🤩 👏

👉 Want to find out more? Find the paper here: https://arxiv.org/abs/2409.11272

#DICEontour #LowResourceLanguages

LOLA -- An Open-Source Massively Multilingual Large Language Model

This paper presents LOLA, a massively multilingual large language model trained on more than 160 languages using a sparse Mixture-of-Experts Transformer architecture. Our architectural and implementation choices address the challenge of harnessing linguistic diversity while maintaining efficiency and avoiding the common pitfalls of multilinguality. Our analysis of the evaluation results shows competitive performance in natural language generation and understanding tasks. Additionally, we demonstrate how the learned expert-routing mechanism exploits implicit phylogenetic linguistic patterns to potentially alleviate the curse of multilinguality. We provide an in-depth look at the training process, an analysis of the datasets, and a balanced exploration of the model's strengths and limitations. As an open-source model, LOLA promotes reproducibility and serves as a robust foundation for future research. Our findings enable the development of compute-efficient multilingual models with strong, scalable performance across languages.

arXiv.org

First conference in 2025!🤩 Many greetings to Daniel, who presented ‘Contextual Augmentation for Entity Linking using Large Language Models’ by Daniel Vollmers, @hamadazahera, Diego Moussallem and Axel Ngonga at #COLING2025 in Abu Dhabi🇦🇪 this week.👏👨‍💻 #DICEontour

👀 Would you like to find out more about the paper? Take a look here: https://papers.dice-research.org/2025/COLING_EL_Augmentation/public.pdf

Iker de la Iglesiak medikuntzako argumentuen ebaluazio automatikoari buruzko posterra aurkeztu du #COLING2025 konferentzian
#COLING2025 konferentzian gaude

Sporting my #ZHAWDigital bottle at #COLING2025. Now keynote by Katrin Erk on „Word meaning, in computational linguistics and beyond“

!B

Proceedings for the Workshop on Generative AI and Knowledge Graph (GenAIK) 2025 co-located with #coling2025 are available via the COLING website:

https://coling-2025-proceedings.s3.us-east-1.amazonaws.com/workshops/GenAIK/program.html

#generativeAI #AI #knowledgegraphs #llms #semanticweb #neurosymbolicAI @fiz_karlsruhe @fizise @albertmeronyo @GenAsefa @GenAIK

Program - Workshop on Generative AI and Knowledge Graphs (GenAIK)

I now saw several talks addressing creativity when „writing“ with LLMs, either as main topic or a contributing one. All fail (or imply refuse) to define or operationalize „creativity.“ They mainly retreat to human annotation/assessment by applying „Humans know it when they see it.“ Which, hm, never mind

#COLING2025 #WritingResearch

Talking about biases

„So you see, COLING traveled around the whole globe during the last 60 years“

Hmmm, the northern hemisphere only, actually, no?

#COLING2025

I am at #COLING2025 and they don’t have stickers!

Yesterday, the GenAIK 2025 workshop on the integration of generative AI and knowledge graphs took place in Abu Dhabi co-located with #coling2025

#GenAIK2025 website: https://genetasefa.github.io/GenAIK2025/

#generativeAI #knowledgegraphs #AI #semanticweb #ontologies #neuroAI #neurosymbolic @GenAsefa @GenAIK @albertmeronyo @lihuchen @fiz_karlsruhe @lysander07 @NFDI4DS

Workshop on Generative AI and Knowledge Graphs (GenAIK)