multilingual AI

  • Bielik-11B-v3.0-Instruct: A Multilingual AI Model


    Bielik-11B-v3.0-InstructBielik-11B-v3.0-Instruct is a sophisticated generative text model with 11 billion parameters, fine-tuned from its base version, Bielik-11B-v3-Base-20250730. This model is a product of the collaboration between the open-science project SpeakLeash and the High Performance Computing center ACK Cyfronet AGH. It has been developed using multilingual text corpora from 32 European languages, with a special focus on Polish, processed by the SpeakLeash team. The project utilizes the Polish PLGrid computing infrastructure, particularly the HPC centers at ACK Cyfronet AGH, highlighting the importance of large-scale computational resources in advancing AI technologies. This matters because it showcases the potential of collaborative efforts in enhancing AI capabilities and the role of national infrastructure in supporting such advancements.

    Read Full Article: Bielik-11B-v3.0-Instruct: A Multilingual AI Model

  • K-EXAONE: Multilingual AI Model by LG AI Research


    LGAI-EXAONE/K-EXAONE-236B-A23B · Hugging FaceK-EXAONE, developed by LG AI Research, is a large-scale multilingual language model featuring a Mixture-of-Experts architecture with 236 billion parameters, 23 billion of which are active during inference. It excels in reasoning, agentic capabilities, and multilingual understanding across six languages, utilizing a 256K context window to efficiently process long documents. The model's architecture is optimized with Multi-Token Prediction, enhancing inference throughput by 1.5 times, and it incorporates Korean cultural contexts to ensure alignment with universal human values. K-EXAONE demonstrates high reliability and safety, making it a robust tool for diverse applications. This matters because it represents a significant advancement in multilingual AI, offering enhanced efficiency and cultural sensitivity in language processing.

    Read Full Article: K-EXAONE: Multilingual AI Model by LG AI Research

  • Plamo3 Support Merged into llama.cpp


    Plamo3 (2B/8B/31B) support has been merged into llama.cppPLaMo 3 NICT 31B Base is a sophisticated language model developed through a collaboration between Preferred Networks, Inc. and the National Institute of Information and Communications Technology (NICT). It is pre-trained on both English and Japanese datasets, showcasing a hybrid architecture that combines Sliding Window Attention (SWA) with traditional attention layers. This integration into llama.cpp signifies an advancement in multilingual model capabilities, enhancing the potential for more nuanced and context-aware language processing. This matters because it represents a significant step forward in creating more versatile and powerful language models that can handle complex linguistic tasks across multiple languages.

    Read Full Article: Plamo3 Support Merged into llama.cpp