Gemma-Kimu-9b v1.0 is a continually pre-trained large language model (LLM) for the Basque language, built upon Google’s Gemma-2-9b foundational model.
This model focuses solely on language adaptation, without instruction-following alignment, serving as the base model for subsequent instruct-tuned versions such as Gemma-Kimu-9b-it.

We continually pre-train the Gemma-2-9b model on a combination of Basque monolingual data and English replay to improve Basque linguistic capabilities while preserving English performance.
This phase enhances the model’s syntactic, lexical, and morphological competence in Basque and establishes a solid foundation for downstream instruction-tuned and task-specific models.

Evaluations show that Gemma-Kimu-9b exhibits significant improvements over the original Gemma-2-9b in Basque language understanding, coherence, and text generation fluency.

Training Data

For continual pre-training, we leveraged a combination of Basque and English data to enhance linguistic performance in Basque while maintaining general English capabilities. The goal is to improve cross-lingual transfer by retaining the model's proficiency in English.

ZelaiHandi ZelaiHandi dataset (San Vicente et al., 2024): ZelaiHandi is the largest collection of freely licensed and high-quality Basque texts gathered from selected web sources. This collection comprises approximately 521 million words which correspond to 1.5 billion tokens (Llama 3.1 tokenizer).

FineWeb dataset (Penedo et al., 2024): FineWeb consists of more than 15T tokens of cleaned and deduplicated English web data from CommonCrawl. We selected a random subset of around 300 million tokens (Llama 3.1 tokenizer)

License

This model is derived from Gemma 2 and is licensed under the Gemma License. Copyright © Google DeepMind. All Rights Reserved.

Acknowledgments

This work is part of the BasqueLLM project, titled "bi-SLM: Optimization of Industrial Processes through Bilingual SLMs" (EXP: 2025-CIE4-000048-01), partially funded by the Guipuzcoa Science, Technology and Innovation Network Program of the Provincial Council of Gipuzkoa. Model training and development were conducted using the Hyperion system at the Donostia International Physics Center (DIPC).

Citation

If you use this model, please cite:

@inproceedings{sarasua2025,
  title={DIPLomA: Efficient Adaptation of Instructed LLMs to Low-Resource Languages via Post-Training Delta Merging},
  author={Sarasua, Ixak and Corral, Ander and Saralegi, Xabier},
  booktitle={Findings of the Association for Computational Linguistics: EMNLP 2025},
  year={2025}
}

Contact

Downloads last month
12
Safetensors
Model size
9B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for orai-nlp/Gemma-Kimu-9b-base

Base model

google/gemma-2-9b
Finetuned
(347)
this model

Datasets used to train orai-nlp/Gemma-Kimu-9b-base

Collection including orai-nlp/Gemma-Kimu-9b-base