Update README.md
Browse files
README.md
CHANGED
|
@@ -9,14 +9,64 @@ tags:
|
|
| 9 |
license: apache-2.0
|
| 10 |
language:
|
| 11 |
- en
|
|
|
|
|
|
|
|
|
|
| 12 |
---
|
| 13 |
|
| 14 |
-
#
|
| 15 |
|
| 16 |
-
- **
|
| 17 |
-
- **License:** apache-2.0
|
| 18 |
-
- **Finetuned from model :** unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit
|
| 19 |
|
| 20 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
|
| 22 |
-
|
|
|
|
| 9 |
license: apache-2.0
|
| 10 |
language:
|
| 11 |
- en
|
| 12 |
+
- tl
|
| 13 |
+
datasets:
|
| 14 |
+
- Linggowiktiks/AnoNa
|
| 15 |
---
|
| 16 |
|
| 17 |
+
# 🦙 Liyama-3B
|
| 18 |
|
| 19 |
+
**Liyama-3B** is a fine-tuned version of Meta’s LLaMA-3B (3.2) model, built to understand and respond fluently in **Tagalog**. It was trained on the **AnoNa** dataset over **3 epochs**, aiming for natural, context-aware instruction-following in Filipino.
|
|
|
|
|
|
|
| 20 |
|
| 21 |
+
---
|
| 22 |
+
|
| 23 |
+
## 🔤 Origin of the Name
|
| 24 |
+
The name **Liyama** is a Tagalified version of *llama*, reflecting both its LLaMA base and its Tagalog-focused language capabilities. It mirrors how Filipino often adapts foreign terms into familiar, phonetic forms—like *camera → kamera*, *lion → leon*, and now, *llama → liyama*.
|
| 25 |
+
|
| 26 |
+
---
|
| 27 |
+
|
| 28 |
+
## 🧠 Training Data: The AnoNa Dataset
|
| 29 |
+
|
| 30 |
+
Liyama-3B was trained solely on **response completions** from the **AnoNa** dataset — a self-instruct corpus generated using **Gemini 1.5** and **2.0**.
|
| 31 |
+
|
| 32 |
+
Inspired by **SimpleQnA**, the dataset contains short, helpful instruction-response pairs. But **AnoNa** introduces several improvements:
|
| 33 |
+
|
| 34 |
+
- ✅ **Less English, More Tagalog** prompts
|
| 35 |
+
- ✅ **Less IFEVAL-style formatting**
|
| 36 |
+
- ✅ **No overuse of modifiers** in instructions
|
| 37 |
+
- ✅ **Balanced task types** to avoid dominant categories
|
| 38 |
+
- ✅ **Complex tasks favored** (65% complex / 35% simple)
|
| 39 |
+
- ✅ **Reduced sycophancy** and generic praise
|
| 40 |
+
- ✅ **Improved follow-up handling**
|
| 41 |
+
- ✅ **AI self-intro appears only when relevant**
|
| 42 |
+
- ✅ **Implicit chain-of-thought reasoning**, not labeled
|
| 43 |
+
- ✅ **Extra task types** added to increase variety
|
| 44 |
+
|
| 45 |
+
This focus creates a model that's practical, straightforward, and tuned for **realistic conversational use in Filipino**, without excessive formatting or irrelevant disclaimers.
|
| 46 |
+
|
| 47 |
+
---
|
| 48 |
+
|
| 49 |
+
## 🗣️ Use Case
|
| 50 |
+
|
| 51 |
+
Liyama-3B is ideal for:
|
| 52 |
+
- Answering questions in Tagalog
|
| 53 |
+
- Writing essays, reflections, and letters in Filipino
|
| 54 |
+
- Following natural instructions, even when mixed with English
|
| 55 |
+
- Chat-based tasks where fluency and tone matter
|
| 56 |
+
- Educational or community apps centered around local language use
|
| 57 |
+
|
| 58 |
+
---
|
| 59 |
+
|
| 60 |
+
## 📦 Model Details
|
| 61 |
+
|
| 62 |
+
| Feature | Value |
|
| 63 |
+
|--------------------|----------------------------|
|
| 64 |
+
| Base Model | LLaMA-3B v3.2 |
|
| 65 |
+
| Fine-tuned Dataset | AnoNa |
|
| 66 |
+
| Epochs | 3 |
|
| 67 |
+
| Language Focus | Tagalog (with some English)|
|
| 68 |
+
| Prompt Format | Responses only |
|
| 69 |
+
|
| 70 |
+
---
|
| 71 |
|
| 72 |
+
Liyama-3B is part of a broader effort to create open, practical Filipino-language models for real use—not just benchmarks. Expect follow-ups tuned for multi-turn chat, reasoning, and creative tasks.
|