|
|
--- |
|
|
license: apache-2.0 |
|
|
base_model: kakaocorp/kanana-safeguard-8b |
|
|
quantized_by: Arc1el |
|
|
quantization_method: bitsandbytes |
|
|
model_type: llama |
|
|
tags: |
|
|
- quantized |
|
|
- 4bit |
|
|
- bitsandbytes |
|
|
- safeguard |
|
|
- korean |
|
|
- safety |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
# Kanana Safeguard Siren 8B - 4bit μμν λ²μ |
|
|
|
|
|
## λͺ¨λΈ κ°μ |
|
|
- **μλ³Έ λͺ¨λΈ**: [kakaocorp/kanana-safeguard-8b](https://huggingface.co/kakaocorp/kanana-safeguard-8b) |
|
|
- **μμν λ°©λ²**: BitsAndBytes 4bit (NF4) |
|
|
- **μμν λꡬ**: bitsandbytes + transformers |
|
|
- **μμΆλ₯ **: μλ³Έ λλΉ μ½ 75% ν¬κΈ° κ°μ (μΆμ ) |
|
|
|
|
|
## λͺ¨λΈ μΈλΆμ 보 |
|
|
|
|
|
### μλ³Έ λͺ¨λΈ μ 보 |
|
|
- **λͺ¨λΈ μν€ν
μ²**: Safeguard model based on transformer architecture |
|
|
- **νλΌλ―Έν° μ**: 8B parameters |
|
|
- **μ£Όμ μ©λ**: μμ μ± κ²μ¦, μ ν΄ μ½ν
μΈ νμ§ |
|
|
- **μΈμ΄**: νκ΅μ΄ μ€μ¬ |
|
|
- **μλ³Έ λͺ¨λΈ λΌμ΄μ μ€**: Apache 2.0 |
|
|
|
|
|
### μμν μ 보 |
|
|
- **μμν νμ
**: 4bit NormalFloat (NF4) |
|
|
- **μ λ°λ**: 4bit weights, 16bit activations |
|
|
|
|
|
## μ±λ₯ λ° λ²€μΉλ§ν¬ |
|
|
|
|
|
### λͺ¨λΈ ν¬κΈ° λΉκ΅ |
|
|
- **μλ³Έ λͺ¨λΈ**: ~16GB (μΆμ ) |
|
|
- **μμν λͺ¨λΈ**: ~4GB (μΆμ ) |
|
|
- **μμΆλ₯ **: μ½ 75% κ°μ |
|
|
|
|
|
### λ©λͺ¨λ¦¬ μ¬μ©λ |
|
|
- **λ‘λ© μ VRAM**: μ½ 4-5GB |
|
|
- **μΆλ‘ μ VRAM**: μ½ 6-8GB (λ°°μΉ ν¬κΈ°μ λ°λΌ λ³λ) |
|
|
- **μμ€ν
RAM**: μ΅μ 8GB κΆμ₯ |
|
|
|
|
|
## μ¬μ©λ² |
|
|
|
|
|
### μ€μΉ |
|
|
```bash |
|
|
pip install transformers accelerate bitsandbytes torch |
|
|
``` |
|
|
|
|
|
### μ½λ μμ |
|
|
```python |
|
|
import torch |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig |
|
|
|
|
|
model_name = "nxtcloud-org/kanana-safeguard-siren-8b-4bit" |
|
|
|
|
|
# BitsAndBytesConfig μ€μ |
|
|
bnb_config = BitsAndBytesConfig( |
|
|
load_in_4bit=True, |
|
|
bnb_4bit_use_double_quant=True, |
|
|
bnb_4bit_quant_type="nf4", |
|
|
bnb_4bit_compute_dtype=torch.bfloat16 |
|
|
) |
|
|
|
|
|
# λͺ¨λΈκ³Ό ν ν¬λμ΄μ λ‘λ |
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
model_name, |
|
|
quantization_config=bnb_config, |
|
|
device_map="auto", |
|
|
trust_remote_code=True |
|
|
) |
|
|
|
|
|
# μ¬μ© μμ - μμ μ± κ²μ¦ |
|
|
text = "μ΄κ²μ κ²μ¦ν ν
μ€νΈμ
λλ€." |
|
|
inputs = tokenizer(text, return_tensors="pt") |
|
|
|
|
|
with torch.no_grad(): |
|
|
outputs = model.generate( |
|
|
**inputs, |
|
|
max_length=512, |
|
|
temperature=0.7, |
|
|
do_sample=True, |
|
|
pad_token_id=tokenizer.eos_token_id |
|
|
) |
|
|
|
|
|
result = tokenizer.decode(outputs[0], skip_special_tokens=True) |
|
|
print(result) |
|
|
``` |
|
|
|
|
|
## μμ€ν
μꡬμ¬ν |
|
|
- **μ΅μ RAM**: 8GB |
|
|
- **κΆμ₯ RAM**: 16GB |
|
|
- **GPU λ©λͺ¨λ¦¬**: 6GB VRAM (RTX 3060 μ΄μ) |
|
|
- **μ§μ νλ«νΌ**: Linux, Windows (CUDA μ§μ GPU) |
|
|
- **Python**: 3.8+ |
|
|
- **CUDA**: 11.1+ |
|
|
|
|
|
## μ νμ¬ν λ° κ³ λ €μ¬ν |
|
|
|
|
|
### μ±λ₯ μ νμ¬ν |
|
|
- 4bit μμνλ‘ μΈν λ―Έλ―Έν μ±λ₯ μ ν κ°λ₯ |
|
|
- μλ³Έ λͺ¨λΈ λλΉ μΆλ‘ νμ§μ΄ μ½κ° κ°μν μ μμ |
|
|
- 볡μ‘ν μμ μ± νλ¨μμ λ ν° μν₯μ λ°μ μ μμ |
|
|
|
|
|
### μ¬μ© κΆμ₯μ¬ν |
|
|
- GPU λ©λͺ¨λ¦¬ μ μ½μ΄ μλ νκ²½μμ μ¬μ© κΆμ₯ |
|
|
- μ€μκ° μμ μ± κ²μ¦μ΄ νμν μ ν리μΌμ΄μ
μ μ ν© |
|
|
- νλ‘λμ
νκ²½ μ¬μ© μ μΆ©λΆν κ²μ¦ νμ |
|
|
|
|
|
### μ£Όμμ¬ν |
|
|
- μ΄ λͺ¨λΈμ μμ μ± κ²μ¦ λͺ©μ μ safeguard λͺ¨λΈμ
λλ€ |
|
|
- μλ³Έ λͺ¨λΈμ μ±λ₯κ³Ό μμ μ± νΉμ±μ μμν νμλ μ μ§νλλ‘ λ
Έλ ₯νμΌλ, μμ ν λμΌνμ§ μμ μ μμ΅λλ€ |
|
|
- μ€μν μμ μ± νλ¨μλ μλ³Έ λͺ¨λΈκ³Όμ κ΅μ°¨ κ²μ¦μ κΆμ₯ν©λλ€ |
|
|
|
|
|
## μ€λ¦¬μ κ³ λ €μ¬ν |
|
|
- μλ³Έ Kakao Corpμ safeguard λͺ¨λΈμ μ€λ¦¬μ κ°μ΄λλΌμΈμ μ€μν©λλ€ |
|
|
- μ΄ λͺ¨λΈμ μ ν΄ μ½ν
μΈ νμ§ λ° μμ μ± κ²μ¦ λͺ©μ μΌλ‘λ§ μ¬μ©λμ΄μΌ ν©λλ€ |
|
|
- μμνλ‘ μΈν μμμΉ λͺ»ν νΈν₯μ΄λ μμ μ± μ΄μκ° λ°μν μ μμΌλ―λ‘ μ£Όμκ° νμν©λλ€ |
|
|
|
|
|
## λΌμ΄μ μ€ |
|
|
μ΄ μμν λͺ¨λΈμ μλ³Έ λͺ¨λΈκ³Ό λμΌν **Apache License 2.0**μ λ°λ¦
λλ€. |
|
|
|
|
|
``` |
|
|
Copyright 2025 Kakao Corp. (Original model) |
|
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License"); |
|
|
you may not use this file except in compliance with the License. |
|
|
You may obtain a copy of the License at |
|
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0 |
|
|
|
|
|
Unless required by applicable law or agreed to in writing, software |
|
|
distributed under the License is distributed on an "AS IS" BASIS, |
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
|
See the License for the specific language governing permissions and |
|
|
limitations under the License. |
|
|
``` |
|
|
|
|
|
## ν¬λ λ§ λ° μΈμ© |
|
|
|
|
|
### μλ³Έ λͺ¨λΈ ν¬λ λ§ |
|
|
```bibtex |
|
|
@misc{kakao-kanana-safeguard-siren-8b, |
|
|
title={Kanana Safeguard Siren 8B}, |
|
|
author={Kakao Corp}, |
|
|
year={2024}, |
|
|
publisher={Hugging Face}, |
|
|
url={https://huggingface.co/kakaocorp/kanana-safeguard-8b} |
|
|
} |
|
|
``` |