File size: 4,756 Bytes
3723383 55da2d9 d0443ec 55da2d9 3723383 55da2d9 d0443ec 55da2d9 d0443ec 55da2d9 d0443ec 55da2d9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 |
---
license: apache-2.0
base_model: kakaocorp/kanana-safeguard-8b
quantized_by: Arc1el
quantization_method: bitsandbytes
model_type: llama
tags:
- quantized
- 4bit
- bitsandbytes
- safeguard
- korean
- safety
pipeline_tag: text-generation
---
# Kanana Safeguard Siren 8B - 4bit μμν λ²μ
## λͺ¨λΈ κ°μ
- **μλ³Έ λͺ¨λΈ**: [kakaocorp/kanana-safeguard-8b](https://huggingface.co/kakaocorp/kanana-safeguard-8b)
- **μμν λ°©λ²**: BitsAndBytes 4bit (NF4)
- **μμν λꡬ**: bitsandbytes + transformers
- **μμΆλ₯ **: μλ³Έ λλΉ μ½ 75% ν¬κΈ° κ°μ (μΆμ )
## λͺ¨λΈ μΈλΆμ 보
### μλ³Έ λͺ¨λΈ μ 보
- **λͺ¨λΈ μν€ν
μ²**: Safeguard model based on transformer architecture
- **νλΌλ―Έν° μ**: 8B parameters
- **μ£Όμ μ©λ**: μμ μ± κ²μ¦, μ ν΄ μ½ν
μΈ νμ§
- **μΈμ΄**: νκ΅μ΄ μ€μ¬
- **μλ³Έ λͺ¨λΈ λΌμ΄μ μ€**: Apache 2.0
### μμν μ 보
- **μμν νμ
**: 4bit NormalFloat (NF4)
- **μ λ°λ**: 4bit weights, 16bit activations
## μ±λ₯ λ° λ²€μΉλ§ν¬
### λͺ¨λΈ ν¬κΈ° λΉκ΅
- **μλ³Έ λͺ¨λΈ**: ~16GB (μΆμ )
- **μμν λͺ¨λΈ**: ~4GB (μΆμ )
- **μμΆλ₯ **: μ½ 75% κ°μ
### λ©λͺ¨λ¦¬ μ¬μ©λ
- **λ‘λ© μ VRAM**: μ½ 4-5GB
- **μΆλ‘ μ VRAM**: μ½ 6-8GB (λ°°μΉ ν¬κΈ°μ λ°λΌ λ³λ)
- **μμ€ν
RAM**: μ΅μ 8GB κΆμ₯
## μ¬μ©λ²
### μ€μΉ
```bash
pip install transformers accelerate bitsandbytes torch
```
### μ½λ μμ
```python
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
model_name = "nxtcloud-org/kanana-safeguard-siren-8b-4bit"
# BitsAndBytesConfig μ€μ
bnb_config = BitsAndBytesConfig(
load_in_4bit=True,
bnb_4bit_use_double_quant=True,
bnb_4bit_quant_type="nf4",
bnb_4bit_compute_dtype=torch.bfloat16
)
# λͺ¨λΈκ³Ό ν ν¬λμ΄μ λ‘λ
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
quantization_config=bnb_config,
device_map="auto",
trust_remote_code=True
)
# μ¬μ© μμ - μμ μ± κ²μ¦
text = "μ΄κ²μ κ²μ¦ν ν
μ€νΈμ
λλ€."
inputs = tokenizer(text, return_tensors="pt")
with torch.no_grad():
outputs = model.generate(
**inputs,
max_length=512,
temperature=0.7,
do_sample=True,
pad_token_id=tokenizer.eos_token_id
)
result = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(result)
```
## μμ€ν
μꡬμ¬ν
- **μ΅μ RAM**: 8GB
- **κΆμ₯ RAM**: 16GB
- **GPU λ©λͺ¨λ¦¬**: 6GB VRAM (RTX 3060 μ΄μ)
- **μ§μ νλ«νΌ**: Linux, Windows (CUDA μ§μ GPU)
- **Python**: 3.8+
- **CUDA**: 11.1+
## μ νμ¬ν λ° κ³ λ €μ¬ν
### μ±λ₯ μ νμ¬ν
- 4bit μμνλ‘ μΈν λ―Έλ―Έν μ±λ₯ μ ν κ°λ₯
- μλ³Έ λͺ¨λΈ λλΉ μΆλ‘ νμ§μ΄ μ½κ° κ°μν μ μμ
- 볡μ‘ν μμ μ± νλ¨μμ λ ν° μν₯μ λ°μ μ μμ
### μ¬μ© κΆμ₯μ¬ν
- GPU λ©λͺ¨λ¦¬ μ μ½μ΄ μλ νκ²½μμ μ¬μ© κΆμ₯
- μ€μκ° μμ μ± κ²μ¦μ΄ νμν μ ν리μΌμ΄μ
μ μ ν©
- νλ‘λμ
νκ²½ μ¬μ© μ μΆ©λΆν κ²μ¦ νμ
### μ£Όμμ¬ν
- μ΄ λͺ¨λΈμ μμ μ± κ²μ¦ λͺ©μ μ safeguard λͺ¨λΈμ
λλ€
- μλ³Έ λͺ¨λΈμ μ±λ₯κ³Ό μμ μ± νΉμ±μ μμν νμλ μ μ§νλλ‘ λ
Έλ ₯νμΌλ, μμ ν λμΌνμ§ μμ μ μμ΅λλ€
- μ€μν μμ μ± νλ¨μλ μλ³Έ λͺ¨λΈκ³Όμ κ΅μ°¨ κ²μ¦μ κΆμ₯ν©λλ€
## μ€λ¦¬μ κ³ λ €μ¬ν
- μλ³Έ Kakao Corpμ safeguard λͺ¨λΈμ μ€λ¦¬μ κ°μ΄λλΌμΈμ μ€μν©λλ€
- μ΄ λͺ¨λΈμ μ ν΄ μ½ν
μΈ νμ§ λ° μμ μ± κ²μ¦ λͺ©μ μΌλ‘λ§ μ¬μ©λμ΄μΌ ν©λλ€
- μμνλ‘ μΈν μμμΉ λͺ»ν νΈν₯μ΄λ μμ μ± μ΄μκ° λ°μν μ μμΌλ―λ‘ μ£Όμκ° νμν©λλ€
## λΌμ΄μ μ€
μ΄ μμν λͺ¨λΈμ μλ³Έ λͺ¨λΈκ³Ό λμΌν **Apache License 2.0**μ λ°λ¦
λλ€.
```
Copyright 2025 Kakao Corp. (Original model)
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
```
## ν¬λ λ§ λ° μΈμ©
### μλ³Έ λͺ¨λΈ ν¬λ λ§
```bibtex
@misc{kakao-kanana-safeguard-siren-8b,
title={Kanana Safeguard Siren 8B},
author={Kakao Corp},
year={2024},
publisher={Hugging Face},
url={https://huggingface.co/kakaocorp/kanana-safeguard-8b}
}
``` |