Update README.md
Browse files
README.md
CHANGED
|
@@ -38,7 +38,7 @@ from transformers import pipeline
|
|
| 38 |
# Initialize the text generation pipeline
|
| 39 |
pipe = pipeline(
|
| 40 |
"text-generation",
|
| 41 |
-
model="mshojaei77/gemma-2-2b-fa-v2",
|
| 42 |
model_kwargs={"torch_dtype": torch.bfloat16},
|
| 43 |
device="cuda", # Or "mps" for Macs with Apple Silicon
|
| 44 |
)
|
|
@@ -53,6 +53,8 @@ outputs = pipe(messages, max_new_tokens=512, chat_template="gemma") # Explicitly
|
|
| 53 |
assistant_response = outputs[0]["generated_text"][-1]["content"].strip()
|
| 54 |
|
| 55 |
print(assistant_response)
|
|
|
|
|
|
|
| 56 |
```
|
| 57 |
|
| 58 |
**Key Usage Notes:**
|
|
|
|
| 38 |
# Initialize the text generation pipeline
|
| 39 |
pipe = pipeline(
|
| 40 |
"text-generation",
|
| 41 |
+
model="mshojaei77/gemma-2-2b-fa-v2",
|
| 42 |
model_kwargs={"torch_dtype": torch.bfloat16},
|
| 43 |
device="cuda", # Or "mps" for Macs with Apple Silicon
|
| 44 |
)
|
|
|
|
| 53 |
assistant_response = outputs[0]["generated_text"][-1]["content"].strip()
|
| 54 |
|
| 55 |
print(assistant_response)
|
| 56 |
+
# Example Output (Illustrative - Output quality may vary significantly):
|
| 57 |
+
# سلام! من خوبم، ممنون. شما چطوری؟ 😊
|
| 58 |
```
|
| 59 |
|
| 60 |
**Key Usage Notes:**
|