update inference with transformers
#162
by
AzizBelaweid
- opened
README.md
CHANGED
|
@@ -51,9 +51,8 @@ print(result)
|
|
| 51 |
```py
|
| 52 |
from transformers import AutoModelForCausalLM
|
| 53 |
|
| 54 |
-
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
|
| 55 |
-
|
| 56 |
-
|
| 57 |
generated_ids = model.generate(tokens, max_new_tokens=1000, do_sample=True)
|
| 58 |
|
| 59 |
# decode with mistral tokenizer
|
|
|
|
| 51 |
```py
|
| 52 |
from transformers import AutoModelForCausalLM
|
| 53 |
|
| 54 |
+
model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", device_map="auto")
|
| 55 |
+
tokens = torch.tensor(tokens).view(1, -1)
|
|
|
|
| 56 |
generated_ids = model.generate(tokens, max_new_tokens=1000, do_sample=True)
|
| 57 |
|
| 58 |
# decode with mistral tokenizer
|