update inference with transformers

#162
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -51,9 +51,8 @@ print(result)
51
  ```py
52
  from transformers import AutoModelForCausalLM
53
 
54
- model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2")
55
- model.to("cuda")
56
-
57
  generated_ids = model.generate(tokens, max_new_tokens=1000, do_sample=True)
58
 
59
  # decode with mistral tokenizer
 
51
  ```py
52
  from transformers import AutoModelForCausalLM
53
 
54
+ model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-Instruct-v0.2", device_map="auto")
55
+ tokens = torch.tensor(tokens).view(1, -1)
 
56
  generated_ids = model.generate(tokens, max_new_tokens=1000, do_sample=True)
57
 
58
  # decode with mistral tokenizer