Prepare support transformers
#24
by
rogeryoungh
- opened
Run the following Python script to run the model with Transformers.
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
import torch
MODEL_PATH = "MiniMaxAI/MiniMax-M2"
model = AutoModelForCausalLM.from_pretrained(
MODEL_PATH,
device_map="auto",
trust_remote_code=True,
)
tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH)
messages = [
{"role": "user", "content": [{"type": "text", "text": "What is your favourite condiment?"}]},
{"role": "assistant", "content": [{"type": "text", "text": "Well, I'm quite partial to a good squeeze of fresh lemon juice. It adds just the right amount of zesty flavour to whatever I'm cooking up in the kitchen!"}]},
{"role": "user", "content": [{"type": "text", "text": "Do you have mayonnaise recipes?"}]}
]
model_inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True).to("cuda")
generated_ids = model.generate(model_inputs, max_new_tokens=100, generation_config=generation_config)
response = tokenizer.batch_decode(generated_ids)[0]
print(response)
rogeryoungh
changed pull request status to
open
It should be generated_ids = model.generate(model_inputs, max_new_tokens=100, generation_config=model.generation_config)
It should be
generated_ids = model.generate(model_inputs, max_new_tokens=100, generation_config=model.generation_config)
Thanks, the docs has been fixed.