| library_name: transformers | |
| license: apache-2.0 | |
| base_model: PharMolix/BioMedGPT-LM-7B | |
| language: | |
| - en | |
| 16-bit version of weights from `PharMolix/BioMedGPT-LM-7B`, for easier download / finetuning / model-merging | |
| Code | |
| ```python | |
| import torch | |
| from transformers import AutoModelForCausalLM, AutoTokenizer | |
| m2 = AutoModelForCausalLM.from_pretrained("PharMolix/BioMedGPT-LM-7B", | |
| torch_dtype=torch.float16, | |
| device_map="auto") | |
| ``` |