--- license: mit --- # internlm2.5_7b_distill_orpo ## Architecture
## Base model [internlm2_5-7b-chat](https://huggingface.co/internlm/internlm2_5-7b-chat) ## Datasets used for training Created a preference optimization dataset [**PKU-SafeRLHF-orpo-72k**](https://huggingface.co/datasets/juneup/PKU-SafeRLHF-orpo-72k) from [PKU-SafeRLHF-single-dimension](https://huggingface.co/datasets/PKU-Alignment/PKU-SafeRLHF-single-dimension) ## Download model ```bash git lfs install git clone https://huggingface.co/juneup/internlm2.5_7b_distill_orpo ``` If you want to clone without large files - just their pointers ```bash GIT_LFS_SKIP_SMUDGE=1 git clone https://huggingface.co/juneup/internlm2.5_7b_distill_orpo ``` ### Download at Ollama ```bash ollama run Juneup/internlm2.5_7b_distill:orpo_q4_k_m