chatglm3-6b-93w

使用 THUDM/chatglm3-6b-base 基座模型在 93 万句子对数据集上训练得到的文本纠错模型。

Release Date

02/14/2024

训练数据

data/train.jsonl

训练方式

qlora

Demo

  • vllm==0.6.4

  • gradio==3.50.0

  • fast_match

  • fastlcs

安装 fast_match

pip install git+https://github.com/zejunwang1/fastMatch

安装 fastlcs

git clone https://github.com/zejunwang1/fastlcs
cd fastlcs/
git checkout deploy
pip install .

启动 demo:

cd chatglm3-6b-93w/inference/
export CUDA_VISIBLE_DEVICES=0
python gradio_demo.py

打开 http://[server_ip]:8080/,查看纠错效果:

Downloads last month
18
Safetensors
Model size
6B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for WangZeJun/chatglm3-6b-93w

Finetuned
(1)
this model

Collection including WangZeJun/chatglm3-6b-93w