Spaces:
Sleeping
Sleeping
| import json | |
| from tokenization import make_tokenizer | |
| add_sentinel_token = 0 | |
| tokenizer = make_tokenizer("ChineseSPTokenizer", None, "tokenizer.model", "50048", | |
| None, add_block_symbols=True, cache_dir="cache", | |
| add_sentinel_token=add_sentinel_token, add_task_mask=True, | |
| add_decoder_mask=False, | |
| fix_command_token=False) | |
| f_out = open("glm_chinese.vocab.txt", "w", encoding="utf-8") | |
| for idx in range(tokenizer.num_tokens): | |
| try: | |
| decode_str = tokenizer.DecodeIds([idx]) | |
| f_out.write("%d\t%s\n" % (idx, decode_str)) | |
| except Exception as e: | |
| print(idx, e) | |