diff options
Diffstat (limited to 'models/trpg-final/checkpoint-200/tokenizer_config.json')
| -rw-r--r-- | models/trpg-final/checkpoint-200/tokenizer_config.json | 58 |
1 files changed, 58 insertions, 0 deletions
diff --git a/models/trpg-final/checkpoint-200/tokenizer_config.json b/models/trpg-final/checkpoint-200/tokenizer_config.json new file mode 100644 index 0000000..f0a3d97 --- /dev/null +++ b/models/trpg-final/checkpoint-200/tokenizer_config.json @@ -0,0 +1,58 @@ +{ + "added_tokens_decoder": { + "0": { + "content": "[PAD]", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "100": { + "content": "[UNK]", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "101": { + "content": "[CLS]", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "102": { + "content": "[SEP]", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "103": { + "content": "[MASK]", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "clean_up_tokenization_spaces": true, + "cls_token": "[CLS]", + "do_basic_tokenize": true, + "do_lower_case": true, + "extra_special_tokens": {}, + "mask_token": "[MASK]", + "model_max_length": 128, + "never_split": null, + "pad_token": "[PAD]", + "sep_token": "[SEP]", + "strip_accents": null, + "tokenize_chinese_chars": true, + "tokenizer_class": "BertTokenizer", + "unk_token": "[UNK]" +} |
