diff options
Diffstat (limited to 'models/trpg-final/tokenizer_config.json')
| -rw-r--r-- | models/trpg-final/tokenizer_config.json | 58 |
1 files changed, 0 insertions, 58 deletions
diff --git a/models/trpg-final/tokenizer_config.json b/models/trpg-final/tokenizer_config.json deleted file mode 100644 index f0a3d97..0000000 --- a/models/trpg-final/tokenizer_config.json +++ /dev/null @@ -1,58 +0,0 @@ -{ - "added_tokens_decoder": { - "0": { - "content": "[PAD]", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - }, - "100": { - "content": "[UNK]", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - }, - "101": { - "content": "[CLS]", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - }, - "102": { - "content": "[SEP]", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - }, - "103": { - "content": "[MASK]", - "lstrip": false, - "normalized": false, - "rstrip": false, - "single_word": false, - "special": true - } - }, - "clean_up_tokenization_spaces": true, - "cls_token": "[CLS]", - "do_basic_tokenize": true, - "do_lower_case": true, - "extra_special_tokens": {}, - "mask_token": "[MASK]", - "model_max_length": 128, - "never_split": null, - "pad_token": "[PAD]", - "sep_token": "[SEP]", - "strip_accents": null, - "tokenize_chinese_chars": true, - "tokenizer_class": "BertTokenizer", - "unk_token": "[UNK]" -} |
