Upload folder using huggingface_hub
Browse files- zh/.msc +0 -0
- zh/.mv +1 -0
- zh/README.md +12 -0
- zh/config.json +48 -0
- zh/configuration.json +1 -0
- zh/model.safetensors +3 -0
- zh/rng_state.pth +3 -0
- zh/scheduler.pt +3 -0
- zh/special_tokens_map.json +37 -0
- zh/tokenizer.json +0 -0
- zh/tokenizer_config.json +56 -0
- zh/trainer_state.json +170 -0
- zh/training_args.bin +3 -0
- zh/vocab.txt +0 -0
zh/.msc
ADDED
|
Binary file (860 Bytes). View file
|
|
|
zh/.mv
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
Revision:master,CreatedAt:1767769058
|
zh/README.md
ADDED
|
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
library_name: transformers
|
| 4 |
+
tags:
|
| 5 |
+
- intent-classification
|
| 6 |
+
- bert
|
| 7 |
+
- chinese
|
| 8 |
+
---
|
| 9 |
+
|
| 10 |
+
# SaleIntent (Chinese)
|
| 11 |
+
|
| 12 |
+
Chinese intent classification model for sales and customer service scenarios.
|
zh/config.json
ADDED
|
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"architectures": [
|
| 3 |
+
"BertForSequenceClassification"
|
| 4 |
+
],
|
| 5 |
+
"attention_probs_dropout_prob": 0.1,
|
| 6 |
+
"bos_token_id": 0,
|
| 7 |
+
"classifier_dropout": null,
|
| 8 |
+
"directionality": "bidi",
|
| 9 |
+
"eos_token_id": 2,
|
| 10 |
+
"hidden_act": "gelu",
|
| 11 |
+
"hidden_dropout_prob": 0.1,
|
| 12 |
+
"hidden_size": 768,
|
| 13 |
+
"id2label": {
|
| 14 |
+
"0": "A",
|
| 15 |
+
"1": "B",
|
| 16 |
+
"2": "C",
|
| 17 |
+
"3": "F",
|
| 18 |
+
"4": "X"
|
| 19 |
+
},
|
| 20 |
+
"initializer_range": 0.02,
|
| 21 |
+
"intermediate_size": 3072,
|
| 22 |
+
"label2id": {
|
| 23 |
+
"A": 0,
|
| 24 |
+
"B": 1,
|
| 25 |
+
"C": 2,
|
| 26 |
+
"F": 3,
|
| 27 |
+
"X": 4
|
| 28 |
+
},
|
| 29 |
+
"layer_norm_eps": 1e-12,
|
| 30 |
+
"max_position_embeddings": 512,
|
| 31 |
+
"model_type": "bert",
|
| 32 |
+
"num_attention_heads": 12,
|
| 33 |
+
"num_hidden_layers": 12,
|
| 34 |
+
"output_past": true,
|
| 35 |
+
"pad_token_id": 0,
|
| 36 |
+
"pooler_fc_size": 768,
|
| 37 |
+
"pooler_num_attention_heads": 12,
|
| 38 |
+
"pooler_num_fc_layers": 3,
|
| 39 |
+
"pooler_size_per_head": 128,
|
| 40 |
+
"pooler_type": "first_token_transform",
|
| 41 |
+
"position_embedding_type": "absolute",
|
| 42 |
+
"problem_type": "single_label_classification",
|
| 43 |
+
"torch_dtype": "float32",
|
| 44 |
+
"transformers_version": "4.51.3",
|
| 45 |
+
"type_vocab_size": 2,
|
| 46 |
+
"use_cache": true,
|
| 47 |
+
"vocab_size": 21128
|
| 48 |
+
}
|
zh/configuration.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"framework": "pytorch", "task": "other"}
|
zh/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d612c090f7b8234c8005f09f04f60952bafd32052dfa2b6137525a41ff7dcc7b
|
| 3 |
+
size 409109468
|
zh/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a5ba5455486eb83ef8e0717df01dfd5ffb5b605910cb0ea0b7d8f206157bb75e
|
| 3 |
+
size 14244
|
zh/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:82f086362fd598cad0852e5445dc8a59f8987edd7993f40edd6de3b6614e6c2f
|
| 3 |
+
size 1064
|
zh/special_tokens_map.json
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": {
|
| 3 |
+
"content": "[CLS]",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"mask_token": {
|
| 10 |
+
"content": "[MASK]",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": {
|
| 17 |
+
"content": "[PAD]",
|
| 18 |
+
"lstrip": false,
|
| 19 |
+
"normalized": false,
|
| 20 |
+
"rstrip": false,
|
| 21 |
+
"single_word": false
|
| 22 |
+
},
|
| 23 |
+
"sep_token": {
|
| 24 |
+
"content": "[SEP]",
|
| 25 |
+
"lstrip": false,
|
| 26 |
+
"normalized": false,
|
| 27 |
+
"rstrip": false,
|
| 28 |
+
"single_word": false
|
| 29 |
+
},
|
| 30 |
+
"unk_token": {
|
| 31 |
+
"content": "[UNK]",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false
|
| 36 |
+
}
|
| 37 |
+
}
|
zh/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
zh/tokenizer_config.json
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": false,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": true,
|
| 47 |
+
"extra_special_tokens": {},
|
| 48 |
+
"mask_token": "[MASK]",
|
| 49 |
+
"model_max_length": 1000000000000000019884624838656,
|
| 50 |
+
"pad_token": "[PAD]",
|
| 51 |
+
"sep_token": "[SEP]",
|
| 52 |
+
"strip_accents": null,
|
| 53 |
+
"tokenize_chinese_chars": true,
|
| 54 |
+
"tokenizer_class": "BertTokenizer",
|
| 55 |
+
"unk_token": "[UNK]"
|
| 56 |
+
}
|
zh/trainer_state.json
ADDED
|
@@ -0,0 +1,170 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_global_step": null,
|
| 3 |
+
"best_metric": null,
|
| 4 |
+
"best_model_checkpoint": null,
|
| 5 |
+
"epoch": 6.0,
|
| 6 |
+
"eval_steps": 500,
|
| 7 |
+
"global_step": 5178,
|
| 8 |
+
"is_hyper_param_search": false,
|
| 9 |
+
"is_local_process_zero": true,
|
| 10 |
+
"is_world_process_zero": true,
|
| 11 |
+
"log_history": [
|
| 12 |
+
{
|
| 13 |
+
"epoch": 0.5793742757821553,
|
| 14 |
+
"grad_norm": 4.606655597686768,
|
| 15 |
+
"learning_rate": 1.8072614909231365e-05,
|
| 16 |
+
"loss": 0.3853,
|
| 17 |
+
"step": 500
|
| 18 |
+
},
|
| 19 |
+
{
|
| 20 |
+
"epoch": 1.0,
|
| 21 |
+
"eval_accuracy": 0.9035830618892509,
|
| 22 |
+
"eval_f1": 0.7438579904034558,
|
| 23 |
+
"eval_loss": 0.26155194640159607,
|
| 24 |
+
"eval_model_preparation_time": 0.0033,
|
| 25 |
+
"eval_runtime": 4.4748,
|
| 26 |
+
"eval_samples_per_second": 343.035,
|
| 27 |
+
"eval_steps_per_second": 21.454,
|
| 28 |
+
"step": 863
|
| 29 |
+
},
|
| 30 |
+
{
|
| 31 |
+
"epoch": 1.1587485515643106,
|
| 32 |
+
"grad_norm": 4.532162666320801,
|
| 33 |
+
"learning_rate": 1.6141367323290847e-05,
|
| 34 |
+
"loss": 0.2576,
|
| 35 |
+
"step": 1000
|
| 36 |
+
},
|
| 37 |
+
{
|
| 38 |
+
"epoch": 1.7381228273464657,
|
| 39 |
+
"grad_norm": 5.762636184692383,
|
| 40 |
+
"learning_rate": 1.4210119737350329e-05,
|
| 41 |
+
"loss": 0.2283,
|
| 42 |
+
"step": 1500
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"epoch": 2.0,
|
| 46 |
+
"eval_accuracy": 0.9198697068403909,
|
| 47 |
+
"eval_f1": 0.7780841955159529,
|
| 48 |
+
"eval_loss": 0.2665092945098877,
|
| 49 |
+
"eval_model_preparation_time": 0.0033,
|
| 50 |
+
"eval_runtime": 4.4634,
|
| 51 |
+
"eval_samples_per_second": 343.91,
|
| 52 |
+
"eval_steps_per_second": 21.508,
|
| 53 |
+
"step": 1726
|
| 54 |
+
},
|
| 55 |
+
{
|
| 56 |
+
"epoch": 2.317497103128621,
|
| 57 |
+
"grad_norm": 0.497444748878479,
|
| 58 |
+
"learning_rate": 1.2278872151409811e-05,
|
| 59 |
+
"loss": 0.1931,
|
| 60 |
+
"step": 2000
|
| 61 |
+
},
|
| 62 |
+
{
|
| 63 |
+
"epoch": 2.8968713789107765,
|
| 64 |
+
"grad_norm": 3.7842884063720703,
|
| 65 |
+
"learning_rate": 1.0347624565469293e-05,
|
| 66 |
+
"loss": 0.1834,
|
| 67 |
+
"step": 2500
|
| 68 |
+
},
|
| 69 |
+
{
|
| 70 |
+
"epoch": 3.0,
|
| 71 |
+
"eval_accuracy": 0.9146579804560261,
|
| 72 |
+
"eval_f1": 0.765570564927072,
|
| 73 |
+
"eval_loss": 0.2603704333305359,
|
| 74 |
+
"eval_model_preparation_time": 0.0033,
|
| 75 |
+
"eval_runtime": 4.4661,
|
| 76 |
+
"eval_samples_per_second": 343.702,
|
| 77 |
+
"eval_steps_per_second": 21.495,
|
| 78 |
+
"step": 2589
|
| 79 |
+
},
|
| 80 |
+
{
|
| 81 |
+
"epoch": 3.4762456546929315,
|
| 82 |
+
"grad_norm": 0.3960210680961609,
|
| 83 |
+
"learning_rate": 8.416376979528776e-06,
|
| 84 |
+
"loss": 0.1353,
|
| 85 |
+
"step": 3000
|
| 86 |
+
},
|
| 87 |
+
{
|
| 88 |
+
"epoch": 4.0,
|
| 89 |
+
"eval_accuracy": 0.9218241042345277,
|
| 90 |
+
"eval_f1": 0.7780743111484496,
|
| 91 |
+
"eval_loss": 0.2716759443283081,
|
| 92 |
+
"eval_model_preparation_time": 0.0033,
|
| 93 |
+
"eval_runtime": 4.48,
|
| 94 |
+
"eval_samples_per_second": 342.633,
|
| 95 |
+
"eval_steps_per_second": 21.429,
|
| 96 |
+
"step": 3452
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"epoch": 4.055619930475087,
|
| 100 |
+
"grad_norm": 5.781081199645996,
|
| 101 |
+
"learning_rate": 6.485129393588259e-06,
|
| 102 |
+
"loss": 0.1446,
|
| 103 |
+
"step": 3500
|
| 104 |
+
},
|
| 105 |
+
{
|
| 106 |
+
"epoch": 4.634994206257242,
|
| 107 |
+
"grad_norm": 9.73343563079834,
|
| 108 |
+
"learning_rate": 4.553881807647741e-06,
|
| 109 |
+
"loss": 0.1029,
|
| 110 |
+
"step": 4000
|
| 111 |
+
},
|
| 112 |
+
{
|
| 113 |
+
"epoch": 5.0,
|
| 114 |
+
"eval_accuracy": 0.9257328990228013,
|
| 115 |
+
"eval_f1": 0.7846248698845295,
|
| 116 |
+
"eval_loss": 0.3429003655910492,
|
| 117 |
+
"eval_model_preparation_time": 0.0033,
|
| 118 |
+
"eval_runtime": 4.4592,
|
| 119 |
+
"eval_samples_per_second": 344.234,
|
| 120 |
+
"eval_steps_per_second": 21.529,
|
| 121 |
+
"step": 4315
|
| 122 |
+
},
|
| 123 |
+
{
|
| 124 |
+
"epoch": 5.214368482039397,
|
| 125 |
+
"grad_norm": 0.018824853003025055,
|
| 126 |
+
"learning_rate": 2.622634221707223e-06,
|
| 127 |
+
"loss": 0.0826,
|
| 128 |
+
"step": 4500
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"epoch": 5.793742757821553,
|
| 132 |
+
"grad_norm": 0.07604879140853882,
|
| 133 |
+
"learning_rate": 6.913866357667054e-07,
|
| 134 |
+
"loss": 0.0782,
|
| 135 |
+
"step": 5000
|
| 136 |
+
},
|
| 137 |
+
{
|
| 138 |
+
"epoch": 6.0,
|
| 139 |
+
"eval_accuracy": 0.9179153094462541,
|
| 140 |
+
"eval_f1": 0.7776778508930786,
|
| 141 |
+
"eval_loss": 0.3899814486503601,
|
| 142 |
+
"eval_model_preparation_time": 0.0033,
|
| 143 |
+
"eval_runtime": 4.4603,
|
| 144 |
+
"eval_samples_per_second": 344.145,
|
| 145 |
+
"eval_steps_per_second": 21.523,
|
| 146 |
+
"step": 5178
|
| 147 |
+
}
|
| 148 |
+
],
|
| 149 |
+
"logging_steps": 500,
|
| 150 |
+
"max_steps": 5178,
|
| 151 |
+
"num_input_tokens_seen": 0,
|
| 152 |
+
"num_train_epochs": 6,
|
| 153 |
+
"save_steps": 500,
|
| 154 |
+
"stateful_callbacks": {
|
| 155 |
+
"TrainerControl": {
|
| 156 |
+
"args": {
|
| 157 |
+
"should_epoch_stop": false,
|
| 158 |
+
"should_evaluate": false,
|
| 159 |
+
"should_log": false,
|
| 160 |
+
"should_save": true,
|
| 161 |
+
"should_training_stop": true
|
| 162 |
+
},
|
| 163 |
+
"attributes": {}
|
| 164 |
+
}
|
| 165 |
+
},
|
| 166 |
+
"total_flos": 1.0898616579001344e+16,
|
| 167 |
+
"train_batch_size": 16,
|
| 168 |
+
"trial_name": null,
|
| 169 |
+
"trial_params": null
|
| 170 |
+
}
|
zh/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:453eb7e84a3532eaf755e0c5e8036dfd3e951d67f790e960278a79f9a2b7f5d6
|
| 3 |
+
size 5240
|
zh/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|