| { |
| "_name_or_path": "roberta-large-btc", |
| "adapters": { |
| "adapters": {}, |
| "config_map": {}, |
| "fusion_config_map": {}, |
| "fusions": {} |
| }, |
| "architectures": [ |
| "RobertaForTokenClassification" |
| ], |
| "attention_probs_dropout_prob": 0.1, |
| "bos_token_id": 0, |
| "classifier_dropout": null, |
| "crf_state_dict": { |
| "_constraint_mask": [ |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 1.0 |
| ], |
| [ |
| 1.0, |
| 1.0, |
| 1.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 1.0, |
| 0.0, |
| 0.0 |
| ], |
| [ |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0, |
| 0.0 |
| ] |
| ], |
| "end_transitions": [ |
| -1.6918301582336426, |
| 0.83443683385849, |
| -1.369484543800354, |
| 0.9246144890785217, |
| 1.069233775138855, |
| -0.16902348399162292, |
| -1.1555670499801636 |
| ], |
| "start_transitions": [ |
| 0.508945107460022, |
| 0.8599593043327332, |
| -0.7935769557952881, |
| 2.3224387168884277, |
| 1.7772325277328491, |
| -0.14362987875938416, |
| 1.1923292875289917 |
| ], |
| "transitions": [ |
| [ |
| -0.16531139612197876, |
| -0.44062453508377075, |
| -0.17101362347602844, |
| 0.25618258118629456, |
| 0.8218539953231812, |
| -0.6791337132453918, |
| -0.18991276621818542 |
| ], |
| [ |
| 0.049584291875362396, |
| -0.029006613418459892, |
| 0.3484196066856384, |
| 0.10896176099777222, |
| 0.252639502286911, |
| -0.29233261942863464, |
| -0.09638770669698715 |
| ], |
| [ |
| -0.38768860697746277, |
| 0.08189011365175247, |
| -0.310952365398407, |
| -0.09067512303590775, |
| -0.34503042697906494, |
| 0.36938661336898804, |
| 0.2102288007736206 |
| ], |
| [ |
| 0.25750938057899475, |
| 0.8332298398017883, |
| -0.7139731049537659, |
| -0.2698363959789276, |
| -0.2478555291891098, |
| -0.15140868723392487, |
| -0.09971318393945694 |
| ], |
| [ |
| -0.08811905235052109, |
| 0.08423032611608505, |
| -0.09703858941793442, |
| 0.8387375473976135, |
| 0.055684782564640045, |
| 0.14125464856624603, |
| -0.33670008182525635 |
| ], |
| [ |
| 0.2373524159193039, |
| 0.07985689491033554, |
| 0.02303706295788288, |
| 0.3609643578529358, |
| 0.3878211975097656, |
| -0.557222306728363, |
| -0.34269899129867554 |
| ], |
| [ |
| -0.26117709279060364, |
| 0.45397937297821045, |
| 1.2661653757095337, |
| 0.4104919135570526, |
| 0.13623365759849548, |
| -0.05334918946027756, |
| -0.017508944496512413 |
| ] |
| ] |
| }, |
| "eos_token_id": 2, |
| "hidden_act": "gelu", |
| "hidden_dropout_prob": 0.1, |
| "hidden_size": 1024, |
| "id2label": { |
| "0": "B-LOC", |
| "1": "B-ORG", |
| "2": "B-PER", |
| "3": "I-LOC", |
| "4": "I-ORG", |
| "5": "I-PER", |
| "6": "O" |
| }, |
| "initializer_range": 0.02, |
| "intermediate_size": 4096, |
| "label2id": { |
| "B-LOC": 0, |
| "B-ORG": 1, |
| "B-PER": 2, |
| "I-LOC": 3, |
| "I-ORG": 4, |
| "I-PER": 5, |
| "O": 6 |
| }, |
| "layer_norm_eps": 1e-05, |
| "max_position_embeddings": 514, |
| "model_type": "roberta", |
| "num_attention_heads": 16, |
| "num_hidden_layers": 24, |
| "pad_token_id": 1, |
| "position_embedding_type": "absolute", |
| "torch_dtype": "float32", |
| "transformers_version": "4.11.3", |
| "type_vocab_size": 1, |
| "use_cache": true, |
| "vocab_size": 50265 |
| } |
|
|