|
{ |
|
"_name_or_path": "roberta-large-btc", |
|
"adapters": { |
|
"adapters": {}, |
|
"config_map": {}, |
|
"fusion_config_map": {}, |
|
"fusions": {} |
|
}, |
|
"architectures": [ |
|
"RobertaForTokenClassification" |
|
], |
|
"attention_probs_dropout_prob": 0.1, |
|
"bos_token_id": 0, |
|
"classifier_dropout": null, |
|
"crf_state_dict": { |
|
"_constraint_mask": [ |
|
[ |
|
1.0, |
|
1.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
0.0, |
|
1.0, |
|
0.0, |
|
1.0 |
|
], |
|
[ |
|
1.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
1.0, |
|
0.0, |
|
1.0, |
|
0.0, |
|
1.0 |
|
], |
|
[ |
|
1.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
0.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
1.0 |
|
], |
|
[ |
|
1.0, |
|
1.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
0.0, |
|
1.0, |
|
0.0, |
|
1.0 |
|
], |
|
[ |
|
1.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
1.0, |
|
0.0, |
|
1.0, |
|
0.0, |
|
1.0 |
|
], |
|
[ |
|
1.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
0.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
1.0 |
|
], |
|
[ |
|
1.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
0.0, |
|
0.0, |
|
1.0, |
|
0.0, |
|
1.0 |
|
], |
|
[ |
|
1.0, |
|
1.0, |
|
1.0, |
|
0.0, |
|
0.0, |
|
0.0, |
|
1.0, |
|
0.0, |
|
0.0 |
|
], |
|
[ |
|
0.0, |
|
0.0, |
|
0.0, |
|
0.0, |
|
0.0, |
|
0.0, |
|
0.0, |
|
0.0, |
|
0.0 |
|
] |
|
], |
|
"end_transitions": [ |
|
-1.6918301582336426, |
|
0.83443683385849, |
|
-1.369484543800354, |
|
0.9246144890785217, |
|
1.069233775138855, |
|
-0.16902348399162292, |
|
-1.1555670499801636 |
|
], |
|
"start_transitions": [ |
|
0.508945107460022, |
|
0.8599593043327332, |
|
-0.7935769557952881, |
|
2.3224387168884277, |
|
1.7772325277328491, |
|
-0.14362987875938416, |
|
1.1923292875289917 |
|
], |
|
"transitions": [ |
|
[ |
|
-0.16531139612197876, |
|
-0.44062453508377075, |
|
-0.17101362347602844, |
|
0.25618258118629456, |
|
0.8218539953231812, |
|
-0.6791337132453918, |
|
-0.18991276621818542 |
|
], |
|
[ |
|
0.049584291875362396, |
|
-0.029006613418459892, |
|
0.3484196066856384, |
|
0.10896176099777222, |
|
0.252639502286911, |
|
-0.29233261942863464, |
|
-0.09638770669698715 |
|
], |
|
[ |
|
-0.38768860697746277, |
|
0.08189011365175247, |
|
-0.310952365398407, |
|
-0.09067512303590775, |
|
-0.34503042697906494, |
|
0.36938661336898804, |
|
0.2102288007736206 |
|
], |
|
[ |
|
0.25750938057899475, |
|
0.8332298398017883, |
|
-0.7139731049537659, |
|
-0.2698363959789276, |
|
-0.2478555291891098, |
|
-0.15140868723392487, |
|
-0.09971318393945694 |
|
], |
|
[ |
|
-0.08811905235052109, |
|
0.08423032611608505, |
|
-0.09703858941793442, |
|
0.8387375473976135, |
|
0.055684782564640045, |
|
0.14125464856624603, |
|
-0.33670008182525635 |
|
], |
|
[ |
|
0.2373524159193039, |
|
0.07985689491033554, |
|
0.02303706295788288, |
|
0.3609643578529358, |
|
0.3878211975097656, |
|
-0.557222306728363, |
|
-0.34269899129867554 |
|
], |
|
[ |
|
-0.26117709279060364, |
|
0.45397937297821045, |
|
1.2661653757095337, |
|
0.4104919135570526, |
|
0.13623365759849548, |
|
-0.05334918946027756, |
|
-0.017508944496512413 |
|
] |
|
] |
|
}, |
|
"eos_token_id": 2, |
|
"hidden_act": "gelu", |
|
"hidden_dropout_prob": 0.1, |
|
"hidden_size": 1024, |
|
"id2label": { |
|
"0": "B-LOC", |
|
"1": "B-ORG", |
|
"2": "B-PER", |
|
"3": "I-LOC", |
|
"4": "I-ORG", |
|
"5": "I-PER", |
|
"6": "O" |
|
}, |
|
"initializer_range": 0.02, |
|
"intermediate_size": 4096, |
|
"label2id": { |
|
"B-LOC": 0, |
|
"B-ORG": 1, |
|
"B-PER": 2, |
|
"I-LOC": 3, |
|
"I-ORG": 4, |
|
"I-PER": 5, |
|
"O": 6 |
|
}, |
|
"layer_norm_eps": 1e-05, |
|
"max_position_embeddings": 514, |
|
"model_type": "roberta", |
|
"num_attention_heads": 16, |
|
"num_hidden_layers": 24, |
|
"pad_token_id": 1, |
|
"position_embedding_type": "absolute", |
|
"torch_dtype": "float32", |
|
"transformers_version": "4.11.3", |
|
"type_vocab_size": 1, |
|
"use_cache": true, |
|
"vocab_size": 50265 |
|
} |
|
|