{ "_name_or_path": "roberta-large-btc", "adapters": { "adapters": {}, "config_map": {}, "fusion_config_map": {}, "fusions": {} }, "architectures": [ "RobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "crf_state_dict": { "_constraint_mask": [ [ 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 0.0, 1.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 1.0 ], [ 1.0, 1.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0 ], [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0 ] ], "end_transitions": [ -1.6918301582336426, 0.83443683385849, -1.369484543800354, 0.9246144890785217, 1.069233775138855, -0.16902348399162292, -1.1555670499801636 ], "start_transitions": [ 0.508945107460022, 0.8599593043327332, -0.7935769557952881, 2.3224387168884277, 1.7772325277328491, -0.14362987875938416, 1.1923292875289917 ], "transitions": [ [ -0.16531139612197876, -0.44062453508377075, -0.17101362347602844, 0.25618258118629456, 0.8218539953231812, -0.6791337132453918, -0.18991276621818542 ], [ 0.049584291875362396, -0.029006613418459892, 0.3484196066856384, 0.10896176099777222, 0.252639502286911, -0.29233261942863464, -0.09638770669698715 ], [ -0.38768860697746277, 0.08189011365175247, -0.310952365398407, -0.09067512303590775, -0.34503042697906494, 0.36938661336898804, 0.2102288007736206 ], [ 0.25750938057899475, 0.8332298398017883, -0.7139731049537659, -0.2698363959789276, -0.2478555291891098, -0.15140868723392487, -0.09971318393945694 ], [ -0.08811905235052109, 0.08423032611608505, -0.09703858941793442, 0.8387375473976135, 0.055684782564640045, 0.14125464856624603, -0.33670008182525635 ], [ 0.2373524159193039, 0.07985689491033554, 0.02303706295788288, 0.3609643578529358, 0.3878211975097656, -0.557222306728363, -0.34269899129867554 ], [ -0.26117709279060364, 0.45397937297821045, 1.2661653757095337, 0.4104919135570526, 0.13623365759849548, -0.05334918946027756, -0.017508944496512413 ] ] }, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 1024, "id2label": { "0": "B-LOC", "1": "B-ORG", "2": "B-PER", "3": "I-LOC", "4": "I-ORG", "5": "I-PER", "6": "O" }, "initializer_range": 0.02, "intermediate_size": 4096, "label2id": { "B-LOC": 0, "B-ORG": 1, "B-PER": 2, "I-LOC": 3, "I-ORG": 4, "I-PER": 5, "O": 6 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 16, "num_hidden_layers": 24, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.11.3", "type_vocab_size": 1, "use_cache": true, "vocab_size": 50265 }