{ "_name_or_path": "Leo97/KcELECTRA-small-v2022-finetuned-in-vehicle", "architectures": [ "ElectraForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "embedding_size": 128, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 256, "id2label": { "0": "SNS\uae30\ub2a5", "1": "\uac80\uc0c9 \uae30\ub2a5", "2": "\uc74c\uc545 \uc11c\ube44\uc2a4", "3": "\uc8fc\ubb38 \uacb0\uc81c", "4": "\ub0a0\uc528 \uc815\ubcf4", "5": "\uc2a4\ud3ec\uce20 \uc815\ubcf4", "6": "\uc74c\uc2dd \uc815\ubcf4", "7": "\uc77c\uc815 \ud655\uc778", "8": "\uc8fc\uc2dd \uc815\ubcf4", "9": "\ub77c\ub514\uc624 \ubc29\uc1a1\uad6d", "10": "\ubb38\uc790 \uae30\ub2a5", "11": "\uc804\ud654 \uae30\ub2a5", "12": "\uc8fc\ubcc0 \ubaa9\uc801\uc9c0 \ud0d0\uc0c9", "13": "\ucc28\ub7c9 \uae30\ubcf8\uae30\ub2a5" }, "initializer_range": 0.02, "intermediate_size": 1024, "label2id": { "SNS\uae30\ub2a5": 0, "\uac80\uc0c9 \uae30\ub2a5": 1, "\ub0a0\uc528 \uc815\ubcf4": 4, "\ub77c\ub514\uc624 \ubc29\uc1a1\uad6d": 9, "\ubb38\uc790 \uae30\ub2a5": 10, "\uc2a4\ud3ec\uce20 \uc815\ubcf4": 5, "\uc74c\uc2dd \uc815\ubcf4": 6, "\uc74c\uc545 \uc11c\ube44\uc2a4": 2, "\uc77c\uc815 \ud655\uc778": 7, "\uc804\ud654 \uae30\ub2a5": 11, "\uc8fc\ubb38 \uacb0\uc81c": 3, "\uc8fc\ubcc0 \ubaa9\uc801\uc9c0 \ud0d0\uc0c9": 12, "\uc8fc\uc2dd \uc815\ubcf4": 8, "\ucc28\ub7c9 \uae30\ubcf8\uae30\ub2a5": 13 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "electra", "num_attention_heads": 4, "num_hidden_layers": 12, "pad_token_id": 0, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "summary_activation": "gelu", "summary_last_dropout": 0.1, "summary_type": "first", "summary_use_proj": true, "tokenizer_class": "BertTokenizer", "torch_dtype": "float32", "transformers_version": "4.26.1", "type_vocab_size": 2, "use_cache": true, "vocab_size": 54343 }