{ "_name_or_path": "neuralmind/bert-base-portuguese-cased", "architectures": [ "BertForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "classifier_dropout": null, "directionality": "bidi", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "cenarios-2017", "1": "topofmind", "2": "multimidia", "3": "mercado", "4": "tec", "5": "ambiente", "6": "opiniao", "7": "o-melhor-de-sao-paulo", "8": "especial", "9": "ilustrada", "10": "infograficos", "11": "equilibrioesaude", "12": "sobretudo", "13": "serafina", "14": "esporte", "15": "guia-de-livros-discos-filmes", "16": "seminariosfolha", "17": "turismo", "18": "rfi", "19": "bichos", "20": "saopaulo", "21": "asmais", "22": "folhinha", "23": "mundo", "24": "ombudsman", "25": "vice", "26": "treinamentocienciaesaude", "27": "tv", "28": "ciencia", "29": "bbc", "30": "ilustrissima", "31": "comida", "32": "dw", "33": "treinamento", "34": "empreendedorsocial", "35": "mulher", "36": "educacao", "37": "guia-de-livros-filmes-discos", "38": "poder", "39": "paineldoleitor", "40": "euronews", "41": "colunas", "42": "cotidiano", "43": "banco-de-dados" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "ambiente": 5, "asmais": 21, "banco-de-dados": 43, "bbc": 29, "bichos": 19, "cenarios-2017": 0, "ciencia": 28, "colunas": 41, "comida": 31, "cotidiano": 42, "dw": 32, "educacao": 36, "empreendedorsocial": 34, "equilibrioesaude": 11, "especial": 8, "esporte": 14, "euronews": 40, "folhinha": 22, "guia-de-livros-discos-filmes": 15, "guia-de-livros-filmes-discos": 37, "ilustrada": 9, "ilustrissima": 30, "infograficos": 10, "mercado": 3, "mulher": 35, "multimidia": 2, "mundo": 23, "o-melhor-de-sao-paulo": 7, "ombudsman": 24, "opiniao": 6, "paineldoleitor": 39, "poder": 38, "rfi": 18, "saopaulo": 20, "seminariosfolha": 16, "serafina": 13, "sobretudo": 12, "tec": 4, "topofmind": 1, "treinamento": 33, "treinamentocienciaesaude": 26, "turismo": 17, "tv": 27, "vice": 25 }, "layer_norm_eps": 1e-12, "max_position_embeddings": 512, "model_type": "bert", "num_attention_heads": 12, "num_hidden_layers": 12, "output_past": true, "pad_token_id": 0, "pooler_fc_size": 768, "pooler_num_attention_heads": 12, "pooler_num_fc_layers": 3, "pooler_size_per_head": 128, "pooler_type": "first_token_transform", "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.41.2", "type_vocab_size": 2, "use_cache": true, "vocab_size": 29794 }