{ "_name_or_path": "pjox/dalembert", "architectures": [ "RobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "lachapelle", "1": "barthez_de_marmorieres", "2": "rallier", "3": "naigeon", "4": "diderot", "5": "willermoz", "6": "bellin", "7": "tressan", "8": "boullanger", "9": "duclos", "10": "cahusac", "11": "faiguet", "12": "romilly_je", "13": "falconet", "14": "venel", "15": "pesselier", "16": "david", "17": "holbach", "18": "leroy_jb", "19": "tarin", "20": "menuret", "21": "lemonnier", "22": "anville", "23": "robert", "24": "roux", "25": "bouillet_j", "26": "desmarest", "27": "paillasson", "28": "perronet", "29": "sauvages", "30": "petit", "31": "vandenesse", "32": "mallet", "33": "millot", "34": "papillon", "35": "margency", "36": "voltaire", "37": "lamotte", "38": "boucher", "39": "bourgelat", "40": "goussier", "41": "malouin", "42": "desmahis", "43": "barthez", "44": "beauzee", "45": "meyzieu", "46": "fenouillot", "47": "morellet", "48": "toussaint", "49": "daubenton_ljm", "50": "formey", "51": "grosley", "52": "durival_j", "53": "watelet", "54": "aumont", "55": "authville", "56": "pestre", "57": "ratte", "58": "dalembert", "59": "lebreton", "60": "gueneau", "61": "lesage", "62": "damilaville", "63": "penchenier", "64": "liebault", "65": "monnoye", "66": "soubeyran", "67": "brisson", "68": "louis", "69": "kurdwanowski", "70": "landois", "71": "berthoud", "72": "pezay", "73": "deleyre", "74": "voglie", "75": "lucotte_jr", "76": "dufour", "77": "dumarsais", "78": "eidous", "79": "necker", "80": "fouquet", "81": "bertrand", "82": "blondel", "83": "argenville", "84": "morand", "85": "forbonnais", "86": "rousseau", "87": "jaucourt", "88": "bouchaud", "89": "lenglet", "90": "lacondamine", "91": "montdorge", "92": "marmontel", "93": "leromain", "94": "montet", "95": "tronchin", "96": "bordeu", "97": "lavirotte", "98": "genson", "99": "leroy_cg", "100": "collot", "101": "lefevre", "102": "allut", "103": "brulle", "104": "villiers", "105": "grimm", "106": "bouchu", "107": "leblond", "108": "abbes", "109": "romilly_j", "110": "quesnay", "111": "yvon", "112": "daubenton_p" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "abbes": 108, "allut": 102, "anville": 22, "argenville": 83, "aumont": 54, "authville": 55, "barthez": 43, "barthez_de_marmorieres": 1, "beauzee": 44, "bellin": 6, "berthoud": 71, "bertrand": 81, "blondel": 82, "bordeu": 96, "bouchaud": 88, "boucher": 38, "bouchu": 106, "bouillet_j": 25, "boullanger": 8, "bourgelat": 39, "brisson": 67, "brulle": 103, "cahusac": 10, "collot": 100, "dalembert": 58, "damilaville": 62, "daubenton_ljm": 49, "daubenton_p": 112, "david": 16, "deleyre": 73, "desmahis": 42, "desmarest": 26, "diderot": 4, "duclos": 9, "dufour": 76, "dumarsais": 77, "durival_j": 52, "eidous": 78, "faiguet": 11, "falconet": 13, "fenouillot": 46, "forbonnais": 85, "formey": 50, "fouquet": 80, "genson": 98, "goussier": 40, "grimm": 105, "grosley": 51, "gueneau": 60, "holbach": 17, "jaucourt": 87, "kurdwanowski": 69, "lachapelle": 0, "lacondamine": 90, "lamotte": 37, "landois": 70, "lavirotte": 97, "leblond": 107, "lebreton": 59, "lefevre": 101, "lemonnier": 21, "lenglet": 89, "leromain": 93, "leroy_cg": 99, "leroy_jb": 18, "lesage": 61, "liebault": 64, "louis": 68, "lucotte_jr": 75, "mallet": 32, "malouin": 41, "margency": 35, "marmontel": 92, "menuret": 20, "meyzieu": 45, "millot": 33, "monnoye": 65, "montdorge": 91, "montet": 94, "morand": 84, "morellet": 47, "naigeon": 3, "necker": 79, "paillasson": 27, "papillon": 34, "penchenier": 63, "perronet": 28, "pesselier": 15, "pestre": 56, "petit": 30, "pezay": 72, "quesnay": 110, "rallier": 2, "ratte": 57, "robert": 23, "romilly_j": 109, "romilly_je": 12, "rousseau": 86, "roux": 24, "sauvages": 29, "soubeyran": 66, "tarin": 19, "toussaint": 48, "tressan": 7, "tronchin": 95, "vandenesse": 31, "venel": 14, "villiers": 104, "voglie": 74, "voltaire": 36, "watelet": 53, "willermoz": 5, "yvon": 111 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.42.4", "type_vocab_size": 1, "use_cache": true, "vocab_size": 32768 }