{ "_name_or_path": "pjox/dalembert", "architectures": [ "RobertaForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "authville", "1": "lesage", "2": "berthoud", "3": "quesnay", "4": "grimm", "5": "lucotte_jr", "6": "falconet", "7": "paillasson", "8": "argenville", "9": "pezay", "10": "lenglet", "11": "liebault", "12": "allut", "13": "tarin", "14": "barthez_de_marmorieres", "15": "montdorge", "16": "romilly_j", "17": "holbach", "18": "brisson", "19": "formey", "20": "eidous", "21": "leblond", "22": "leromain", "23": "penchenier", "24": "petit", "25": "meyzieu", "26": "lemonnier", "27": "vandenesse", "28": "jaucourt", "29": "lachapelle", "30": "menuret", "31": "rousseau", "32": "morellet", "33": "bouchaud", "34": "naigeon", "35": "boullanger", "36": "voltaire", "37": "beauzee", "38": "daubenton_ljm", "39": "toussaint", "40": "morand", "41": "kurdwanowski", "42": "desmarest", "43": "goussier", "44": "lebreton", "45": "leroy_cg", "46": "montet", "47": "robert", "48": "voglie", "49": "landois", "50": "damilaville", "51": "cahusac", "52": "dufour", "53": "necker", "54": "bordeu", "55": "aumont", "56": "desmahis", "57": "romilly_je", "58": "blondel", "59": "roux", "60": "collot", "61": "papillon", "62": "sauvages", "63": "rallier", "64": "genson", "65": "fenouillot", "66": "watelet", "67": "abbes", "68": "soubeyran", "69": "boucher", "70": "anville", "71": "brulle", "72": "bertrand", "73": "tressan", "74": "fouquet", "75": "deleyre", "76": "faiguet", "77": "grosley", "78": "ratte", "79": "dalembert", "80": "dumarsais", "81": "bellin", "82": "bouillet_j", "83": "lavirotte", "84": "millot", "85": "daubenton_p", "86": "yvon", "87": "margency", "88": "diderot", "89": "gueneau", "90": "forbonnais", "91": "barthez", "92": "bourgelat", "93": "bouchu", "94": "david", "95": "monnoye", "96": "venel", "97": "malouin", "98": "duclos", "99": "durival_j", "100": "villiers", "101": "perronet", "102": "lacondamine", "103": "leroy_jb", "104": "marmontel", "105": "mallet", "106": "pestre", "107": "lefevre", "108": "tronchin", "109": "pesselier", "110": "willermoz", "111": "louis", "112": "lamotte" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "abbes": 67, "allut": 12, "anville": 70, "argenville": 8, "aumont": 55, "authville": 0, "barthez": 91, "barthez_de_marmorieres": 14, "beauzee": 37, "bellin": 81, "berthoud": 2, "bertrand": 72, "blondel": 58, "bordeu": 54, "bouchaud": 33, "boucher": 69, "bouchu": 93, "bouillet_j": 82, "boullanger": 35, "bourgelat": 92, "brisson": 18, "brulle": 71, "cahusac": 51, "collot": 60, "dalembert": 79, "damilaville": 50, "daubenton_ljm": 38, "daubenton_p": 85, "david": 94, "deleyre": 75, "desmahis": 56, "desmarest": 42, "diderot": 88, "duclos": 98, "dufour": 52, "dumarsais": 80, "durival_j": 99, "eidous": 20, "faiguet": 76, "falconet": 6, "fenouillot": 65, "forbonnais": 90, "formey": 19, "fouquet": 74, "genson": 64, "goussier": 43, "grimm": 4, "grosley": 77, "gueneau": 89, "holbach": 17, "jaucourt": 28, "kurdwanowski": 41, "lachapelle": 29, "lacondamine": 102, "lamotte": 112, "landois": 49, "lavirotte": 83, "leblond": 21, "lebreton": 44, "lefevre": 107, "lemonnier": 26, "lenglet": 10, "leromain": 22, "leroy_cg": 45, "leroy_jb": 103, "lesage": 1, "liebault": 11, "louis": 111, "lucotte_jr": 5, "mallet": 105, "malouin": 97, "margency": 87, "marmontel": 104, "menuret": 30, "meyzieu": 25, "millot": 84, "monnoye": 95, "montdorge": 15, "montet": 46, "morand": 40, "morellet": 32, "naigeon": 34, "necker": 53, "paillasson": 7, "papillon": 61, "penchenier": 23, "perronet": 101, "pesselier": 109, "pestre": 106, "petit": 24, "pezay": 9, "quesnay": 3, "rallier": 63, "ratte": 78, "robert": 47, "romilly_j": 16, "romilly_je": 57, "rousseau": 31, "roux": 59, "sauvages": 62, "soubeyran": 68, "tarin": 13, "toussaint": 39, "tressan": 73, "tronchin": 108, "vandenesse": 27, "venel": 96, "villiers": 100, "voglie": 48, "voltaire": 36, "watelet": 66, "willermoz": 110, "yvon": 86 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "pad_token_id": 1, "position_embedding_type": "absolute", "problem_type": "single_label_classification", "torch_dtype": "float32", "transformers_version": "4.44.2", "type_vocab_size": 1, "use_cache": true, "vocab_size": 32768 }