{ "_name_or_path": "microsoft/layoutlmv3-base", "architectures": [ "LayoutLMv3ForSequenceClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "coordinate_size": 128, "eos_token_id": 2, "has_relative_attention_bias": true, "has_spatial_attention_bias": true, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "additional_documentation_for_claims", "1": "additional_documents_for_authorization", "2": "authorization", "3": "clinical_notes", "4": "continuity_of_care", "5": "discharge_bundle", "6": "dme_orders", "7": "f2f_notes", "8": "facesheets", "9": "facility_after_visit_summary", "10": "home_health_orders", "11": "icd_code_request", "12": "insurance", "13": "lab_orders", "14": "lab_reports", "15": "medication_change_orders", "16": "medications", "17": "medline", "18": "medline_documents", "19": "orders", "20": "ostomy_orders", "21": "poc", "22": "progress_note", "23": "referral_bundle", "24": "verbal_orders_signed", "25": "visit_note", "26": "wound_orders" }, "initializer_range": 0.02, "input_size": 224, "intermediate_size": 3072, "label2id": { "additional_documentation_for_claims": 0, "additional_documents_for_authorization": 1, "authorization": 2, "clinical_notes": 3, "continuity_of_care": 4, "discharge_bundle": 5, "dme_orders": 6, "f2f_notes": 7, "facesheets": 8, "facility_after_visit_summary": 9, "home_health_orders": 10, "icd_code_request": 11, "insurance": 12, "lab_orders": 13, "lab_reports": 14, "medication_change_orders": 15, "medications": 16, "medline": 17, "medline_documents": 18, "orders": 19, "ostomy_orders": 20, "poc": 21, "progress_note": 22, "referral_bundle": 23, "verbal_orders_signed": 24, "visit_note": 25, "wound_orders": 26 }, "layer_norm_eps": 1e-05, "max_2d_position_embeddings": 1024, "max_position_embeddings": 514, "max_rel_2d_pos": 256, "max_rel_pos": 128, "model_type": "layoutlmv3", "num_attention_heads": 12, "num_channels": 3, "num_hidden_layers": 12, "pad_token_id": 1, "patch_size": 16, "rel_2d_pos_bins": 64, "rel_pos_bins": 32, "second_input_size": 112, "shape_size": 128, "text_embed": true, "torch_dtype": "float32", "transformers_version": "4.47.1", "type_vocab_size": 1, "visual_embed": true, "vocab_size": 50265 }