diff --git a/LoRA_adapters/dinov2-large-finetuned-lora-artist/adapter_config.json b/LoRA_adapters/dinov2-large-finetuned-lora-artist/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..86abe9913a713ad3d07a28e28ce4ac132cf4b3b7
--- /dev/null
+++ b/LoRA_adapters/dinov2-large-finetuned-lora-artist/adapter_config.json
@@ -0,0 +1,32 @@
+{
+  "alpha_pattern": {},
+  "auto_mapping": {
+    "base_model_class": "Dinov2ForImageClassification",
+    "parent_library": "transformers.models.dinov2.modeling_dinov2"
+  },
+  "base_model_name_or_path": "facebook/dinov2-large",
+  "bias": "none",
+  "fan_in_fan_out": false,
+  "inference_mode": true,
+  "init_lora_weights": true,
+  "layers_pattern": null,
+  "layers_to_transform": null,
+  "loftq_config": {},
+  "lora_alpha": 16,
+  "lora_dropout": 0.1,
+  "megatron_config": null,
+  "megatron_core": "megatron.core",
+  "modules_to_save": [
+    "classifier"
+  ],
+  "peft_type": "LORA",
+  "r": 16,
+  "rank_pattern": {},
+  "revision": null,
+  "target_modules": [
+    "query",
+    "value"
+  ],
+  "task_type": null,
+  "use_rslora": false
+}
\ No newline at end of file