From 2d08e09f30086d067dde56c321bb6db02c5e3de8 Mon Sep 17 00:00:00 2001
From: Tetiana Yemelianenko <tyemel.mzeom@gmail.com>
Date: Tue, 12 Nov 2024 14:29:05 +0000
Subject: [PATCH] Upload New File

---
 .../adapter_config.json                       | 32 +++++++++++++++++++
 1 file changed, 32 insertions(+)
 create mode 100644 LoRA_adapters/dinov2-large-finetuned-lora-genre/adapter_config.json

diff --git a/LoRA_adapters/dinov2-large-finetuned-lora-genre/adapter_config.json b/LoRA_adapters/dinov2-large-finetuned-lora-genre/adapter_config.json
new file mode 100644
index 0000000..86abe99
--- /dev/null
+++ b/LoRA_adapters/dinov2-large-finetuned-lora-genre/adapter_config.json
@@ -0,0 +1,32 @@
+{
+  "alpha_pattern": {},
+  "auto_mapping": {
+    "base_model_class": "Dinov2ForImageClassification",
+    "parent_library": "transformers.models.dinov2.modeling_dinov2"
+  },
+  "base_model_name_or_path": "facebook/dinov2-large",
+  "bias": "none",
+  "fan_in_fan_out": false,
+  "inference_mode": true,
+  "init_lora_weights": true,
+  "layers_pattern": null,
+  "layers_to_transform": null,
+  "loftq_config": {},
+  "lora_alpha": 16,
+  "lora_dropout": 0.1,
+  "megatron_config": null,
+  "megatron_core": "megatron.core",
+  "modules_to_save": [
+    "classifier"
+  ],
+  "peft_type": "LORA",
+  "r": 16,
+  "rank_pattern": {},
+  "revision": null,
+  "target_modules": [
+    "query",
+    "value"
+  ],
+  "task_type": null,
+  "use_rslora": false
+}
\ No newline at end of file
-- 
GitLab