From 6784508d155a621f51a4c0ef67d8f4ee6cac458d Mon Sep 17 00:00:00 2001 From: Tetiana Yemelianenko <tyemel.mzeom@gmail.com> Date: Tue, 12 Nov 2024 14:30:05 +0000 Subject: [PATCH] Upload New File --- .../adapter_config.json | 32 +++++++++++++++++++ 1 file changed, 32 insertions(+) create mode 100644 LoRA_adapters/dinov2-large-finetuned-lora-style/adapter_config.json diff --git a/LoRA_adapters/dinov2-large-finetuned-lora-style/adapter_config.json b/LoRA_adapters/dinov2-large-finetuned-lora-style/adapter_config.json new file mode 100644 index 0000000..86abe99 --- /dev/null +++ b/LoRA_adapters/dinov2-large-finetuned-lora-style/adapter_config.json @@ -0,0 +1,32 @@ +{ + "alpha_pattern": {}, + "auto_mapping": { + "base_model_class": "Dinov2ForImageClassification", + "parent_library": "transformers.models.dinov2.modeling_dinov2" + }, + "base_model_name_or_path": "facebook/dinov2-large", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.1, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": [ + "classifier" + ], + "peft_type": "LORA", + "r": 16, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "query", + "value" + ], + "task_type": null, + "use_rslora": false +} \ No newline at end of file -- GitLab