diff --git a/LoRA_adapters/dinov2-large-finetuned-lora-artist/adapter_config.json b/LoRA_adapters/dinov2-large-finetuned-lora-artist/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..86abe9913a713ad3d07a28e28ce4ac132cf4b3b7 --- /dev/null +++ b/LoRA_adapters/dinov2-large-finetuned-lora-artist/adapter_config.json @@ -0,0 +1,32 @@ +{ + "alpha_pattern": {}, + "auto_mapping": { + "base_model_class": "Dinov2ForImageClassification", + "parent_library": "transformers.models.dinov2.modeling_dinov2" + }, + "base_model_name_or_path": "facebook/dinov2-large", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_dropout": 0.1, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": [ + "classifier" + ], + "peft_type": "LORA", + "r": 16, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "query", + "value" + ], + "task_type": null, + "use_rslora": false +} \ No newline at end of file