Advertisement
Guest User

XL LoRA guide

a guest
Jul 30th, 2023
692
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 3.09 KB | None | 0 0
  1. {
  2. "LoRA_type": "Standard",
  3. "adaptive_noise_scale": 0,
  4. "additional_parameters": "",
  5. "block_alphas": "",
  6. "block_dims": "",
  7. "block_lr_zero_threshold": "",
  8. "bucket_no_upscale": true,
  9. "bucket_reso_steps": 64,
  10. "cache_latents": true,
  11. "cache_latents_to_disk": true,
  12. "caption_dropout_every_n_epochs": 0.0,
  13. "caption_dropout_rate": 0,
  14. "caption_extension": ".txt",
  15. "clip_skip": "1",
  16. "color_aug": false,
  17. "conv_alpha": 64,
  18. "conv_block_alphas": "",
  19. "conv_block_dims": "",
  20. "conv_dim": 64,
  21. "decompose_both": false,
  22. "dim_from_weights": false,
  23. "down_lr_weight": "",
  24. "enable_bucket": true,
  25. "epoch": 6,
  26. "factor": -1,
  27. "flip_aug": false,
  28. "full_bf16": false,
  29. "full_fp16": false,
  30. "gradient_accumulation_steps": 1.0,
  31. "gradient_checkpointing": true,
  32. "keep_tokens": "0",
  33. "learning_rate": 2.0,
  34. "logging_dir": "",
  35. "lora_network_weights": "",
  36. "lr_scheduler": "constant_with_warmup",
  37. "lr_scheduler_num_cycles": "",
  38. "lr_scheduler_power": "",
  39. "lr_warmup": 0,
  40. "max_bucket_reso": 2048,
  41. "max_data_loader_n_workers": "0",
  42. "max_resolution": "1024,1024",
  43. "max_timestep": 1000,
  44. "max_token_length": "75",
  45. "max_train_epochs": "",
  46. "mem_eff_attn": false,
  47. "mid_lr_weight": "",
  48. "min_bucket_reso": 256,
  49. "min_snr_gamma": 10,
  50. "min_timestep": 0,
  51. "mixed_precision": "bf16",
  52. "model_list": "custom",
  53. "module_dropout": 0.1,
  54. "multires_noise_discount": 0.2,
  55. "multires_noise_iterations": 8,
  56. "network_alpha": 128,
  57. "network_dim": 128,
  58. "network_dropout": 0,
  59. "no_token_padding": false,
  60. "noise_offset": 0.0357,
  61. "noise_offset_type": "Multires",
  62. "num_cpu_threads_per_process": 2,
  63. "optimizer": "Adafactor",
  64. "optimizer_args": "\"scale_parameter=False\", \"relative_step=False\", \"warmup_init=False\" ",
  65. "output_dir": "D:/Downloads/Training/Salma Hayek",
  66. "output_name": "Salma-v1",
  67. "persistent_data_loader_workers": false,
  68. "pretrained_model_name_or_path": "D:/stable-diffusion/ComfyUI/ComfyUI/models/checkpoints/sdXL_v10.safetensors",
  69. "prior_loss_weight": 1.0,
  70. "random_crop": false,
  71. "rank_dropout": 0.1,
  72. "reg_data_dir": "",
  73. "resume": "",
  74. "sample_every_n_epochs": 0,
  75. "sample_every_n_steps": 0,
  76. "sample_prompts": "",
  77. "sample_sampler": "euler_a",
  78. "save_every_n_epochs": 1,
  79. "save_every_n_steps": 0,
  80. "save_last_n_steps": 0,
  81. "save_last_n_steps_state": 0,
  82. "save_model_as": "safetensors",
  83. "save_precision": "fp16",
  84. "save_state": false,
  85. "scale_v_pred_loss_like_noise_pred": false,
  86. "scale_weight_norms": 0,
  87. "sdxl": true,
  88. "sdxl_cache_text_encoder_outputs": true,
  89. "sdxl_no_half_vae": true,
  90. "seed": "",
  91. "shuffle_caption": false,
  92. "stop_text_encoder_training": 0,
  93. "text_encoder_lr": 0.0,
  94. "train_batch_size": 4,
  95. "train_data_dir": "D:/Downloads/Training/Salma Hayek",
  96. "train_on_input": true,
  97. "training_comment": "",
  98. "unet_lr": 2.0,
  99. "unit": 1,
  100. "up_lr_weight": "",
  101. "use_cp": true,
  102. "use_wandb": false,
  103. "v2": false,
  104. "v_parameterization": false,
  105. "vae_batch_size": 0,
  106. "wandb_api_key": "",
  107. "weighted_captions": false,
  108. "xformers": true
  109. }
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement