capitan01r

prodigy_bf16_targeted

Feb 2nd, 2026
132
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 6.19 KB | None | 0 0
  1. job: "extension"
  2. config:
  3. name: "targeted_layers_bf16"
  4. process:
  5. - type: "diffusion_trainer"
  6. training_folder: "/app/ai-toolkit/output"
  7. sqlite_db_path: "./aitk_db.db"
  8. device: "cuda"
  9. trigger_word: null
  10. performance_log_every: 10
  11. network:
  12. type: "lora"
  13. linear: 64
  14. linear_alpha: 64
  15. conv: 32
  16. conv_alpha: 32
  17. lokr_full_rank: true
  18. lokr_factor: -1
  19. network_kwargs:
  20. only_if_contains:
  21. - "layers.14.attention.to_q"
  22. - "layers.14.attention.to_k"
  23. - "layers.14.attention.to_v"
  24. - "layers.14.attention.to_out.0"
  25. - "layers.14.feed_forward.w2"
  26. - "layers.14.feed_forward.w3"
  27. - "layers.15.attention.to_q"
  28. - "layers.15.attention.to_k"
  29. - "layers.15.attention.to_v"
  30. - "layers.15.attention.to_out.0"
  31. - "layers.15.feed_forward.w2"
  32. - "layers.15.feed_forward.w3"
  33. - "layers.16.attention.to_q"
  34. - "layers.16.attention.to_k"
  35. - "layers.16.attention.to_v"
  36. - "layers.16.attention.to_out.0"
  37. - "layers.16.feed_forward.w2"
  38. - "layers.16.feed_forward.w3"
  39. - "layers.17.attention.to_q"
  40. - "layers.17.attention.to_k"
  41. - "layers.17.attention.to_v"
  42. - "layers.17.attention.to_out.0"
  43. - "layers.17.feed_forward.w2"
  44. - "layers.17.feed_forward.w3"
  45. - "layers.18.attention.to_q"
  46. - "layers.18.attention.to_k"
  47. - "layers.18.attention.to_v"
  48. - "layers.18.attention.to_out.0"
  49. - "layers.18.feed_forward.w2"
  50. - "layers.18.feed_forward.w3"
  51. - "layers.19.attention.to_q"
  52. - "layers.19.attention.to_k"
  53. - "layers.19.attention.to_v"
  54. - "layers.19.attention.to_out.0"
  55. - "layers.19.feed_forward.w2"
  56. - "layers.19.feed_forward.w3"
  57. - "layers.20.attention.to_q"
  58. - "layers.20.attention.to_k"
  59. - "layers.20.attention.to_v"
  60. - "layers.20.attention.to_out.0"
  61. - "layers.20.feed_forward.w2"
  62. - "layers.20.feed_forward.w3"
  63. - "layers.21.attention.to_q"
  64. - "layers.21.attention.to_k"
  65. - "layers.21.attention.to_v"
  66. - "layers.21.attention.to_out.0"
  67. - "layers.21.feed_forward.w2"
  68. - "layers.21.feed_forward.w3"
  69. - "layers.22.attention.to_q"
  70. - "layers.22.attention.to_k"
  71. - "layers.22.attention.to_v"
  72. - "layers.22.attention.to_out.0"
  73. - "layers.22.feed_forward.w2"
  74. - "layers.22.feed_forward.w3"
  75. - "layers.23.attention.to_q"
  76. - "layers.23.attention.to_k"
  77. - "layers.23.attention.to_v"
  78. - "layers.23.attention.to_out.0"
  79. - "layers.23.feed_forward.w2"
  80. - "layers.23.feed_forward.w3"
  81. - "layers.24.attention.to_q"
  82. - "layers.24.attention.to_k"
  83. - "layers.24.attention.to_v"
  84. - "layers.24.attention.to_out.0"
  85. - "layers.24.feed_forward.w2"
  86. - "layers.24.feed_forward.w3"
  87. - "layers.25.attention.to_q"
  88. - "layers.25.attention.to_k"
  89. - "layers.25.attention.to_v"
  90. - "layers.25.attention.to_out.0"
  91. - "layers.25.feed_forward.w2"
  92. - "layers.25.feed_forward.w3"
  93. save:
  94. dtype: "bf16"
  95. save_every: 250
  96. max_step_saves_to_keep: 12
  97. save_format: "diffusers"
  98. push_to_hub: false
  99. datasets:
  100. - folder_path: "/app/ai-toolkit/datasets/new"
  101. mask_path: null
  102. mask_min_value: 0.1
  103. default_caption: ""
  104. caption_ext: "txt"
  105. caption_dropout_rate: 0.05
  106. cache_latents_to_disk: true
  107. is_reg: false
  108. network_weight: 1
  109. resolution:
  110. - 512
  111. controls: []
  112. shrink_video_to_frames: true
  113. num_frames: 1
  114. flip_x: false
  115. flip_y: false
  116. num_repeats: 1
  117. train:
  118. batch_size: 1
  119. bypass_guidance_embedding: false
  120. steps: 5000
  121. gradient_accumulation: 1
  122. train_unet: true
  123. train_text_encoder: false
  124. gradient_checkpointing: true
  125. noise_scheduler: "flowmatch"
  126. optimizer: "prodigy"
  127. timestep_type: "sigmoid"
  128. content_or_style: "balanced"
  129. optimizer_params:
  130. weight_decay: 0.01
  131. d_coef: 2
  132. safeguard_warmup: true
  133. use_bias_correction: true
  134. betas:
  135. - 0.9
  136. - 0.99
  137. unload_text_encoder: false
  138. cache_text_embeddings: true
  139. lr: 1
  140. ema_config:
  141. use_ema: false
  142. ema_decay: 0.99
  143. skip_first_sample: true
  144. force_first_sample: false
  145. disable_sampling: false
  146. dtype: "bf16"
  147. diff_output_preservation: false
  148. diff_output_preservation_multiplier: 1
  149. diff_output_preservation_class: "person"
  150. switch_boundary_every: 1
  151. loss_type: "mse"
  152. logging:
  153. log_every: 1
  154. use_ui_logger: true
  155. model:
  156. name_or_path: "Tongyi-MAI/Z-Image-Turbo"
  157. quantize: false
  158. qtype: "qfloat8"
  159. quantize_te: false
  160. qtype_te: "qfloat8"
  161. arch: "zimage:turbo"
  162. low_vram: false
  163. model_kwargs: {}
  164. layer_offloading: false
  165. layer_offloading_text_encoder_percent: 1
  166. layer_offloading_transformer_percent: 1
  167. assistant_lora_path: "ostris/zimage_turbo_training_adapter/zimage_turbo_training_adapter_v2.safetensors"
  168. sample:
  169. sampler: "flowmatch"
  170. sample_every: 250
  171. width: 1024
  172. height: 1024
  173. samples:
  174. - prompt: "from a close range we see, a beautiful woman staring at the camera"
  175. seed: 42
  176. walk_seed: true
  177. guidance_scale: 1
  178. sample_steps: 8
  179. num_frames: 1
  180. fps: 1
  181. meta:
  182. name: "targeted_layers_bf16"
  183. version: "1.0"
Advertisement
Add Comment
Please, Sign In to add comment