| { |
| "i2v": true, |
| "use_audio": true, |
| "random_prefix_frames": true, |
| "sp_size": 1, |
| "text_encoder_path": "/tmp/pretrained_models/Wan2.1-T2V-14B/models_t5_umt5-xxl-enc-bf16.pth", |
| "image_encoder_path": "None", |
| "dit_path": "/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00001-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00002-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00003-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00004-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00005-of-00006.safetensors,/tmp/pretrained_models/Wan2.1-T2V-14B/diffusion_pytorch_model-00006-of-00006.safetensors", |
| "model_config": { |
| "in_dim": 33, |
| "audio_hidden_size": 32 |
| }, |
| "train_architecture": "lora", |
| "lora_target_modules": "q,k,v,o,ffn.0,ffn.2", |
| "init_lora_weights": "kaiming", |
| "lora_rank": 128, |
| "lora_alpha": 64.0 |
| } |
|
|
|
|