kundaja-green commited on
Commit
59def2d
·
1 Parent(s): 3124b04

Update start.sh to add arguments

Browse files
Files changed (1) hide show
  1. start.sh +5 -6
start.sh CHANGED
@@ -1,24 +1,20 @@
1
  #!/bin/bash
2
- # --- Final Definitive Startup Script (v23.6 - Points to the correct TOML file) ---
3
 
4
  set -e
5
  echo "--- Startup Script Initialized ---"
6
 
7
- # This path is correct ONLY if persistent storage is enabled at /data
8
  MODELS_DIR="/data/models"
9
  OUTPUT_DIR="/data/output"
10
 
11
- # Create our subdirectories inside the persistent storage
12
  mkdir -p $MODELS_DIR
13
  mkdir -p $OUTPUT_DIR
14
 
15
- # --- Define paths to the model files ---
16
  DIT_PATH="$MODELS_DIR/wan2.1_i2v_720p_14B_fp8_e4m3fn.safetensors"
17
  VAE_PATH="$MODELS_DIR/Wan2.1_VAE.pth"
18
  CLIP_PATH="$MODELS_DIR/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth"
19
  T5_PATH="$MODELS_DIR/models_t5_umt5-xxl-enc-bf16.pth"
20
 
21
- # --- Download models if they don't exist ---
22
  echo "--- Checking for model files in persistent storage... ---"
23
  if [ ! -f "$DIT_PATH" ]; then
24
  echo "Downloading DiT model..."
@@ -40,7 +36,7 @@ fi
40
  echo "--- All models are present. Starting training... ---"
41
  ls -lh $MODELS_DIR
42
 
43
- # --- Run the training command with CORRECTED dataset config ---
44
  accelerate launch wan_train_network.py \
45
  --task="i2v-14B" \
46
  --dit="$DIT_PATH" \
@@ -50,6 +46,9 @@ accelerate launch wan_train_network.py \
50
  --dataset_config="dataset/huggingfacetoml.toml" \
51
  --output_dir="$OUTPUT_DIR" \
52
  --output_name="my-I2V-Lora" \
 
 
 
53
  --max_train_epochs="70" \
54
  --learning_rate="1e-5" \
55
  --optimizer_type="adamw" \
 
1
  #!/bin/bash
2
+ # --- Final Definitive Startup Script (v23.7 - Re-adds necessary LoRA arguments) ---
3
 
4
  set -e
5
  echo "--- Startup Script Initialized ---"
6
 
 
7
  MODELS_DIR="/data/models"
8
  OUTPUT_DIR="/data/output"
9
 
 
10
  mkdir -p $MODELS_DIR
11
  mkdir -p $OUTPUT_DIR
12
 
 
13
  DIT_PATH="$MODELS_DIR/wan2.1_i2v_720p_14B_fp8_e4m3fn.safetensors"
14
  VAE_PATH="$MODELS_DIR/Wan2.1_VAE.pth"
15
  CLIP_PATH="$MODELS_DIR/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth"
16
  T5_PATH="$MODELS_DIR/models_t5_umt5-xxl-enc-bf16.pth"
17
 
 
18
  echo "--- Checking for model files in persistent storage... ---"
19
  if [ ! -f "$DIT_PATH" ]; then
20
  echo "Downloading DiT model..."
 
36
  echo "--- All models are present. Starting training... ---"
37
  ls -lh $MODELS_DIR
38
 
39
+ # --- Run the training command with CORRECT LoRA arguments ---
40
  accelerate launch wan_train_network.py \
41
  --task="i2v-14B" \
42
  --dit="$DIT_PATH" \
 
46
  --dataset_config="dataset/huggingfacetoml.toml" \
47
  --output_dir="$OUTPUT_DIR" \
48
  --output_name="my-I2V-Lora" \
49
+ --network_module="networks.lora_wan" \
50
+ --network_dim="32" \
51
+ --network_alpha="4" \
52
  --max_train_epochs="70" \
53
  --learning_rate="1e-5" \
54
  --optimizer_type="adamw" \