Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +81 -0
- .gitignore +5 -0
- .ipynb_checkpoints/README-checkpoint.md +127 -0
- .ipynb_checkpoints/Untitled-checkpoint.ipynb +42 -0
- .python-version +1 -0
- README.md +127 -0
- Untitled.ipynb +42 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00000500.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00001000.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00001500.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00002000.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00002500.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00003000.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00003500.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00004000.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00004500.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00005000.safetensors +3 -0
- ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00005500.safetensors +3 -0
- ani_bright_landscape_w14_outputs/sample/.ipynb_checkpoints/ani_bright_landscape_w14_lora_e000000_00_20250614193330_000-checkpoint.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000100_00_20250614200312_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000100_00_20250614205429_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000200_00_20250614201444_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000200_00_20250614210603_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000300_00_20250614202631_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000300_00_20250614211751_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000400_00_20250614203815_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000400_00_20250614212935_000.png +0 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000500_00_20250614214104_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000600_00_20250614215236_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000700_00_20250614220428_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000800_00_20250614221624_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000900_00_20250614222805_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001000_00_20250614223938_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001100_00_20250614225110_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001200_00_20250614230254_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001300_00_20250614231431_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001400_00_20250614232604_000.png +0 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001500_00_20250614233807_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001600_00_20250614234934_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001700_00_20250615000056_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001800_00_20250615001248_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001900_00_20250615002428_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002000_00_20250615003609_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002100_00_20250615004756_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002200_00_20250615005932_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002300_00_20250615011124_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002400_00_20250615012246_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002500_00_20250615013434_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002600_00_20250615014624_000.png +3 -0
- ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002700_00_20250615015805_000.png +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,84 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
docs/kisekaeichi_ref.png filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
docs/kisekaeichi_result.png filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
docs/kisekaeichi_start.png filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_000400_00_20250607144540_000.png filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_000900_00_20250607154029_000.png filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_001000_00_20250607155150_000.png filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_001100_00_20250607160313_000.png filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_001200_00_20250607161443_000.png filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_002400_00_20250607182338_000.png filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_002500_00_20250607183410_000.png filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_005700_00_20250608001933_000.png filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_005800_00_20250608003038_000.png filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_005900_00_20250608004117_000.png filter=lfs diff=lfs merge=lfs -text
|
| 49 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_006000_00_20250608005157_000.png filter=lfs diff=lfs merge=lfs -text
|
| 50 |
+
ani_landscape_w14_outputs/sample/ani_landscape_w14_lora_e000000_00_20250607140357_000.png filter=lfs diff=lfs merge=lfs -text
|
| 51 |
+
ani_bright_landscape_w14_outputs/sample/.ipynb_checkpoints/ani_bright_landscape_w14_lora_e000000_00_20250614193330_000-checkpoint.png filter=lfs diff=lfs merge=lfs -text
|
| 52 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000100_00_20250614200312_000.png filter=lfs diff=lfs merge=lfs -text
|
| 53 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000100_00_20250614205429_000.png filter=lfs diff=lfs merge=lfs -text
|
| 54 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000200_00_20250614201444_000.png filter=lfs diff=lfs merge=lfs -text
|
| 55 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000200_00_20250614210603_000.png filter=lfs diff=lfs merge=lfs -text
|
| 56 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000300_00_20250614202631_000.png filter=lfs diff=lfs merge=lfs -text
|
| 57 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000300_00_20250614211751_000.png filter=lfs diff=lfs merge=lfs -text
|
| 58 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000400_00_20250614203815_000.png filter=lfs diff=lfs merge=lfs -text
|
| 59 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000500_00_20250614214104_000.png filter=lfs diff=lfs merge=lfs -text
|
| 60 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000600_00_20250614215236_000.png filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000700_00_20250614220428_000.png filter=lfs diff=lfs merge=lfs -text
|
| 62 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000800_00_20250614221624_000.png filter=lfs diff=lfs merge=lfs -text
|
| 63 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000900_00_20250614222805_000.png filter=lfs diff=lfs merge=lfs -text
|
| 64 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001000_00_20250614223938_000.png filter=lfs diff=lfs merge=lfs -text
|
| 65 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001100_00_20250614225110_000.png filter=lfs diff=lfs merge=lfs -text
|
| 66 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001200_00_20250614230254_000.png filter=lfs diff=lfs merge=lfs -text
|
| 67 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001300_00_20250614231431_000.png filter=lfs diff=lfs merge=lfs -text
|
| 68 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001500_00_20250614233807_000.png filter=lfs diff=lfs merge=lfs -text
|
| 69 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001600_00_20250614234934_000.png filter=lfs diff=lfs merge=lfs -text
|
| 70 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001700_00_20250615000056_000.png filter=lfs diff=lfs merge=lfs -text
|
| 71 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001800_00_20250615001248_000.png filter=lfs diff=lfs merge=lfs -text
|
| 72 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001900_00_20250615002428_000.png filter=lfs diff=lfs merge=lfs -text
|
| 73 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002000_00_20250615003609_000.png filter=lfs diff=lfs merge=lfs -text
|
| 74 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002100_00_20250615004756_000.png filter=lfs diff=lfs merge=lfs -text
|
| 75 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002200_00_20250615005932_000.png filter=lfs diff=lfs merge=lfs -text
|
| 76 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002300_00_20250615011124_000.png filter=lfs diff=lfs merge=lfs -text
|
| 77 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002400_00_20250615012246_000.png filter=lfs diff=lfs merge=lfs -text
|
| 78 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002500_00_20250615013434_000.png filter=lfs diff=lfs merge=lfs -text
|
| 79 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002600_00_20250615014624_000.png filter=lfs diff=lfs merge=lfs -text
|
| 80 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002700_00_20250615015805_000.png filter=lfs diff=lfs merge=lfs -text
|
| 81 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002800_00_20250615020938_000.png filter=lfs diff=lfs merge=lfs -text
|
| 82 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002900_00_20250615022111_000.png filter=lfs diff=lfs merge=lfs -text
|
| 83 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003000_00_20250615023241_000.png filter=lfs diff=lfs merge=lfs -text
|
| 84 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003100_00_20250615024417_000.png filter=lfs diff=lfs merge=lfs -text
|
| 85 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003200_00_20250615025601_000.png filter=lfs diff=lfs merge=lfs -text
|
| 86 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003300_00_20250615030731_000.png filter=lfs diff=lfs merge=lfs -text
|
| 87 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003400_00_20250615031919_000.png filter=lfs diff=lfs merge=lfs -text
|
| 88 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003500_00_20250615033107_000.png filter=lfs diff=lfs merge=lfs -text
|
| 89 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003600_00_20250615034258_000.png filter=lfs diff=lfs merge=lfs -text
|
| 90 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003700_00_20250615035416_000.png filter=lfs diff=lfs merge=lfs -text
|
| 91 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003800_00_20250615040611_000.png filter=lfs diff=lfs merge=lfs -text
|
| 92 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_003900_00_20250615041741_000.png filter=lfs diff=lfs merge=lfs -text
|
| 93 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004000_00_20250615042926_000.png filter=lfs diff=lfs merge=lfs -text
|
| 94 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004100_00_20250615044128_000.png filter=lfs diff=lfs merge=lfs -text
|
| 95 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004200_00_20250615045316_000.png filter=lfs diff=lfs merge=lfs -text
|
| 96 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004300_00_20250615050438_000.png filter=lfs diff=lfs merge=lfs -text
|
| 97 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004400_00_20250615051620_000.png filter=lfs diff=lfs merge=lfs -text
|
| 98 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004500_00_20250615052749_000.png filter=lfs diff=lfs merge=lfs -text
|
| 99 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004600_00_20250615053935_000.png filter=lfs diff=lfs merge=lfs -text
|
| 100 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004700_00_20250615055120_000.png filter=lfs diff=lfs merge=lfs -text
|
| 101 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004800_00_20250615060257_000.png filter=lfs diff=lfs merge=lfs -text
|
| 102 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_004900_00_20250615061430_000.png filter=lfs diff=lfs merge=lfs -text
|
| 103 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005000_00_20250615062622_000.png filter=lfs diff=lfs merge=lfs -text
|
| 104 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005100_00_20250615063804_000.png filter=lfs diff=lfs merge=lfs -text
|
| 105 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005200_00_20250615064945_000.png filter=lfs diff=lfs merge=lfs -text
|
| 106 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005300_00_20250615070047_000.png filter=lfs diff=lfs merge=lfs -text
|
| 107 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005400_00_20250615071255_000.png filter=lfs diff=lfs merge=lfs -text
|
| 108 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005500_00_20250615072436_000.png filter=lfs diff=lfs merge=lfs -text
|
| 109 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005600_00_20250615073627_000.png filter=lfs diff=lfs merge=lfs -text
|
| 110 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005700_00_20250615074808_000.png filter=lfs diff=lfs merge=lfs -text
|
| 111 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_005800_00_20250615075935_000.png filter=lfs diff=lfs merge=lfs -text
|
| 112 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_e000000_00_20250614193330_000.png filter=lfs diff=lfs merge=lfs -text
|
| 113 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_e000000_00_20250614193614_000.png filter=lfs diff=lfs merge=lfs -text
|
| 114 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_e000000_00_20250614194106_000.png filter=lfs diff=lfs merge=lfs -text
|
| 115 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_e000000_00_20250614195132_000.png filter=lfs diff=lfs merge=lfs -text
|
| 116 |
+
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_e000000_00_20250614204247_000.png filter=lfs diff=lfs merge=lfs -text
|
.gitignore
ADDED
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
__pycache__/
|
| 2 |
+
.venv
|
| 3 |
+
venv/
|
| 4 |
+
logs/
|
| 5 |
+
uv.lock
|
.ipynb_checkpoints/README-checkpoint.md
ADDED
|
@@ -0,0 +1,127 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Anime Light Landscape Text-to-Video Generation
|
| 2 |
+
|
| 3 |
+
This repository contains the necessary steps and scripts to generate anime-style videos using the Anime_Landscape text-to-video model with LoRA (Low-Rank Adaptation) weights. The model produces anime-style videos based on textual prompts with distinctive geometric and neon aesthetic.
|
| 4 |
+
|
| 5 |
+
## Prerequisites
|
| 6 |
+
|
| 7 |
+
Before proceeding, ensure that you have the following installed on your system:
|
| 8 |
+
|
| 9 |
+
• **Ubuntu** (or a compatible Linux distribution)
|
| 10 |
+
• **Python 3.x**
|
| 11 |
+
• **pip** (Python package manager)
|
| 12 |
+
• **Git**
|
| 13 |
+
• **Git LFS** (Git Large File Storage)
|
| 14 |
+
• **FFmpeg**
|
| 15 |
+
|
| 16 |
+
## Installation
|
| 17 |
+
|
| 18 |
+
1. **Update and Install Dependencies**
|
| 19 |
+
|
| 20 |
+
```bash
|
| 21 |
+
sudo apt-get update && sudo apt-get install cbm git-lfs ffmpeg
|
| 22 |
+
```
|
| 23 |
+
|
| 24 |
+
2. **Clone the Repository**
|
| 25 |
+
|
| 26 |
+
```bash
|
| 27 |
+
git clone https://huggingface.co/svjack/Anime_Bright_Landscape_wan_2_1_14_B_text2video_lora
|
| 28 |
+
cd Anime_Bright_Landscape_wan_2_1_14_B_text2video_lora
|
| 29 |
+
```
|
| 30 |
+
|
| 31 |
+
3. **Install Python Dependencies**
|
| 32 |
+
|
| 33 |
+
```bash
|
| 34 |
+
pip install torch torchvision
|
| 35 |
+
pip install -r requirements.txt
|
| 36 |
+
pip install ascii-magic matplotlib tensorboard huggingface_hub datasets
|
| 37 |
+
pip install moviepy==1.0.3
|
| 38 |
+
pip install sageattention==1.0.6
|
| 39 |
+
```
|
| 40 |
+
|
| 41 |
+
4. **Download Model Weights**
|
| 42 |
+
|
| 43 |
+
```bash
|
| 44 |
+
wget https://huggingface.co/Wan-AI/Wan2.1-T2V-14B/resolve/main/models_t5_umt5-xxl-enc-bf16.pth
|
| 45 |
+
wget https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth
|
| 46 |
+
wget https://huggingface.co/Wan-AI/Wan2.1-T2V-14B/resolve/main/Wan2.1_VAE.pth
|
| 47 |
+
wget https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_t2v_1.3B_bf16.safetensors
|
| 48 |
+
wget https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_t2v_14B_bf16.safetensors
|
| 49 |
+
```
|
| 50 |
+
|
| 51 |
+
## Usage
|
| 52 |
+
|
| 53 |
+
To generate a video, use the `wan_generate_video.py` script with the appropriate parameters.
|
| 54 |
+
|
| 55 |
+
#### Interactive Mode
|
| 56 |
+
For experimenting with different prompts:
|
| 57 |
+
```bash
|
| 58 |
+
python wan_generate_video.py --fp8 --task t2v-14B --video_size 480 832 --video_length 81 --infer_steps 35 \
|
| 59 |
+
--save_path save --output_type both \
|
| 60 |
+
--dit wan2.1_t2v_14B_bf16.safetensors --vae Wan2.1_VAE.pth \
|
| 61 |
+
--t5 models_t5_umt5-xxl-enc-bf16.pth \
|
| 62 |
+
--attn_mode torch \
|
| 63 |
+
--lora_weight ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00005500.safetensors \
|
| 64 |
+
--lora_multiplier 1.0 \
|
| 65 |
+
--interactive
|
| 66 |
+
```
|
| 67 |
+
|
| 68 |
+
```prompt
|
| 69 |
+
```
|
| 70 |
+
|
| 71 |
+
```prompt
|
| 72 |
+
```
|
| 73 |
+
|
| 74 |
+
```prompt
|
| 75 |
+
```
|
| 76 |
+
|
| 77 |
+
```prompt
|
| 78 |
+
```
|
| 79 |
+
|
| 80 |
+
```prompt
|
| 81 |
+
```
|
| 82 |
+
|
| 83 |
+
|
| 84 |
+
|
| 85 |
+
## Key Parameters
|
| 86 |
+
|
| 87 |
+
* `--fp8`: Enable FP8 precision (recommended)
|
| 88 |
+
* `--task`: Model version (`t2v-1.3B`)
|
| 89 |
+
* `--video_size`: Output resolution (e.g., `480 832`)
|
| 90 |
+
* `--video_length`: Number of frames (typically 81)
|
| 91 |
+
* `--infer_steps`: Quality vs speed trade-off (35-50)
|
| 92 |
+
* `--lora_weight`: Path to Kinich LoRA weights
|
| 93 |
+
* `--lora_multiplier`: Strength of LoRA effect (1.0 recommended)
|
| 94 |
+
* `--prompt`: Should include "In the style of Kinich" for best results
|
| 95 |
+
|
| 96 |
+
## Style Characteristics
|
| 97 |
+
|
| 98 |
+
For optimal results, prompts should describe:
|
| 99 |
+
- Characters with geometric neon hair patterns
|
| 100 |
+
- Black outfits with gold/teal designs
|
| 101 |
+
- Futuristic or high-energy backgrounds
|
| 102 |
+
- Vibrant color palettes with glowing elements
|
| 103 |
+
- Dynamic poses and expressions
|
| 104 |
+
|
| 105 |
+
## Output
|
| 106 |
+
|
| 107 |
+
Generated videos and frames will be saved in the specified `save_path` directory with:
|
| 108 |
+
- MP4 video file
|
| 109 |
+
- Individual frames as PNG images
|
| 110 |
+
|
| 111 |
+
## Troubleshooting
|
| 112 |
+
|
| 113 |
+
• Verify all model weights are correctly downloaded
|
| 114 |
+
• Ensure sufficient GPU memory (>=12GB recommended)
|
| 115 |
+
• Check for version conflicts in Python packages
|
| 116 |
+
|
| 117 |
+
## License
|
| 118 |
+
|
| 119 |
+
This project is licensed under the MIT License.
|
| 120 |
+
|
| 121 |
+
## Acknowledgments
|
| 122 |
+
|
| 123 |
+
• **Hugging Face** for model hosting
|
| 124 |
+
• **Wan-AI** for base models
|
| 125 |
+
• **svjack** for LoRA adaptation
|
| 126 |
+
|
| 127 |
+
For support, please open an issue in the repository.
|
.ipynb_checkpoints/Untitled-checkpoint.ipynb
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cells": [
|
| 3 |
+
{
|
| 4 |
+
"cell_type": "code",
|
| 5 |
+
"execution_count": null,
|
| 6 |
+
"id": "a0d2adc9-e517-4906-b730-3fbc16a1a7e5",
|
| 7 |
+
"metadata": {},
|
| 8 |
+
"outputs": [],
|
| 9 |
+
"source": [
|
| 10 |
+
"python wan_generate_video.py --fp8 --task t2v-1.3B --video_size 480 832 --video_length 81 --infer_steps 35 \\\n",
|
| 11 |
+
"--save_path save --output_type both \\\n",
|
| 12 |
+
"--dit aniWan2114BFp8E4m3fn_t2v13B.safetensors --vae Wan2.1_VAE.pth \\\n",
|
| 13 |
+
"--t5 models_t5_umt5-xxl-enc-bf16.pth \\\n",
|
| 14 |
+
"--attn_mode torch \\\n",
|
| 15 |
+
"--lora_weight Kinich_w1_3_outputs/Kinich_w1_3_lora-000070.safetensors \\\n",
|
| 16 |
+
"--lora_multiplier 1.0 \\\n",
|
| 17 |
+
"--interactive\n"
|
| 18 |
+
]
|
| 19 |
+
}
|
| 20 |
+
],
|
| 21 |
+
"metadata": {
|
| 22 |
+
"kernelspec": {
|
| 23 |
+
"display_name": "wan2gp",
|
| 24 |
+
"language": "python",
|
| 25 |
+
"name": "wan2gp"
|
| 26 |
+
},
|
| 27 |
+
"language_info": {
|
| 28 |
+
"codemirror_mode": {
|
| 29 |
+
"name": "ipython",
|
| 30 |
+
"version": 3
|
| 31 |
+
},
|
| 32 |
+
"file_extension": ".py",
|
| 33 |
+
"mimetype": "text/x-python",
|
| 34 |
+
"name": "python",
|
| 35 |
+
"nbconvert_exporter": "python",
|
| 36 |
+
"pygments_lexer": "ipython3",
|
| 37 |
+
"version": "3.10.9"
|
| 38 |
+
}
|
| 39 |
+
},
|
| 40 |
+
"nbformat": 4,
|
| 41 |
+
"nbformat_minor": 5
|
| 42 |
+
}
|
.python-version
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
3.10
|
README.md
ADDED
|
@@ -0,0 +1,127 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Anime Light Landscape Text-to-Video Generation
|
| 2 |
+
|
| 3 |
+
This repository contains the necessary steps and scripts to generate anime-style videos using the Anime_Landscape text-to-video model with LoRA (Low-Rank Adaptation) weights. The model produces anime-style videos based on textual prompts with distinctive geometric and neon aesthetic.
|
| 4 |
+
|
| 5 |
+
## Prerequisites
|
| 6 |
+
|
| 7 |
+
Before proceeding, ensure that you have the following installed on your system:
|
| 8 |
+
|
| 9 |
+
• **Ubuntu** (or a compatible Linux distribution)
|
| 10 |
+
• **Python 3.x**
|
| 11 |
+
• **pip** (Python package manager)
|
| 12 |
+
• **Git**
|
| 13 |
+
• **Git LFS** (Git Large File Storage)
|
| 14 |
+
• **FFmpeg**
|
| 15 |
+
|
| 16 |
+
## Installation
|
| 17 |
+
|
| 18 |
+
1. **Update and Install Dependencies**
|
| 19 |
+
|
| 20 |
+
```bash
|
| 21 |
+
sudo apt-get update && sudo apt-get install cbm git-lfs ffmpeg
|
| 22 |
+
```
|
| 23 |
+
|
| 24 |
+
2. **Clone the Repository**
|
| 25 |
+
|
| 26 |
+
```bash
|
| 27 |
+
git clone https://huggingface.co/svjack/Anime_Bright_Landscape_wan_2_1_14_B_text2video_lora
|
| 28 |
+
cd Anime_Bright_Landscape_wan_2_1_14_B_text2video_lora
|
| 29 |
+
```
|
| 30 |
+
|
| 31 |
+
3. **Install Python Dependencies**
|
| 32 |
+
|
| 33 |
+
```bash
|
| 34 |
+
pip install torch torchvision
|
| 35 |
+
pip install -r requirements.txt
|
| 36 |
+
pip install ascii-magic matplotlib tensorboard huggingface_hub datasets
|
| 37 |
+
pip install moviepy==1.0.3
|
| 38 |
+
pip install sageattention==1.0.6
|
| 39 |
+
```
|
| 40 |
+
|
| 41 |
+
4. **Download Model Weights**
|
| 42 |
+
|
| 43 |
+
```bash
|
| 44 |
+
wget https://huggingface.co/Wan-AI/Wan2.1-T2V-14B/resolve/main/models_t5_umt5-xxl-enc-bf16.pth
|
| 45 |
+
wget https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/models_clip_open-clip-xlm-roberta-large-vit-huge-14.pth
|
| 46 |
+
wget https://huggingface.co/Wan-AI/Wan2.1-T2V-14B/resolve/main/Wan2.1_VAE.pth
|
| 47 |
+
wget https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_t2v_1.3B_bf16.safetensors
|
| 48 |
+
wget https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_t2v_14B_bf16.safetensors
|
| 49 |
+
```
|
| 50 |
+
|
| 51 |
+
## Usage
|
| 52 |
+
|
| 53 |
+
To generate a video, use the `wan_generate_video.py` script with the appropriate parameters.
|
| 54 |
+
|
| 55 |
+
#### Interactive Mode
|
| 56 |
+
For experimenting with different prompts:
|
| 57 |
+
```bash
|
| 58 |
+
python wan_generate_video.py --fp8 --task t2v-14B --video_size 480 832 --video_length 81 --infer_steps 35 \
|
| 59 |
+
--save_path save --output_type both \
|
| 60 |
+
--dit wan2.1_t2v_14B_bf16.safetensors --vae Wan2.1_VAE.pth \
|
| 61 |
+
--t5 models_t5_umt5-xxl-enc-bf16.pth \
|
| 62 |
+
--attn_mode torch \
|
| 63 |
+
--lora_weight ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00005500.safetensors \
|
| 64 |
+
--lora_multiplier 1.0 \
|
| 65 |
+
--interactive
|
| 66 |
+
```
|
| 67 |
+
|
| 68 |
+
```prompt
|
| 69 |
+
```
|
| 70 |
+
|
| 71 |
+
```prompt
|
| 72 |
+
```
|
| 73 |
+
|
| 74 |
+
```prompt
|
| 75 |
+
```
|
| 76 |
+
|
| 77 |
+
```prompt
|
| 78 |
+
```
|
| 79 |
+
|
| 80 |
+
```prompt
|
| 81 |
+
```
|
| 82 |
+
|
| 83 |
+
|
| 84 |
+
|
| 85 |
+
## Key Parameters
|
| 86 |
+
|
| 87 |
+
* `--fp8`: Enable FP8 precision (recommended)
|
| 88 |
+
* `--task`: Model version (`t2v-1.3B`)
|
| 89 |
+
* `--video_size`: Output resolution (e.g., `480 832`)
|
| 90 |
+
* `--video_length`: Number of frames (typically 81)
|
| 91 |
+
* `--infer_steps`: Quality vs speed trade-off (35-50)
|
| 92 |
+
* `--lora_weight`: Path to Kinich LoRA weights
|
| 93 |
+
* `--lora_multiplier`: Strength of LoRA effect (1.0 recommended)
|
| 94 |
+
* `--prompt`: Should include "In the style of Kinich" for best results
|
| 95 |
+
|
| 96 |
+
## Style Characteristics
|
| 97 |
+
|
| 98 |
+
For optimal results, prompts should describe:
|
| 99 |
+
- Characters with geometric neon hair patterns
|
| 100 |
+
- Black outfits with gold/teal designs
|
| 101 |
+
- Futuristic or high-energy backgrounds
|
| 102 |
+
- Vibrant color palettes with glowing elements
|
| 103 |
+
- Dynamic poses and expressions
|
| 104 |
+
|
| 105 |
+
## Output
|
| 106 |
+
|
| 107 |
+
Generated videos and frames will be saved in the specified `save_path` directory with:
|
| 108 |
+
- MP4 video file
|
| 109 |
+
- Individual frames as PNG images
|
| 110 |
+
|
| 111 |
+
## Troubleshooting
|
| 112 |
+
|
| 113 |
+
• Verify all model weights are correctly downloaded
|
| 114 |
+
• Ensure sufficient GPU memory (>=12GB recommended)
|
| 115 |
+
• Check for version conflicts in Python packages
|
| 116 |
+
|
| 117 |
+
## License
|
| 118 |
+
|
| 119 |
+
This project is licensed under the MIT License.
|
| 120 |
+
|
| 121 |
+
## Acknowledgments
|
| 122 |
+
|
| 123 |
+
• **Hugging Face** for model hosting
|
| 124 |
+
• **Wan-AI** for base models
|
| 125 |
+
• **svjack** for LoRA adaptation
|
| 126 |
+
|
| 127 |
+
For support, please open an issue in the repository.
|
Untitled.ipynb
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cells": [
|
| 3 |
+
{
|
| 4 |
+
"cell_type": "code",
|
| 5 |
+
"execution_count": null,
|
| 6 |
+
"id": "a0d2adc9-e517-4906-b730-3fbc16a1a7e5",
|
| 7 |
+
"metadata": {},
|
| 8 |
+
"outputs": [],
|
| 9 |
+
"source": [
|
| 10 |
+
"python wan_generate_video.py --fp8 --task t2v-1.3B --video_size 480 832 --video_length 81 --infer_steps 35 \\\n",
|
| 11 |
+
"--save_path save --output_type both \\\n",
|
| 12 |
+
"--dit aniWan2114BFp8E4m3fn_t2v13B.safetensors --vae Wan2.1_VAE.pth \\\n",
|
| 13 |
+
"--t5 models_t5_umt5-xxl-enc-bf16.pth \\\n",
|
| 14 |
+
"--attn_mode torch \\\n",
|
| 15 |
+
"--lora_weight Kinich_w1_3_outputs/Kinich_w1_3_lora-000070.safetensors \\\n",
|
| 16 |
+
"--lora_multiplier 1.0 \\\n",
|
| 17 |
+
"--interactive\n"
|
| 18 |
+
]
|
| 19 |
+
}
|
| 20 |
+
],
|
| 21 |
+
"metadata": {
|
| 22 |
+
"kernelspec": {
|
| 23 |
+
"display_name": "wan2gp",
|
| 24 |
+
"language": "python",
|
| 25 |
+
"name": "wan2gp"
|
| 26 |
+
},
|
| 27 |
+
"language_info": {
|
| 28 |
+
"codemirror_mode": {
|
| 29 |
+
"name": "ipython",
|
| 30 |
+
"version": 3
|
| 31 |
+
},
|
| 32 |
+
"file_extension": ".py",
|
| 33 |
+
"mimetype": "text/x-python",
|
| 34 |
+
"name": "python",
|
| 35 |
+
"nbconvert_exporter": "python",
|
| 36 |
+
"pygments_lexer": "ipython3",
|
| 37 |
+
"version": "3.10.9"
|
| 38 |
+
}
|
| 39 |
+
},
|
| 40 |
+
"nbformat": 4,
|
| 41 |
+
"nbformat_minor": 5
|
| 42 |
+
}
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00000500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8268e71a0ce3d721d00f18fee64216845eb671b9def5381cf97cdb43c275d3e9
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00001000.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:70804ff7ce7c684d17f1082cc9e3d7a98ecf5099612bfb081e0b9b675048b2b2
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00001500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:232c7e6ef6d4735d27ce4f53fc67aa67d5a4f1b79e87ce72928a1914ab0327d5
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00002000.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:087d5da7ba98f7d7d0150c393946e3ae8cde6ef78ba152ee5c0993e0ba6b2d81
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00002500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b1dd78a5d1f4a6ee7ec0806b7d4e2e7b0dafe0d831e58b71a9a9639d235d5957
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00003000.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a1804b549e7be6ff6099e5040b22549c9a2f238ff753200ce76a4de39c5af6ba
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00003500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3786a74e3d3bb23dead2d3c65187eaa0d90c4cdb224b084b5b71e7e803265415
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00004000.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ad89c2cce43b3baf5a4f211a219503c869b69321226a460abd78b8547c0d9c2f
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00004500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c62ff62e759c2432d403998945e4e740dd65e57fc7a6d0628fd83b7cf067e78e
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00005000.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a3d94100772a310d08afb66b9b3654bf5785011c0628be852c8e856caad1aef
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/ani_bright_landscape_w14_lora-step00005500.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c6aa5ad7f728324b71421dbae565198e54be0cc2b257d7e4406003a00af37345
|
| 3 |
+
size 613557784
|
ani_bright_landscape_w14_outputs/sample/.ipynb_checkpoints/ani_bright_landscape_w14_lora_e000000_00_20250614193330_000-checkpoint.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000100_00_20250614200312_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000100_00_20250614205429_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000200_00_20250614201444_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000200_00_20250614210603_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000300_00_20250614202631_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000300_00_20250614211751_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000400_00_20250614203815_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000400_00_20250614212935_000.png
ADDED
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000500_00_20250614214104_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000600_00_20250614215236_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000700_00_20250614220428_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000800_00_20250614221624_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_000900_00_20250614222805_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001000_00_20250614223938_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001100_00_20250614225110_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001200_00_20250614230254_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001300_00_20250614231431_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001400_00_20250614232604_000.png
ADDED
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001500_00_20250614233807_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001600_00_20250614234934_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001700_00_20250615000056_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001800_00_20250615001248_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_001900_00_20250615002428_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002000_00_20250615003609_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002100_00_20250615004756_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002200_00_20250615005932_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002300_00_20250615011124_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002400_00_20250615012246_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002500_00_20250615013434_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002600_00_20250615014624_000.png
ADDED
|
Git LFS Details
|
ani_bright_landscape_w14_outputs/sample/ani_bright_landscape_w14_lora_002700_00_20250615015805_000.png
ADDED
|
Git LFS Details
|