gj3ka1 commited on
Commit
3d65edd
verified
1 Parent(s): 8880d1d

Release 0.1.0 beta

Browse files

Release 0.1.0 beta of "AnimaEngine" AI with based-model LoRa and tencent/HunyuanVideo using trojblue dataset

.gitattributes CHANGED
@@ -1,35 +1,37 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.ipynb filter=lfs diff=lfs merge=lfs -text
37
+ *.webp filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,71 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ datasets:
4
+ - trojblue/test-HunyuanVideo-anime-stills
5
+ base_model:
6
+ - tencent/HunyuanVideo
7
+ pipeline_tag: text-to-video
8
+ tags:
9
+ - hunyuan-video
10
+ - text-to-video
11
+ - lora
12
+ - diffusers
13
+ - template:diffusion-lora
14
+ instance_prompt: anime
15
+ widget:
16
+ - text: 'anime girl 1girl, alcohol carton, blush, braid, bridge, crosswalk, dress, green dress, holding carton, long hair, long sleeves, multiple girls, night, open mouth, outdoors, pedestrian bridge, purple eyes, red hair, single braid, solo focus, spaghetti strap'
17
+ output:
18
+ url: samples/ComfyUI_00017_.webp
19
+ - text: '<prompt in metadata>'
20
+ output:
21
+ url: samples/ComfyUI_00024_.webp
22
+ - text: '<prompt in metadata>'
23
+ output:
24
+ url: samples/ComfyUI_00068_.webp
25
+ - text: 'anime scene of a vibrant carnival with colorful rides, games, and food stalls, and a clown handing balloons to a group of laughing children.'
26
+ output:
27
+ url: samples/ComfyUI_00071_.webp
28
  ---
29
+
30
+ # **AnimaEngine**
31
+
32
+ <Gallery />
33
+
34
+
35
+ [v0.1 - testing version]
36
+
37
+
38
+
39
+ An anime-style lora trained on anime screencaps and illustrations, aimed to create vibrant, bright and colorful anime style motions. It's good at generating single person motions (and girls better than boys).
40
+
41
+
42
+
43
+ ## Usage
44
+
45
+ To use the lora (and to use HunyuanVideo in general) in ComfyUI, it's recommended to install the [VideoHelperSuite](https://github.com/Kosinkadink/ComfyUI-VideoHelperSuite), and update to torch 2.5.1+cu124:
46
+
47
+ ```
48
+ pip install -U torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/cu124
49
+ ```
50
+
51
+
52
+
53
+ A sample workflow for using the lora can be found in the Huggingface repo:
54
+
55
+ - [v0.1/ComfyUI_00024_.webp 路 trojblue/HunyuanVideo-lora-AnimeShot at main](https://huggingface.co/trojblue/HunyuanVideo-lora-AnimeShot/blob/main/v0.1/ComfyUI_00024_.webp)
56
+
57
+
58
+
59
+ ## Prompting
60
+
61
+ Use prompts in the format of `anime <subject> <description>` to get the best results, use resolution `544x960` (and usually horizontal works a little bit better than vertical). for example:
62
+
63
+
64
+
65
+ - anime girl with pink twin tails and green eyes, wearing a school uniform, holding a stack of books in a bustling library filled with sunlight streaming through tall windows.
66
+ - anime boy with silver hair and blue eyes, wearing a casual hoodie, sitting on a park bench, feeding pigeons with a gentle smile.
67
+ - anime girl 1girl, alcohol carton, blush, braid, bridge, crosswalk, dress, green dress, holding carton, long hair, long sleeves, multiple girls, night, open mouth, outdoors, pedestrian bridge, purple eyes, red hair, single braid, solo focus, spaghetti strap
68
+
69
+ ## Limitations
70
+
71
+ It's trained as a test model so sometimes when body movements are large it gets disconnected. Also some concepts are less anime-like compared to others. I do plan to update the model later with more training time and dataset.
samples/.DS_Store ADDED
Binary file (6.15 kB). View file
 
samples/ComfyUI_00017_.webp ADDED

Git LFS Details

  • SHA256: 3a1a9f16178c837161df8f98de2583c6126f662c5247a6114f8aace24ea70bcb
  • Pointer size: 132 Bytes
  • Size of remote file: 3.51 MB
samples/ComfyUI_00024_.webp ADDED

Git LFS Details

  • SHA256: 7d8767ac9bdfad92007ea0b8c3ead4f8cca472423aa630133f64a3afca796786
  • Pointer size: 132 Bytes
  • Size of remote file: 1.23 MB
samples/ComfyUI_00068_.webp ADDED

Git LFS Details

  • SHA256: 193c23601cf3a8e868540c87a7ea6fb8e230b49a88c9b49f92ff760f14d19adf
  • Pointer size: 132 Bytes
  • Size of remote file: 1.47 MB
samples/ComfyUI_00071_.webp ADDED

Git LFS Details

  • SHA256: 5c90434a7c4c9fdbbe514687f32d5c156ad6b32bca4149d186bd08f9bf256b29
  • Pointer size: 132 Bytes
  • Size of remote file: 4.01 MB
v0.1/ComfyUI_00024_.webp ADDED

Git LFS Details

  • SHA256: 7d8767ac9bdfad92007ea0b8c3ead4f8cca472423aa630133f64a3afca796786
  • Pointer size: 132 Bytes
  • Size of remote file: 1.23 MB
v0.1/adapter_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": false,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 32,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.0,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 32,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "linear1",
27
+ "linear2",
28
+ "img_mlp.fc1",
29
+ "img_attn_proj",
30
+ "txt_attn_proj",
31
+ "txt_mlp.fc2",
32
+ "img_mod.linear",
33
+ "img_attn_qkv",
34
+ "img_mlp.fc2",
35
+ "modulation.linear",
36
+ "txt_mlp.fc1",
37
+ "txt_mod.linear",
38
+ "txt_attn_qkv"
39
+ ],
40
+ "task_type": null,
41
+ "use_dora": false,
42
+ "use_rslora": false
43
+ }
v0.1/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55eaaaea0420811a2bcd3225a616a6694c696a433843ca71421926309a32faff
3
+ size 322519480