private-model / HunYuanDiT-v1.2 /HunYuan1.2.json
ACCA225's picture
Upload HunYuan1.2.json
33848a9 verified
raw
history blame
8.06 kB
{
"last_node_id": 34,
"last_link_id": 14,
"nodes": [
{
"id": 28,
"type": "VAEDecode",
"pos": [
287,
387
],
"size": {
"0": 210,
"1": 46
},
"flags": {},
"order": 8,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 4
},
{
"name": "vae",
"type": "VAE",
"link": 7
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
5
],
"shape": 3
}
],
"properties": {
"Node name for S&R": "VAEDecode"
}
},
{
"id": 27,
"type": "EmptyLatentImage",
"pos": [
-147,
310
],
"size": {
"0": 315,
"1": 106
},
"flags": {},
"order": 0,
"mode": 0,
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
3
],
"shape": 3
}
],
"properties": {
"Node name for S&R": "EmptyLatentImage"
},
"widgets_values": [
1024,
1024,
1
]
},
{
"id": 26,
"type": "CLIPTextEncode",
"pos": [
-479,
574
],
"size": {
"0": 400,
"1": 200
},
"flags": {},
"order": 5,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 13
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"links": [
2
],
"shape": 3
}
],
"properties": {
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"错误的眼睛,糟糕的人脸,毁容,糟糕的艺术,变形,多余的肢体,模糊的颜色,模糊,重复,病态,残缺,"
]
},
{
"id": 32,
"type": "DiffusersVAELoader",
"pos": [
-123,
463
],
"size": {
"0": 303.9019775390625,
"1": 58
},
"flags": {},
"order": 1,
"mode": 0,
"outputs": [
{
"name": "VAE",
"type": "VAE",
"links": [
7
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "DiffusersVAELoader"
},
"widgets_values": [
"diffusion_pytorch_model.bin"
]
},
{
"id": 24,
"type": "KSampler",
"pos": [
160,
515
],
"size": {
"0": 315,
"1": 262
},
"flags": {},
"order": 7,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 12,
"slot_index": 0
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 1
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 2
},
{
"name": "latent_image",
"type": "LATENT",
"link": 3
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"links": [
4
],
"shape": 3
}
],
"properties": {
"Node name for S&R": "KSampler"
},
"widgets_values": [
42,
"fixed",
50,
6,
"ddpm",
"normal",
1
]
},
{
"id": 34,
"type": "LoraLoader",
"pos": [
94,
125
],
"size": {
"0": 315,
"1": 126
},
"flags": {},
"order": 4,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 10
},
{
"name": "clip",
"type": "CLIP",
"link": 11
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
12
],
"shape": 3,
"slot_index": 0
},
{
"name": "CLIP",
"type": "CLIP",
"links": [
13,
14
],
"shape": 3,
"slot_index": 1
}
],
"properties": {
"Node name for S&R": "LoraLoader"
},
"widgets_values": [
"downloaded_file.safetensors",
0,
0
]
},
{
"id": 29,
"type": "PreviewImage",
"pos": [
529,
200
],
"size": {
"0": 539.4452514648438,
"1": 385.4747619628906
},
"flags": {},
"order": 9,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 5
}
],
"properties": {
"Node name for S&R": "PreviewImage"
}
},
{
"id": 25,
"type": "CLIPTextEncode",
"pos": [
-416,
43
],
"size": {
"0": 400,
"1": 200
},
"flags": {},
"order": 6,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 14
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"links": [
1
],
"shape": 3
}
],
"properties": {
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"1 girl"
]
},
{
"id": 33,
"type": "DiffusersCLIPLoader",
"pos": [
-501,
311
],
"size": {
"0": 315,
"1": 82
},
"flags": {},
"order": 2,
"mode": 0,
"outputs": [
{
"name": "CLIP",
"type": "CLIP",
"links": [
11
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "DiffusersCLIPLoader"
},
"widgets_values": [
"pytorch_model-Copy1.bin",
"pytorch_model.bin"
]
},
{
"id": 31,
"type": "DiffusersCheckpointLoader",
"pos": [
-474,
433
],
"size": {
"0": 315,
"1": 82
},
"flags": {},
"order": 3,
"mode": 0,
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
10
],
"shape": 3,
"slot_index": 0
}
],
"properties": {
"Node name for S&R": "DiffusersCheckpointLoader"
},
"widgets_values": [
"official.pt",
"v1.2"
]
}
],
"links": [
[
1,
25,
0,
24,
1,
"CONDITIONING"
],
[
2,
26,
0,
24,
2,
"CONDITIONING"
],
[
3,
27,
0,
24,
3,
"LATENT"
],
[
4,
24,
0,
28,
0,
"LATENT"
],
[
5,
28,
0,
29,
0,
"IMAGE"
],
[
7,
32,
0,
28,
1,
"VAE"
],
[
10,
31,
0,
34,
0,
"MODEL"
],
[
11,
33,
0,
34,
1,
"CLIP"
],
[
12,
34,
0,
24,
0,
"MODEL"
],
[
13,
34,
1,
26,
0,
"CLIP"
],
[
14,
34,
1,
25,
0,
"CLIP"
]
],
"groups": [],
"config": {},
"extra": {
"ds": {
"scale": 0.672749994932571,
"offset": [
664.5132440927649,
73.75977428794693
]
}
},
"version": 0.4
}