You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+16-5Lines changed: 16 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -126,7 +126,7 @@ cmake .. -DSD_METAL=ON
126
126
cmake --build . --config Release
127
127
```
128
128
129
-
### Using Flash Attention
129
+
##### Using Flash Attention
130
130
131
131
Enabling flash attention reduces memory usage by at least 400 MB. At the moment, it is not supported when CUBLAS is enabled because the kernel implementation is missing.
132
132
@@ -142,7 +142,7 @@ usage: ./bin/sd [arguments]
142
142
143
143
arguments:
144
144
-h, --help show this help message and exit
145
-
-M, --mode [txt2img or img2img]generation mode (default: txt2img)
145
+
-M, --mode [MODEL] run mode (txt2img or img2img or convert, default: txt2img)
146
146
-t, --threads N number of threads to use during computation (default: -1).
147
147
If threads <= 0, then threads will be set to the number of CPU physical cores
148
148
-m, --model [MODEL] path to model
@@ -168,7 +168,8 @@ arguments:
168
168
-s SEED, --seed SEED RNG seed (default: 42, use random seed for < 0)
169
169
-b, --batch-count COUNT number of images to generate.
--clip-skip N number of layers to skip of clip model (default: 0)
171
+
--clip-skip N ignore last layers of CLIP network; 1 ignores none, 2 ignores one layer (default: -1)
172
+
<= 0 represents unspecified, will be 1 for SD1.x, 2 for SD2.x
172
173
--vae-tiling process vae in tiles to reduce memory usage
173
174
-v, --verbose print extra info
174
175
```
@@ -183,6 +184,16 @@ You can specify the model weight type using the `--type` parameter. The weights
183
184
- `q5_0` or `q5_1` for 5-bit integer quantization
184
185
- `q4_0` or `q4_1` for 4-bit integer quantization
185
186
187
+
#### Convert to GGUF
188
+
189
+
You can also convert weights in the formats `ckpt/safetensors/diffusers` to gguf and perform quantization in advance, avoiding the need for quantization every time you load them.
sd -m ../models/v1-5-pruned-emaonly.safetensors -p "a lovely cat" --taesd ../models/diffusion_pytorch_model.safetensors
259
270
```
260
271
261
-
## Using ESRGAN to upscale results
272
+
####Using ESRGAN to upscale results
262
273
263
274
You can use ESRGAN to upscale the generated images. At the moment, only the [RealESRGAN_x4plus_anime_6B.pth](https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.2.4/RealESRGAN_x4plus_anime_6B.pth) model is supported. Support for more models of this architecture will be added soon.
0 commit comments