Crosstyan
commited on
Commit
·
b946723
1
Parent(s):
7395ca4
update new models
Browse files- .gitattributes +1 -0
- README.md +7 -1
- bp_mk3.safetensors +3 -0
- bp_mk5.safetensors +3 -0
.gitattributes
CHANGED
|
@@ -1,3 +1,4 @@
|
|
| 1 |
*.png filter=lfs diff=lfs merge=lfs -text
|
| 2 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 3 |
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 1 |
*.png filter=lfs diff=lfs merge=lfs -text
|
| 2 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
| 3 |
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
| 4 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
|
@@ -16,6 +16,12 @@ widget:
|
|
| 16 |
|
| 17 |
# BPModel
|
| 18 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
BPModel is an experimental Stable Diffusion model based on [ACertainty](https://huggingface.co/JosephusCheung/ACertainty) from [Joseph Cheung](https://huggingface.co/JosephusCheung).
|
| 20 |
|
| 21 |
Why is the Model even existing? There are loads of Stable Diffusion model out there, especially anime style models.
|
|
@@ -145,7 +151,7 @@ LaTeNt SpAcE!
|
|
| 145 |
Use [`bp_1024_with_vae_te.ckpt`](bp_1024_with_vae_te.ckpt) if you don't have VAE and text encoder with you, still
|
| 146 |
EMA weight is not included and it's fp16.
|
| 147 |
|
| 148 |
-
If you want to continue training, use [`bp_1024_e10_ema.ckpt`](bp_1024_e10_ema.ckpt) which is the ema weight
|
| 149 |
and with fp32 precision.
|
| 150 |
|
| 151 |
For better performance, it is strongly recommended to use Clip skip (CLIP stop at last layers) 2.
|
|
|
|
| 16 |
|
| 17 |
# BPModel
|
| 18 |
|
| 19 |
+
## Update
|
| 20 |
+
|
| 21 |
+
2023-01-02: I wasted more GPU hours to train it a little bit more (overfitting). Check out [bp_mk3.safetensors](bp_mk3.safetensors) and [bp_mk5.safetensors](bp_mk5.safetensors). Prepare yourself own VAE! Update your WebUI if you can't load [safetensors](https://github.com/huggingface/safetensors).
|
| 22 |
+
|
| 23 |
+
## Introduction
|
| 24 |
+
|
| 25 |
BPModel is an experimental Stable Diffusion model based on [ACertainty](https://huggingface.co/JosephusCheung/ACertainty) from [Joseph Cheung](https://huggingface.co/JosephusCheung).
|
| 26 |
|
| 27 |
Why is the Model even existing? There are loads of Stable Diffusion model out there, especially anime style models.
|
|
|
|
| 151 |
Use [`bp_1024_with_vae_te.ckpt`](bp_1024_with_vae_te.ckpt) if you don't have VAE and text encoder with you, still
|
| 152 |
EMA weight is not included and it's fp16.
|
| 153 |
|
| 154 |
+
If you want to continue training, use [`bp_1024_e10_ema.ckpt`](bp_1024_e10_ema.ckpt) which is the ema unet weight
|
| 155 |
and with fp32 precision.
|
| 156 |
|
| 157 |
For better performance, it is strongly recommended to use Clip skip (CLIP stop at last layers) 2.
|
bp_mk3.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:97848d7d80b242a1483d0307509b422fee12a0e7096ff202397a6e395a71aea9
|
| 3 |
+
size 1719136903
|
bp_mk5.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f87dabceffd299a8f3f72f031829338e34ad3c1e2541815af08fa694d65fb4c0
|
| 3 |
+
size 1719136903
|