MaxedOut's picture
Update README.md
d9a95b7
|
raw
history blame
5.63 kB
---
license: other
tags:
- comfyui
- flux
- sdxl
- gguf
- stable diffusion
- t5xxl
- controlnet
- unet
- vae
- model hub
- one click
- upscaler
---
---
💾 Want this whole repo pre-downloaded and installed?
👉 [Support on Patreon](https://www.patreon.com/MaxedOut99) for one-click installers, exclusive LoRAs, and ComfyUI workflows built for mortals.
📺 New to ComfyUI? [Watch the YouTube walkthrough](https://youtu.be/osTqa3SQKx4) and skip the confusion.
---
> ⚠️ **Work in Progress**
> This repo is actively being developed. Especially the the model card. Use as you see fit but know some things won't be accurate or up to date yet.
# ComfyUI-Starter-Packs
> A curated vault of the most essential models for ComfyUI users. Flux1, SDXL, ControlNets, Clips, GGUFs all in one place. Carefully organized.
> Hit the heart at the top next to this repo's name if you find it useful. A small action done by many will let me know working on this repo is helpful.
---
## 🪜 What's Inside
This repo is a **purposeful collection** of the most important models organized in folders so whatever you need is all within that folder:
### Flux1
- **Unet Models**: Dev, Schnell, Depth, Canny, Fill
- **GGUF Versions**: Q3, Q5, Q6 for each major branch
- **Clip + T5XXL** encoders (standard + GGUF versions)
- **Loras**: Only if are especially useful or improve the model.
### SDXL
- **Reccomended Checkpoints to get you started** Pony Realism and Juggernaut
- **Base + Refiner** official models
- **ControlNets**: Depth, Canny, OpenPose, Normal, etc.
### Extra
- VAE, upscalers, and anything required to support workflows
---
## 🏋️ Unet Recommendations (Based on VRAM)
| VRAM | Use Case | Model Type |
|------|----------|-------------|
| 16GB+ | Full FP8 | flux1-dev-fp8.safetensors |
| 12GB | Balanced Q5_K_S | GGUF flux1-dev-Q5_K_S.gguf |
| 8GB | Light Q3_K_S | GGUF flux1-dev-Q3_K_S.gguf |
GGUF models are significantly lighter and designed for **low-VRAM** systems.
## 🧠 T5XXL Recommendations (Based on Ram)
| System RAM | Use Case | Model Type |
|------------|----------|-------------|
| 64GB | Max quality | t5xxl_fp16.safetensors |
| 32GB | High quality (can crash if multitasking) | t5xxl_fp16.safetensors or t5xxl_fp8_scaled.safetensors |
| 16GB | Balanced | t5xxl_fp8_scaled.safetensors |
| <16GB | Low-memory / Safe mode | GGUF Q5_K_S or Q3_K_S |
Quantized t5xxl only directly affects prompt adherence.
> ⚠️ These are **recommended tiers**, not hard rules. RAM usage depends on your active processes, ComfyUI extensions, batch sizes, and other factors.
> If you're getting random crashes, try scaling down one tier.
---
## 🏛 Folder Structure
```
Adetailer/
├─ Ultralytics/bbox/
│ ├─ face_yolov8m.pt
│ └─ hand_yolov8s.pt
└─ sams/
└─ sam_vit_b_01ec64.pth
Flux1/
├─ PuLID/
│ └─ pulid_flux_v0.9.1.safetensors
├─ Style_Models/
│ └─ flux1-redux-dev.safetensors
├─ clip/
│ ├─ ViT-L-14-TEXT-detail-improved-hiT-GmP-HF.safetensors
│ ├─ clip_l.safetensors
│ ├─ t5xxl_fp16.safetensors
│ ├─ t5xxl_fp8_e4m3fn_scaled.safetensors
│ └─ GGUF/
│ ├─ t5-v1_1-xxl-encoder-Q3_K_L.gguf
│ └─ t5-v1_1-xxl-encoder-Q5_K_M.gguf
├─ clip_vision/
│ └─ sigclip_vision_patch14_384.safetensors
├─ vae/
│ └─ ae.safetensors
└─ unet/
├─ Dev/
│ ├─ flux1-dev-fp8.safetensors
│ └─ GGUF/
│ ├─ flux1-dev-Q3_K_S.gguf
│ └─ flux1-dev-Q5_K_S.gguf
├─ Fill/
│ ├─ flux1-fill-dev-fp8.safetensors
│ └─ GGUF/
│ ├─ flux1-fill-dev-Q3_K_S.gguf
│ └─ flux1-fill-dev-Q5_K_S.gguf
├─ Canny/
│ ├─ flux1-canny-dev-fp8.safetensors
│ └─ GGUF/
│ ├─ flux1-canny-dev-Q4_0-GGUF.gguf
│ └─ flux1-canny-dev-Q5_0-GGUF.gguf
├─ Depth/
│ ├─ flux1-depth-dev-fp8.safetensors
│ └─ GGUF/
│ ├─ flux1-depth-dev-Q4_0-GGUF.gguf
│ └─ flux1-depth-dev-Q5_0-GGUF.gguf
└─ Schnell/
├─ flux1-schnell-fp8-e4m3fn.safetensors
└─ GGUF/
├─ flux1-schnell-Q3_K_S.gguf
└─ flux1-schnell-Q5_K_S.gguf
```
---
## 📈 Model Previews (Coming Soon)
I might add a single grid-style graphic showing example outputs:
- **Dev vs Schnell**: Quality vs Speed
- **Depth / Canny / Fill**: Source image → processed map → output
- **SDXL examples**: Realism, Stylized, etc.
All preview images will be grouped into a single efficient visual block for each group.
---
## 📢 Want It Even Easier?
Skip the manual downloads.
🎁 **[Patreon.com/MaxedOut](https://patreon.com)** — (Coming Soon) Get:
- One-click installers for all major Flux & SDXL workflows
- ComfyUI workflows built for beginners and pros
- Behind-the-scenes model picks and tips
---
## ❓ FAQ
**Q: Why not every GGUF?**
A: Because Q3, Q5, and Q6 cover the most meaningful range. No bloat.
**Q: Are these the official models?**
A: Yes. Most are sourced directly from creators, or validated mirrors.
**Q: Will this grow?**
A: Yes. But only with purpose. Not a collection of every model off the face of the earth.
**Q: Why aren’t there more Loras here?**
A: Stylized or niche Loras are showcased on Patreon, where we do deeper dives and examples. Some may get added here later if they become foundational.
---
## ✨ Final Thoughts
You shouldn’t need to hunt through 12 Civitai pages and 6 hugging face repos just to build your ComfyUI folder.
This repo fixes that.