Update README.md
Browse files
README.md
CHANGED
@@ -12,6 +12,7 @@ pipeline_tag: text-generation
|
|
12 |
|
13 |
[ExLlamaV2 is an inference library for running local LLMs on modern consumer GPUs.](https://github.com/turboderp-org/exllamav2)
|
14 |
|
|
|
15 |
| Filename | Quant type | File Size | Vram*|
|
16 |
| -------- | ---------- | --------- | -------- |
|
17 |
| [phi-4_hb8_3bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_3bpw) | 3.00 bits per weight | 6.66 GB | **10,3 GB** |
|
@@ -20,6 +21,7 @@ pipeline_tag: text-generation
|
|
20 |
| [phi-4_hb8_6bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_6bpw) | 6.00 bits per weight | 11.8 GB | **15,1 GB** |
|
21 |
| [phi-4_hb8_7bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_7bpw) | 7.00 bits per weight | 13.5 GB | **16,7 GB** |
|
22 |
| [phi-4_hb8_8bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_8bpw) | 8.00 bits per weight | 15.2 GB | **18,2 GB** |
|
|
|
23 |
*at 16k context
|
24 |
|
25 |
# Phi-4 Model Card
|
|
|
12 |
|
13 |
[ExLlamaV2 is an inference library for running local LLMs on modern consumer GPUs.](https://github.com/turboderp-org/exllamav2)
|
14 |
|
15 |
+
|
16 |
| Filename | Quant type | File Size | Vram*|
|
17 |
| -------- | ---------- | --------- | -------- |
|
18 |
| [phi-4_hb8_3bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_3bpw) | 3.00 bits per weight | 6.66 GB | **10,3 GB** |
|
|
|
21 |
| [phi-4_hb8_6bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_6bpw) | 6.00 bits per weight | 11.8 GB | **15,1 GB** |
|
22 |
| [phi-4_hb8_7bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_7bpw) | 7.00 bits per weight | 13.5 GB | **16,7 GB** |
|
23 |
| [phi-4_hb8_8bpw](https://huggingface.co/cmh/phi-4_exl2/tree/hb8_8bpw) | 8.00 bits per weight | 15.2 GB | **18,2 GB** |
|
24 |
+
|
25 |
*at 16k context
|
26 |
|
27 |
# Phi-4 Model Card
|