ae362b929321e78dac00be23203737cbe27de9ffb3e77d8c79669ad3bc6da06e
Browse files- README.md +7 -7
- quant_strategy.json +1 -1
README.md
CHANGED
|
@@ -1,13 +1,13 @@
|
|
| 1 |
-
---
|
| 2 |
-
base_model: GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0
|
| 3 |
-
license: apache-2.0
|
| 4 |
-
tags:
|
| 5 |
-
- mlx
|
| 6 |
-
---
|
| 7 |
|
| 8 |
# GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx
|
| 9 |
|
| 10 |
-
This quantized low-bit model [GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx) was converted to MLX format from [`GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0
|
| 11 |
Refer to the [original model card](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0) for more details on the model.
|
| 12 |
|
| 13 |
## Use with mlx
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0
|
| 3 |
+
license: apache-2.0
|
| 4 |
+
tags:
|
| 5 |
+
- mlx
|
| 6 |
+
---
|
| 7 |
|
| 8 |
# GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx
|
| 9 |
|
| 10 |
+
This quantized low-bit model [GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx) was converted to MLX format from [`GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0`](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0) using gbx-lm version **0.3.4**.
|
| 11 |
Refer to the [original model card](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0) for more details on the model.
|
| 12 |
|
| 13 |
## Use with mlx
|
quant_strategy.json
CHANGED
|
@@ -7601,4 +7601,4 @@
|
|
| 7601 |
}
|
| 7602 |
}
|
| 7603 |
}
|
| 7604 |
-
}
|
|
|
|
| 7601 |
}
|
| 7602 |
}
|
| 7603 |
}
|
| 7604 |
+
}
|