Upload folder using huggingface_hub
Browse files- adapter_config.json +1 -0
- adapter_model.pt +3 -0
- consolidated.00.pth +3 -0
- params.json +12 -0
- tokenizer.model +3 -0
adapter_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"r": 64, "lora_alpha": 128, "target_modules": ["q_proj", "v_proj", "o_proj"], "peft_type": "LORA", "base_model_name_or_path": "meta-llama/Llama-3.2-1B-Instruct"}
|
adapter_model.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ce2bfbd7b3d6da6458a3bb002c6ef859c6a4523d6cb0af3e212b527d08d723e
|
3 |
+
size 22050243
|
consolidated.00.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc17d497df5e4175b3a8acb4f5865b26f7fc1b009b25bef814b95fde10e8a1f3
|
3 |
+
size 2471677246
|
params.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"dim": 2048,
|
3 |
+
"n_layers": 16,
|
4 |
+
"n_heads": 32,
|
5 |
+
"n_kv_heads": 8,
|
6 |
+
"vocab_size": 128256,
|
7 |
+
"ffn_dim_multiplier": 1.5,
|
8 |
+
"multiple_of": 256,
|
9 |
+
"norm_eps": 1e-05,
|
10 |
+
"rope_theta": 500000.0,
|
11 |
+
"use_scaled_rope": true
|
12 |
+
}
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:82e9d31979e92ab929cd544440f129d9ecd797b69e327f80f17e1c50d5551b55
|
3 |
+
size 2183982
|