cpt core 4
Browse files- README.md +3 -3
- scripts/cpt_core_model_4.py +1 -1
- scripts/requirements-litgpt.in +0 -9
README.md
CHANGED
|
@@ -400,9 +400,9 @@ litgpt convert_pretrained_checkpoint ../out/pretrain-core-3/final ../out/pretrai
|
|
| 400 |
```
|
| 401 |
|
| 402 |
```bash
|
| 403 |
-
litgpt convert_from_litgpt ../out/pretrain-core-3/final ../out/
|
| 404 |
-
cp ../config-
|
| 405 |
-
cp -rv ../tokenizer/* ../out/
|
| 406 |
python -B convert_pth_to_safetensors.py
|
| 407 |
```
|
| 408 |
|
|
|
|
| 400 |
```
|
| 401 |
|
| 402 |
```bash
|
| 403 |
+
litgpt convert_from_litgpt ../out/pretrain-core-3/final ../out/cpt-core-pre-4
|
| 404 |
+
cp ../config-4.json ../out/cpt-core-pre-4/config.json
|
| 405 |
+
cp -rv ../tokenizer/* ../out/cpt-core-pre-4
|
| 406 |
python -B convert_pth_to_safetensors.py
|
| 407 |
```
|
| 408 |
|
scripts/cpt_core_model_4.py
CHANGED
|
@@ -13,7 +13,7 @@ dataset_block_size = 16385
|
|
| 13 |
max_seq_length = 16385
|
| 14 |
dtype = torch.bfloat16
|
| 15 |
load_in_4bit = False
|
| 16 |
-
model_name = '../out/
|
| 17 |
output_dir = '../out/cpt-core-4'
|
| 18 |
|
| 19 |
|
|
|
|
| 13 |
max_seq_length = 16385
|
| 14 |
dtype = torch.bfloat16
|
| 15 |
load_in_4bit = False
|
| 16 |
+
model_name = '../out/cpt-core-pre-4'
|
| 17 |
output_dir = '../out/cpt-core-4'
|
| 18 |
|
| 19 |
|
scripts/requirements-litgpt.in
CHANGED
|
@@ -16,12 +16,3 @@ torchao
|
|
| 16 |
sophia-opt
|
| 17 |
# dolphinflow
|
| 18 |
lm_eval[ifeval,math]
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
# unsloth
|
| 22 |
-
unsloth
|
| 23 |
-
xformers
|
| 24 |
-
trl
|
| 25 |
-
peft
|
| 26 |
-
accelerate
|
| 27 |
-
bitsandbytes
|
|
|
|
| 16 |
sophia-opt
|
| 17 |
# dolphinflow
|
| 18 |
lm_eval[ifeval,math]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|