File size: 2,502 Bytes
f83b282 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 |
# M-skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5-sft
This model was created as part of the **skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5** experiment using the SkillFactory experiment management system.
## Model Details
- **Training Method**: LLaMAFactory SFT (Supervised Fine-Tuning)
- **Stage Name**: sft
- **Experiment**: skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5
## Training Configuration
{"model_name_or_path": "Qwen/Qwen2.5-1.5B-Instruct", "trust_remote_code": true, "stage": "sft", "do_train": true, "finetuning_type": "full", "deepspeed": "/home/skeh/skill-factory/thirdparty/LLaMA-Factory/examples/deepspeed/ds_z2_config.json", "dataset": "TAUR_dev__skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5", "template": "qwen", "cutoff_len": 16384, "max_samples": 1000000, "overwrite_cache": true, "preprocessing_num_workers": 1, "dataloader_num_workers": 0, "disable_tqdm": false, "output_dir": "/datasets/sedrick/skillfactory/temp/llamafactory/checkpoints", "logging_steps": 10, "save_steps": 100000, "plot_loss": true, "overwrite_output_dir": true, "per_device_train_batch_size": 1, "gradient_accumulation_steps": 1, "learning_rate": 1e-06, "num_train_epochs": 2, "lr_scheduler_type": "cosine", "warmup_ratio": 0.05, "weight_decay": 0.0001, "adam_beta1": 0.9, "adam_beta2": 0.95, "bf16": true, "ddp_timeout": 180000000, "gradient_checkpointing": true, "save_only_model": true, "enable_masked_ranges": false, "save_strategy": "steps", "save_total_limit": 5, "sf_tracker_dataset_id": "TAUR-dev/D-ExpTracker__skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5__v1", "sf_eval_before_training": false, "sf_wandb_project": "skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5_sft", "sf_eval_steps": null, "run_name": "skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5_sft"}
## Experiment Tracking
🔗 **View complete experiment details**: [Experiment Tracker Dataset](https://huggingface.co/datasets/TAUR-dev/D-ExpTracker__skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5__v1)
## Usage
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("TAUR-dev/M-skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5-sft")
model = AutoModelForCausalLM.from_pretrained("TAUR-dev/M-skillfactory_sft_countdown_3arg_promptvariants_qrepeat3_reflections5-sft")
```
|