base_model: AlignmentResearch/Llama-3.3-Tiny-Instruct | |
# Random LoRA Adapter for Reference Model | |
This is a randomly initialized LoRA adapter for the `AlignmentResearch/Llama-3.3-Tiny-Instruct` model, specifically designed for use as a reference model. | |
## Details | |
- **Base model**: AlignmentResearch/Llama-3.3-Tiny-Instruct | |
- **Adapter type**: Reference | |
- **Seed**: 0 | |
- **LoRA rank**: 16 | |
- **LoRA alpha**: 32 | |
- **target modules**: all-linear | |
## Usage | |
```python | |
from peft import PeftModel | |
from transformers import AutoModelForCausalLM, AutoTokenizer | |
# Load base model | |
base_model = AutoModelForCausalLM.from_pretrained("AlignmentResearch/Llama-3.3-Tiny-Instruct") | |
tokenizer = AutoTokenizer.from_pretrained("AlignmentResearch/Llama-3.3-Tiny-Instruct") | |
# Load LoRA adapter | |
model = PeftModel.from_pretrained(base_model, "AlignmentResearch/Llama-3.3-Tiny-Instruct-lora-reference-0") | |
``` | |
This reference adapter was created for testing purposes and contains random weights. | |