File size: 330 Bytes
3700967
 
 
 
 
 
 
1
2
3
4
5
6
7
Mistral-7B finetuned on a dataset of BTS fanfic at 32k context.

This model uses the alpaca format:
```
{"instruction": "An interaction between a user providing instructions, and an imaginative assistant providing responses.", "input": "...", "output": "..."}
```
Note RoPE scaling parameter **8.0**, with RoPE scaling type linear